var/home/core/zuul-output/0000755000175000017500000000000015135714716014537 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015135726716015506 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log.gz0000644000175000017500000304675115135726537020304 0ustar corecore_wikubelet.log_o[;r)Br'o b-n(!9t%Cs7}g/غIs,r.k9GfD ~6I_翪|mvşo#oVݏKf+ovpZjC4%_̿f\ϘקjzuQ6/㴻|]=ry+/vWŊ7 .=*EbqZnx.h{nۯSa ׋D*%(Ϗ_϶ݬvGR)$DD D~m{]iX\|U. $ॄKЗ/83Jp ώI8&xėv=E|;FmZl8T*v (6pk**+ Le*gUWi [ӊg*ͼ~aT(;`KZ)&@i{ C2i1Gdē _%Kٻւ(Ĩ$#TLX h~lys%v6:SFA֗f΀QՇ2Kݙ$ӎ;IXN :7sL0x.`6)ɚL}ӄ]C }I4Vv@%٘e#dc0Fn 촂iHSr`岮X7̝4?qKf, # qe䧤 ss]QzH.ad!rJBi`V +|i}}THW{y|*/BP3m3A- ZPmN^iL[NrrݝE)~QGGAj^3}wy/{47[q)&c(޸0"$5ڪҾη*t:%?vEmO5tqÜ3Cyu '~qlN?}|nLFR6f8yWxYd ;K44|CK4UQviYDZh$#*)e\W$IAT;s0Gp}=9ڠedۜ+EaH#QtDV:?7#w4r_۾8ZJ%PgS!][5ߜQZ݇~- MR9z_Z;57xh|_/CWuU%v[_((G yMi@'3Pmz8~Y >hl%}Р`sMC77Aztԝp ,}Nptt%q6& ND lM;ָPZGa(X(2*91n,50/mx'})')SĔv}S%xhRe)a@r AF' ]J)ӨbqMWNjʵ2PK-guZZg !M)a(!H/?R?Q~}% ;]/ľv%T&hoP~(*טj=dߛ_SRzSa™:']*}EXɧM<@:jʨΨrPE%NT&1H>g":ͨ ҄v`tYoTq&OzcP_k(PJ'ήYXFgGہwħkIM*򸆔l=q VJީ#b8&RgX2qBMoN w1ђZGd m 2P/Ɛ!" aGd;0RZ+ 9O5KiPc7CDG.b~?|ђP? -8%JNIt"`HP!]ZrͰ4j8!*(jPcǷ!)'xmv>!0[r_G{j 6JYǹ>zs;tc.mctie:x&"bR4S uV8/0%X8Ua0NET݃jYAT` &AD]Ax95mvXYs"(A+/_+*{b }@UP*5ì"M|܊W7|}N{mL=d]' =MS2[3(/hoj$=Zm Mlh>P>Qwf8*c4˥Ęk(+,«.c%_~&^%80=1Jgͤ39(&ʤdH0Ζ@.!)CGt?~=ˢ>f>\bN<Ⱦtë{{b2hKNh`0=/9Gɺɔ+'Х[)9^iX,N&+1Id0ֶ|}!oѶvhu|8Qz:^S-7;k>U~H><~5i ˿7^0*]h,*aklVIKS7d'qAWEݰLkS :}%J6TIsbFʶ褢sFUC)(k-C"TQ[;4j39_WiZSس:$3w}o$[4x:bl=pd9YfAMpIrv̡}XI{B%ZԎuHvhd`Η|ʣ)-iaE';_j{(8xPA*1bv^JLj&DY3#-1*I+g8a@(*%kX{ Z;#es=oi_)qb㼃{buU?zT u]68 QeC Hl @R SFZuU&uRz[2(A1ZK(O5dc}QQufCdX($0j(HX_$GZaPo|P5q @3ǟ6 mR!c/24مQNֆ^n,hU֝cfT :):[gCa?\&IpW$8!+Uph*/ o/{")qq҈78݇hD}H sTB*F$6 2C` |ɧJ~iM cO;m#NV?d?TCg5otޔC1s`u.EkB6ga׬9J2&vV,./ӐoQJ*Dw*^sCeyWtɖ9F.[-cʚmD (QMW`zP~n"U'8%kEq*Lr;TY *BCCpJhxUpܺDoGdlaQ&8#v| (~~yZ-VW"T- 0@4V{g6R/wD_tՄ/F+HP'AE+ J j"b~PO"d.wEр%}5zWˬQOS)ZbF p$^(2JцQImuzhpyXڈ2ͤh}/[g1ieQ*-=hiך5J))?' c9*ϒ]rύW -e]hx&gs7,6BxzxօoFMA['҉F=NGD4sTq1HPld=Q,DQ IJipqc2*;/!~x]y7D7@u邗`unn_ư-a9t_/.9tTo]r8-X{TMYtt =0AMUk}G9^UA,;Tt,"Dxl DfA\w; &`Ͱ٢x'H/jh7hM=~ ֟y[dI~fHIqC۶1Ik\)3 5Ķ']?SؠC"j_6Ÿ9؎]TTjl\D^x6ANbC ]tVUKe$,\ܺI `Qز@UӬ@B {~6caR!=A>\+܁<lW Gϸ}^w'̅dk  C 7fbU{3Se[s %'!?xL 2ڲ]>i+m^CM&WTj7ȗE!NC6P}H`k(FUM gul)b ;2n6'k}ˍ[`-fYX_pL +1wu(#'3"fxsuҮױdy.0]?ݽb+ uV4}rdM$ѢIA$;~Lvigu+]NC5 nNჶT@~ܥ 7-mU,\rXmQALglNʆ P7k%v>"WCyVtnV K`pC?fE?~fjBwU&'ᚡilRї`m] leu]+?T4v\% ;qF0qV(]pP4W =d#t ru\M{Nj.~27)p|Vn60֭l$4԰vg`i{ 6uwŇctyX{>GXg&[ņzP8_ Y+yn~F8I !6WB3C%X)ybLFB%X2U6vw8uUF+X|YukXxVO(+gIQp؎Z{TcR@MSRδ~+1æ|mq՗5$B᲋eY(|*磎\Dži`dZe j'V!Mu@ KV{XץF .Jg< ƜINs:b zĄu3=Az4 u5'og^s7`Rzu-anOIq;6z( rx߅ euPvIɦ7聀t>G;_H;2ʗ6 h6QװxmR JQUbTP2j˔Ni)C)HKE"$ӝ!@2<Bq 2oh80,kNA7,?ע|tC3.㤣TiHEIǢƅKgGnGHRjg#qSn02DŔw:ؽ 5l)Fa/TTmCԤ{"9b{ywSXE*m#3U ùRIvޏrJ`k|wJKH:O*OKy`( ݢe*{ ua ȻݔhvOkU~OǠI/aǕ-JMX _.6KsjA Qsmd  O#F.Uf28ZAgy>y,d$C?O'abk5"'\}6M~VlʵTMF R k%S5<`d+0o,AրcbvJ2O`gA2Ȏp@ lv+ H\FpG)ۏjk_c51̃^cn ba-X/#=Im41NLu\9ETp^poAOO&Aޕ6ql?N/e1N2iDEu&ݛȘPˬ-Ő\B`xr`"F'Iٺ*DnA)yzr^!3Ír!S$,.:+d̋BʺJ#SX*8ҁW7~>oOFe-<uJQ|FZEP__gi(`0/ƍcv7go2G$ N%v$^^&Q 4AMbvvɀ1J{ڔhэK'9*W )IYO;E4z⛢79"hK{BFEmBAΛ3>IO j u߿d{=t-n3Pnef9[}=%G*9sX,¬xS&9'E&"/"ncx}"mV5tŘ:wcZ К G)]$mbXE ^ǽ8%>,0FЕ 6vAVKVCjrD25#Lrv?33Iam:xy`|Q'eű^\ơ' .gygSAixپ im41;P^azl5|JE2z=.wcMԧ ax& =`|#HQ*lS<.U׻`>ajϿ '!9MHK:9#s,jV剤C:LIeHJ"M8P,$N;a-zݸJWc :.<sR6 լ$gu4M*B(A ݖΑِ %H;S*ڳJt>$M!^*n3qESfU, Iĭb#UFJPvBgZvn aE5}~2E|=D' ܇q>8[¿yp/9Om/5|k \6xH.Z'OeCD@cq:Y~<1LٖY9# xe8g IKTQ:+Xg:*}.<M{ZH[^>m0G{ ̷hiOO|9Y"mma[sSbb'Rv&{@6; KE.a\}:<]Oyve3h9}E[kMD,5 %sO{킒 8.K?]i/`׎tp NvԻV4|<{H@#*h{Yp/E%dlh\bU:E%h@&SEK [ Ƣ xg{z%ǻViX~鮦w35QE~qp[ʕ@}ZL! Z0!A⼏q)[f &E1K3i+`JG P/EG 4 9LڑKL|`PОnG#|}qOR{Q|2_tH߫%pD?1%(@nfxOrs25rMլf{sk7݇fjӞh2HkeL'Wʿ}Ƞ%>9cSH|cEyQp 'ˢd:,v-us"Iidw>%zM@9IqrGq:&_p3õB!>9'0LL]M[lwWVR9I5YpVgtuZfG{RoZr3ٮr;wW:͋nqCRu1y=㊻Ij z[|W%q0 CJV٨3,ib{eH7 mҝ(3ɏO/̗-=OR\dIoHZ6n`R֑&#.Mv0vԬ]I˟vrK}F9X|FI#g.Gi)%!iK|o}|ֵ7!ېATJKB2Z/"BfB(gdj۸=}'),-iX'|M2roK\e5Pt:*qSH PgƉU'VKξ ,!3`˞t1Rx}fvvPXdQSg6EDT:dׁz^DjXp͇G|X5Q9K$)U?o': .,wؓaՁ_ 3]Q16ZYafuvrq^ѷQT},!H]6{Jw>%wK{)rH+"B4H7-]r}7v8|׾~Us?yWfv3>xpRҧH-EeJ~4YIozi:nq Vq8swHOzf ̙eX-4`TDGq G.tݻgq74ŠqBFf8 9Fk Afq#ϛa$!qNCJ4bnvB @W,v&- 6wCBjxk9ᤉ ,Asy3YޜZ4ΓVYf'h?kNg?҆8oC!IMo:^G10EY↘H:L@D+dˠUHs[hiҕ|֏G/G`' m5p|:9U8PZ7Yݷ/7cs=v{lLHڕ/=("lX-Ǐ_whaE 98 (oѢ/Р΅ 7ցl6618ł_1/=fu).s¯?.S[{'g=Ҥ):d8h\y6]t1T7IUV:;.1& ,5΀j:<< +Y?58In'bXIǣO{&V\DŽ0,9f O_"[l:h¢8wݓ19\:f6:+ .3}=uvKc ٹeS<>ij(o'ciS<{1$E[nP b?8E'xv[K+E{,Qƙ1*dcs_Z'407|qBOgYU|U--sG8`u! qGYܷw;ȌCPc_|(RaIBKb+{P.T! =ĦiTob d<>SHr][KqWs7ѝBYǭ~RR"p9dFg|K- obY_vM 4>/]e/dy,8!xŋ5 R<^mYo 3c9(F?h04SRm+0^PTi-"] O('@BKD6 {NmʐzRj.aQcb^CZ-uvpr CѐٱlGNzIveca=%1Qi F>wTLHUGӃ\sA֎Xpljlv ^tSȻ \cPwίwX"{>9V0ټ_`#U8VdTtD_GU9V ұ{q:ObUi7s )B ۊZlzIA4S#x,T3ѱ ԶJ=rs>Nb: Q6ˌ߉J%.Dl2ȱ%ܱ&6XƟ6qg(USok+Po$lwvmi8W_VT18V =| ub6QWCnY'"*aN08wuSEAVخ m3 o\` sHc# fqT .,ŀU|⦍߶/*~48âF,#[:y_YIpʼn)dk!J'Z5=r&; (y*b*O_ULT.ÔD[%s1,jЅ@k0Ցu֯dtKl$Y5O*GUڇvI`b0ο0~oI`b#FOf_$0!i rS/wvҍ%Eb/Ec|U9F-)L)ŘF`U:VK jeFrԋ7EDYpԽ.D\dNyj荊EEg]bÔF˩ք%EGƶ*NX)Hc(<|q@Oޯr^3>Uf1w;mCja:-1_k٘%VbZ˙#G6 `q+MPU~l!.?I_Pĝ"] rT [eTr؟˰ ]\ h! v˱>5S1px fnk}sRmA>d2UAkؖvlX܇Bz1U_#Xӫ+al H d\k/I,k,ρ|`zR/$@8VU^rcG"E7\qtS:ڝUyy >Vc11*?xYa8U`Jw/AcL~|;yj8TR#s"Q.ϊ/Yrx+u6*27fǪC%+A~*Zآ'ѭnۡ|< a1s\ T5҃FZh?EV"sd!@БU ^p%pO3|B5=2怕nwRqR9~ i±za+HFNi>. EWz:V^&YEs5Ȭ N *7{!fRБBSۘ† Er/IGU}APQT]|XN X]FbKjKdO U6[3TTX)|*H'2U0:VunBl  `5/@ա06VNO8VGON@KgjyK?Wq1egI+ I.*F~L!Gf"LD&U 6tGd#fR*c ^tSLjnKS9 Ȼ \ >lr&}+̼d"I va,Jm_u)d靕َ| Vw85F3Liƙb<;dM-})C?Fw*IJ_3UG'+¨[9| >80\+ xJpΕ`p~mg˗%F Rg(6=/r+%a>w Ohght uЍaRs ^d6GXAf?V_mW puȇ S:tŴvŀU#-*mZ5k5r)_x*8ͼx@(k:_TX%[paRu~}#Ѥr %A%`;MxB[CzR怕#H% }8@*AM.SEhd,rKrʇ)br\+! s1CtӒNc_:F*`Nv;ogQFa2V%ZniE|nZ&-I,t*ώlo Lhnٓ'Xm R ˍ-~ά}hs\5TT%~am.>!LcoJrKmqvez܅E9t6FZXgsreHhlٷ+ [}r:̓?W~e6>0E8`Jq-(ed;W¨:Ä&]䒿e;0:|$Ȃ1L-%;Ƅ{dɱL;V[bp>!n&աIJX1$9;[?- й vRCxKVV+#lj@_RL;IQ8ŢΌXD@Z< (1ZRÜ:OUM/vư{'jYXE4S/8 7: `/ +G\ U>]B2/n2=8) B gJ3bcKo̹ʇ\B~Is 2sO/I!}xV&\b<9$4Nve^آ]$LGF@LjKٕyzH 31Հm-XıUXF|\A-2) ' RG6h?āUŔyj[j_ӂ~ яA弆^bDyzǖQ8`jXbsK?l58,?YP5䜭ve9YFznTEf3Ja\,@2,?WYؾNr<V` =V[oB5!Z\ļǪЎr8@*ucѡv\[|s L-+y{5K@dzp`r"mũɸHNd"yc Pu>x2;W`_VR<aӗ&D<=h-Rר|/r _ǖھcߖ]G@Ն;UQG1 '3Jە Q88ASUȿ!:WѥLf21;d9OU᧯MR3V:<}xXh//T+coY5Ȧ4/m0NE(G2[+G~H'5ipӘ͏O +Px SPp.,?Uv|$kmnœ/Ke39ifަ۝9; %RNdI%}(q$t u&Hi5JeZEɚUe{ $UŮ*Iq9ݕkwU Jn H`W" xFVe]S5RL%d7;k=*Fëfuʺ;nϒ/yTH!4AY5Hr%s]!b!>))daŖx4 ξ'qEồt:1L{htޤ0b s/^šOc˴s:s;7s& מK4f|8a"Vfi<~Ӷw,ϲ}Ϸ_B!ty#x/}d01=gK$_ ݗN9X1A%Rd2}NwDS&s0d__mT/I3+pP0)?HX"ܪiDD N^m'*|0b{1[ Ŷ3ib*BOXǓ lW:,fV,r01-v\O0a\:"~V0w4+"͋v[? :GqDQ/Z~wDȌ#ak*+bM4W e #ፊz>嵌uߎ:췌OGdͨ@yӹ4d."E")%0)^bUzVOP)539 (Q]X]_G0WQWo,љYӔ=E\O^_5J0x%ýj1 oNFIQ04R +-5o+Rjma Ig]uJڦ%n Nү4 (Gb@bT|E9ʇsVr<:V*Ÿ1݇ Ś"*G%<< l}cG_uWSJ NQ ĭ G4gi1b!g\ig 5[z{1bΒDr)8hQ} Ԗe8Kz|^^M4ώp(% ;<}>n'dy8/x!9댄\ ʙaw4\\sEНI&80}98=ؗuQ09"o$!U0IxH~(z=c5 kbesH'+8V!&C ~hJcM?tٛ#w XD)nո\(SZ5Yw [ߜEܕTS頊g=5i`6O|c{*]:ffͧ2gM%xEqN%6߯5G79qD{݊Q09v- fϨ1}'HhwWYSm˽*A"{sR)1 u-`(3 7wP)OhҀTLN1B'O(޺&%r FBLmM@0K$ !&x3őH`voHӤ6^ H2 pY0W (RyYgOum>L+~.8 AHc#98 T3,+U4pKL|΀~7}ڟ js{`f)O /-7º͓y%`;+44ϥxgw[Hs>~{0{X#Q[uߕF}FUimuik$O!O=Lk`uH|񥨇X|wY vNcw h:~LM˙ ӽ?dk+w5Łi?IߋQVHRvXteZmsJua9V4\5DHxzWxSu5 /Xu y9˺1E7̈=e3+NpGekIm|>< .i!TL4k ϻЮDvTB]BVxPVWck+'%鼕Ro6 !kA2jSPƌcj/(?W9"]?O$l,bkzcɦD:v/ѹڪ 2kj_JuHq[GB&͚}"}=~h`Z x ж$ԞV/%U_z_u9kg]K4u=NLκ#u;< 310ȋoe!X5idVHyQɥT<<&Hx9K6:/HQxilɫqsy$fhb <I%2Gڢ 3fE[ )KRԡK@6WjxZ~7µyZ-n1EU*%f?FvYc?ǓoYN@$fi&VkØY0L-3ucK^=X>gIa&*2oLO4rO!&mw":OV5e:PDg1$FsSZ90xV}SDmB嬉k`kP[%jg=XMiW]lL*mRy4+@ļqi]dӘb`v:BSc f5p< m=$k٪/]Dq_`)nf'nsI-5a)7gfIypvtY"LW@AEʱ"\}vW^l RKm &! ؇⾎x&nʈVl}Z+IOGPhȘcicoʼnMWO7V T6ZHef`aulk@D&r^T7_ 30.S\[AݥnvVb N5 11|ØG[UDU-^n7(6xz{}>y*Y\eN*F \n[ Uגb:(F* C"A{q*e2=G}iҭE".Kglgr׶:axҲx$ܷDNlMB {#ێ}P . [ ,/p5_*r`PG'st1$f CD]sSjҹ_zloQ(Ɩέ]vq9DgMu<ͺC ֤Ү__(:Dt:p@d"pMz]ʖF̳uӲf-_XSzފRm+9~fwhtuW 9*v',ey,hik2!,Qn$\;E ڦff2yp,2MRͲ!w#]|t98ŭR!bFXw7a<ׁxDγ <H:x+bal{SL#֘ܣKjn`漫C&oM33ɳL7y|'GBS0ͽa`Cv2+Y\Vkz7M= b m|[6Y`g9X댺MgD=lӓye!#kOd\țAK6EYdfe@:؈0ʘ;:.2-I$QK3IZ*\ L$NAoV]sc۸kaSn[ܭD0ܭ@­+y\Ba[1Px݊$mKh2tIuj@=e@(`>B_؋]˔8X%]JљOwg콺?'FZQ'NU9a]AmK ! %O&Mzm"r'~<ś T-Y@nA΀ 1popes_!³hhP6X brކ,ض,-j p2ԩX#xaxJY*|ղ !j4yOݩEB4LRN xP_3f.^cbtoMI'*X 'fyOW hkeTlQ{#:EU]O(rm?"`U6x-yczYldIt+pYj*= ?d?=;`lwvj(>PP^eC_A#/ "mŜ:~1Elq|+AyIz@DnO$l[ |C{;HT@/{G~??L._蕳&2~; A {^']E "pƩKK; ?e=EHηSmk>xyg:kzn ax}rQ@. z=1FÕx2") Wྰ `i?|ӏ("^>.`#K>@f#q> zʩGyOwK'Rng /( Lt ѧ$ PQDaQ=0 HE Dx=ɵ©oSH-l@(wG.HD47J۝7;{q{ ,qGg dN#PD߀7;y.yû(z_~e/m|ga,9>E:Fp]i>Fg)wըNNQZ2Wss]"SգFLm8R3-(q`9 ^G<3"+[Ę9xH.i@N`&.Id粺;dlH=!#f< Q-3.HbQX=m;K8^dg#l,W`|6ؗ*=p" qkZ35\͊t]$uY6-=Q6T(u+f/쏈đ}˅ ri ,P8 )V5h],*0>soTc}P>KqY_W_}BAQ> Ȥzl<*A"ػfpt2=o0ncv @l-889$ʶN ]Jwl0Cis[>[8NJ - H&I: 8Us )=CDVRR,\ow\VO>|8|R: tB}/a̽hx N< Iq^/ȩw67tlq(UT%%?u;pBɄM #! >k?.t{\MAg N~bQ~_Ex )g9|rʓrVd'nү%`&j*z/Kg{5f5 ݩijr p61CcX1NU[*u\۬ .ܪUȺ%w޹ɾJ&9h,g"o8 Y^-# }=wj(1Z"g_<)W X8Wj~n vY} /|`z#΀' G(,eh۞0x%x(no%q|"o9:̧ExV`xҷ}B$ W1jxnSv ']0Vo`ge v)s'N|ۋ8ɩ\p 'ċm{fiQÉ8TL L³IRaOat0bv2]3ϙi 5R hpw=[..\8̆~ a{ a0^@ɏQU_=pi 'h&XSϱ9N6,we=/K;M6O=[F֠K=6VZHc>ﰀ;[̧w[|æpxmzMyp}LpzO|-ޠn*!BcC5*j_ !*HHcS%"/0ߪt~d Il v{$4^I ~YDYMލwoV{ I[cU,N1tϨRZPBi(4i{wtk~uԀu ݩ%eEd?>pᣯ2ur SU:V3)$DEΆ[Λu>-8~}p{֢H3_7|5n ċ)C{"e4 e({si/@=w62bvctTO̚GEr%<͡k &/ \8xs*kx۬zzVKȳ&}Y7{V1;7vsFA xP=1%ҒB2@fC8[OllʂC7%hyw}U*=ͰjQckϻCez<;" Y8K6Ivض.._*(6sgC/A1ˇjTYlԾi;n%eDquMP8zhgZ@AcQDR]籔!,Gӝ&"hly;x%ia'-#nyȦhH=+pj~q-0hcSg wLiJ%_TPYZ0x[ \RɁr8LR avĎx.LN}>v<*t.oA謚h:>fG0RU*5 ׂZ]%m(@cT 1m|+76cqЎpU_0-HC  m ʰL!;DEJb/cH26SqirDum-= 35`(b{ܒGx#EmmXM`dj\R JP8"˴tSFi-Zf@:<5Kvhpxqie?ˆ #e{w2C;/ o uhJ 6!nyA ?Ouzƹͨ}Y޳^e|GzfvW-=duןd!G{%d*ԙ5N*(ђ^SL`>RV|"i)͵Jm1j 9kWb*y4ꇛRZŘil͌QI7`bH 'wgw>}{8TYȧB3ܷ2 ic|f CZ[2 {4H3CjO\Mb.UvglYIL~L*Z2Befm)2b$mN[+ItV llS7EW(Y˙#v^' Sjwem%9_ܗUE2EK~$%kf#+aKu%=,jw&e;jV@ XjWcT/I06 hif*;FB=~K F32e]} HUkJ5UB" ^Y! 9ob+:j<~^@=A(bxo1o=+t,P[ )DRCϹMLZ!H~8:ê@bP@/ jtLаEҙ?\߲80rSC U5XK1!EҔ )'czA 5g e"j#|)ԪÿZc}JԈ2Z} di&QS;W\ T .; WL0 fƚVk>K IV߲UwcHo ڎC7o 05LbJ뇶XĒ+APqHS 'YЭ90vS00 UYC" NfXCd`81h5x绕Z*͓P$}$JHd Ia19L@Īj1dR`xcqrҪ¢O),|ı'"XUX)a5'^O攗?GIOq,Ϡ}2'ݩ㿗 \j/. 2a8Hs!^Se[& 26}qIg!H:3sܨ0Vc*J/nTBEbb"FI xʭkZl!s͖As]R`(Hs0 5%Hb#j܅wG4|‚1K#׺%;-Rr謚 W=cBU8s6t"LM5_YpGLxܖ;C7_CD\\W+Vxxx,U%ĔQoSc|x ;cMKE|I}|Yps3Fm`j;'8 H1: 4m" aMAEs6j^YptfU3 #pĉs'lO)zxڻU-B-F(P| s!7Ď*BjId)AE <|޲FlWʂ PN@ AD z\Vj:YI}BsX?YTs7[x07*YI!WRPļLi1v#IMQB>?l'or_ڹ{I ]lК-uۨԼW9A\$Hb_Svr۠K IAl Dd#1%)RPـ MRd65l~'ARQ N͈̻T1oȆ8@Ct^*/=8db-;QӉrwsbl(@F=+ CG4rHt>yp׽`&MS>L#teU.Fe߈a6E_8) x/]V!LPakD Χ4ݷG?Y_+!)\tbH2]3!}.9& =Xc08R]PxU`|jpm v$e.–EmZ`"G?S: :wHϼG Kc}g.J:%(ݗχp>R6}_A:.M_==Vbxz$ CvvU ;R Ưg1E!|wʂuKnqu\|= -߇::h$т>q-hWʃ1hmcnq'G)"ޥfdAZfSbo5ƧUbMsU/2jis7>ej'/mT~0=B'cɆy&s<7  ~nDbPcP䄆4+)ROQBD"/t6M@R`&16mSY**9A#pNUPP{gធwNU+ iʇQ'.*L '_.&@hBQ*NH:˄ǯ,w$(t~Jx$ \jٗ;@dңREnEʹdӻbR0h_e|jλ%q塎@bX=Syu"|տ^8[y~ȭ(2 d:go4bK ALfgr98q(C IbR1̴̭'nv[kr n&jrlR&N < HpzU'{G&nqa2C:Fch0risRG] #$ʷ'eE:@ f1z:9UR(H$:¥Cvoο8aVM=m,'3gFA 8!ZFYavǔL4wm: 'e1к`EqGW|BJ&H:sҥ'/͌q$%g/{ςϯ 3{bv‹*C*l> Yy1K/ v}5;Ε/GEFº_W([owUO:oYXrؠ0hM )u$1c{c[^ *wbM68csQ:69U=GLt<8Ym S((ܞ f\"%ce*&:4檬CEt>e~1>VU~DRu`׵NxCj:v5 w_&m6I}fB%/G +Rs_pJd 41 ǎa߳ª)?+4*繸Dyy{5ʃ{k^E*6׹8g?ow/48r飞\횼%{8LdZGnI1UY2_{DOA<iK}:J]cJF}aFv7tڱ:Uu^j^'3"U1~{X2;3XS\K.Gl&FΤA6j,N!8W(_dɣZ{pQcpۧ{E[Yr,DQ]nQdt>a1L?WԲ\).N/^ Q?^_YpL07sN&\v967EҘv5)+L17.O:Kymd"t>^s~REX?!d'_Jn IdzgB7<8O6z9>M'>/4^oQ6g#ħ͊}ϼ;#媒m$uq {lK.l# (K(tb堊|bPfUNR< wz60e`ܪx $% vrgl/VO" UAZ$93m4/V5!̳jkvמ:;T.@⸘V;(c(cnyr}s?_|K4ZXґw]ΉE™\4ndr=l3CL!HaC06 5G:mm;xwx O!^o ;,ȪG@WkU+lj(djXpvUk܋uGճª.z'{q9.G(; KaOQ~󫚷yM)6M[֏o7G%+a)_AO.^JVRQ߈cz-(i@M}9w=,m6_P LD^Vچ 3vӶ醙 U:usNU #|Fo/;MCVVTf-վ>(߮@0dALoFꗧWMy}2"K -K^MZt51=`z do\MB;q!'H }>2Ym“|sJrh_8}.x"漎ya! `arI)@~9KyM>2n<<;fm$MI$H$PckahbYה[o z8γ*Hi# MW2P`F+Ǘ j~fH<痗PdHBWtD֠|X0̋BUIhs?0D),@Y4`}. e?34a>8.iB>H9u2$_}*]*'L#8ą_A  3$ˌ}С}E EᇉZw$Xзp:n D/s#LƓJK`;l?;_Zn`~Cs4Dt%BUe+BDw5$~V8_Ȇ+>>M,tP ֘K<)Tc8줸WK^0s՛4KRQq9濗p ?5tӵ[ב%rYuTXW kVTRI7@Wo %Dg0@^ &w%2{p.߁yqGҕ}sVX[-1+egqňK3;?\0eLnςk%mS3X $َ5.).Riߔ9]˄Su(UXo] *O_sgŅx2O6ylfW=FY`x?ˆ̾?>da9o8=`| ՜qvnYM1f\7v%KSzA#w%J,M|)(br{33gJHEpfΒٍϜ%qk3g }PڅEд Zʦ:ƆҒgCixySLƤrk*z-T(Md*\wLF*ƵSlZɪP h9;!;PVIW^ĔOJ+dnluJ~ 6HD(m^ :6`.EAa@0R0 g4۴a9ꗋ&\W;@-DJhk]em >i!} Eap{g~>yVO'NP[Bd`! S>fRY&|n\PZNƅiɢj}W#Rdk3<0.Zdh~X B*v:D`\7UJ5LxN 6j%ip% 奣N׿_ŝil/ML(ͣRK[R͋ wǩUc,vt7;3ܳh@+hs;V;C8mmqY"pA T.h'}L`_URw:+C߅u*~xl@{oDaU:88~*rszfUK{fwA-wz̬4KLS;vWIC, +à z#(wRb"^,(qUB ~VmlP+}p-U8F/ѳhK,kj}b3CW꺍Z ۨ:J2*ת띴T-珞`a0}m&{Ue *b veW;6>2nf7'qio2Вbug -ݦQ Sn|rZZgm+;(OVr%KiŪr%Q4ؚrL8Xu$5Tp%@_Q#h%U[5SmkYjlf՟f}Pێr0*+M h:0Bwtu +%i-ik[sl{= M' Ȕ 9(۬ ]))FZf4K6c|% 4͌/spK֩1uR/R=;O/Y.Ojj_EåWCLjϮޯt""ڝU^5tԍ8LPi/0,:8d]+Ǭ=O |d[: Sy&~]r3qfJ"a⇜"/?6rp:} lRAI-kX]V'?R4ٻ㣕˱ZJwX,CD\XcB:SXb ZL|M-M\6QD;Jc[u+&+X|;HT|sqffRm N]CZ34G;wI=rIg󪙸ʁ ZM4X@8̤1c%C:1c =ĵm1a F5ڜykhWBSc >C] j=p&i+gܜ3ᜓ8Aڈ3Y}xϙ+ <"E)Bt4ʁlZe\K8qr&Q9 d hDoi短P|8 ~ 8K20j RYx5g dhvX\ZȅH{BbzãabX0q$fjG0꜊sj-PJ1Z)ĺ+#g0=JPh6*LfHΪ|p` `n sJi̼rTB8%(4 &9I=0[9VCvF^یv1҃ VE}"4 !40${Ch bw,WZ! Ug^D!cq\ PjE SU{M &a2tt5JXof:mP]&TN~7/G,x{Dpt9"#܌|BH12O`;(/ۆ]bp"9\%ZOnn!%@w1C!"M/wë/W]Sv%e;W;Wst.Ԟ|ɔ5O`p"UO1]kzRz^YOw 3kNx!浧}1}T(.f*|xw<ɯä}6/kTvc"jt˸?}x#h3 Ĥ aV7w +KF2ͻ$׷h / ;F"1aR)Ci5Р݊ g 081`tE=L , sXyy P-^YHHϽg18 +bbsIf 7"(*뼟Obn2[N(2HHQm#h&JĬwc΅Huƻw]h.{D$.}!NRiHWQQÂa7qiRH.R`:1f{%8j0i,p#Jd@^Ec0b}KEVySMu>scqƸ{yRAq ʧ"bFYD3WPMK6Ήv&x4G :Jb`P/tvQO 5Zw@`w@\{w4Ƶv ! f8HF-E(x 'k 2V`c.m|`3I͚/uOvF68QDz`m>omKo/jck%DF Ѳ4ZK:9*d")F"ךr{`W&'tK V_mԁz)1>tAY5(10&!XHF2GW\S)dRo"NZKRzhL)#?{Wߣ6俊)~And'n6J:(K{ fvwÜZ2]]]*@""!,q OB1HTQD8:s0V#8H!e=Rte_teq^9}shdxb;L-ɍ8#+: ohu<пЭ/C\%Y]e1!XR<ـ ڰ#G+Xž[E tET?BXHH5Q CE "XLl ,e`rC)j3AR{G~R#XStoV"X@Ej(m|O7 0jZs]H RD %XZK#QiY f'f6c)u.QD*%j6[!-W(m!JEJMTc ih"@Dh10qh3**g:V*t<qy"ǡ1,!3 Ba"aS! "Z>#ad HR%D+O*\( s0&s6<5J4p?I2%:K4@u!F2D $)NpBQːHC@0 A(uHϯT#X6]J7 *QI:¹=JF W%U]1ڭp%5`U!l FpjJ~.NR#7oT>]IJVhy!% :`%Qay&.߀X$5XA(Q$TlUk g׀`5K.LOک7^]JK`csIEM:X$z…H>]%5x`<ߊhC T'k5=Fn.G0+*,ޙVP,+C>]L5! |+µhHT"ShxJKDu"YW`rԂnv)h>F}mAs[ NGMPa9.MKM{v>T"RˆȤX7Z`BNV\lRjA=^hRna]ĉv]4UOzgn.mi]0D_bN 3U0s%͓@lZ띍+Y0;zZ^ w&BVzJVh^5xS꫐Io>_VEUWGc 2.§{4+X7ө6'p5=KĹ?(HC^g%R) rYĝy 7@z0 -0[ZUpy6@ҳ~l2f=m>&gFdN[m[m } GZP<^GQ٣ ۫UɅ+KǛ?g٬/\^2_y]{[Ec*>a- uH@+sFr~޼X(Y**ơA X4!Dl8cjQmnA2M[.׶XWGq!VfdGoQa-}֓}c]="h>~|ߑ{\ȕdLqߝoQw>:;' Wu%;Sm)ync6 jH"*4&`l]5WC*ahR&&2h`U`RPr׳!)v7[V(r{vNŸ7i=<2-3Xe#b lԊ%W6Drwհps!n'1КM 9Hsv0dgG"Ǝ'A1 {m}fy \`5]l%iz[O'@SG) 09~y[5ҿ,%} ן@;U{XgOSC}8įgnݩ7)c7I?F?o~*˳ZY%rio_B6*ÛWHٌ!RF%\/*ED?*]KjJW0[k ir'\k )SW՗ M'_:K2f K0tOWW˽dEԳ:omb*U^O/DU/wxr% 뢤6<&%}MqP]BH}Zy%L|!Z5EKNNaTMOꥵ;⌧j&YYTٰyx0 l:NEwa"[Fs6l6?U? ,L^d˝OGVDR4TZP+deda+C 7Nedͮc;S‰{och&- + 4ި:$^ 2o]{k7r)~1zyCQBt1`Ơpαt\V^zYǘ8uFI#R}/=iUc*D/i ?'f|;]sHyWrK ܣf}9KXM27pE:wWb_9v:16ߎoM?2gc Ze$ݛ_K ߴvsymܼY;shkxe/E#k/ gV lq4x5H-X* d6R8XZU uU^'UY??rx =76*a 3G7%z`h[^Fch^b ʼ{ 1]ᾖc1ˍ[-(7T]񉷈i l{`k-^{Y5FO ä$ øA޼[(w)- #PlF< B9B9:Z:ZTjS'մ,7T%r6d+B' ct .0V&<.PH`c,P,60sX1Tk%i n? =v}MZy0k#iW~?YmEVboF$@e$[lbW?qxYhCr5ѳ6ƼŵRHU2WC{JηiyViWkbifkw;לbџv;m6!*Pܒɠ>%X(E(PD(dE@O b}pr(r <LCEpR*JHJH!MHrjuA;s\O>ɲۡl hG%Vf= ܼ8l36Efmp}\z6 G(,2_dWv_eXr"8hHzz?c?a:L+к / ̼AR{Pӻ__Y1u\':W1岽D$h"$q`O4`p:T42$!cFy,cI6\U FdԺ̻ + QUx=شpB3@? kNb9qyM&g&~HI Ik< `hOl59XX[3i͆0[ lzI  FA93qDI0D׶u0jf8S(h%sc9lk.o7f;5%߾}0۩Tvo~sN\qo(Hq*IW~lI ՜rXW oKՂ=UyRPx n6< n76< >a*#nó[Ԇgmxv݆gmxp?L-$k3@[S fm3 Tk km$I'ck{ ػ7"[&)(*VJYH_fBT1+3_dϸ_ xkȂaEaYvt|ɞ_/o_Z|{_W{nFVS_zAi!EnfI| x9N|*UQz%{X VbϾE!YbDQXyWiK:wMm'>72rkA2%7+)%y*N|*y<*9 %D""|K|J_%Ōy0ħ  q~fqEg7n@A|itu=wco![!˹ꦷ_SoBx;Q]3u_>g,S '}'@y-N|*.?;}!R֜c%Jyd s+IZ_Y yq!gh=6"$(>⸀3L)C<PStlQ[B K7rJ(K٢߃QS!x$/I*2+N2!{( f[ǫOYO^x2{b &%3T>92`hF'QS!! S^T3Y?tv1}ߛ#(ooojs }߿nW+__T^"#Κ.FGHB!NIE3^B [ׄO%0_9^c뜸۠Xm8"'xbfgkTk1$+s/Kf;}wrSp5?7#L4`XIzqq>ڐ)xw(xw! 4!B Ƙ{ħbtmrJǠ3y.ut"1ON$Jxȉ<7+?yTކt,O0^o9ѨC+1Y3-f3+NRA)@tBBt0yQ =(j^!_?MLl~Oknq:޳ԲӨ]=ox=B?MBωˣhS$`P` U=q7Lm>Zk΄n FɭA 5g<vi s_xXu!3#hJhck8 S BvpF+u/FEO}=d;AI%XFE %g1(uC}PQSA5i.QARpHH۔Y>#>[7[<6HK2 0]cjҠp 2`B"t]HD-/ѸˈKR) (Xr)k-%V$aLB'3ħB.e32 Q 8]F[❄)e&8j=#]y7Ҳm+|c@8J8/[&bFHS2~2J|"bGEUm,eg^?ƹ4$ԴgϾZؽ$tart\y"6rpoy/N|*jq݈1W_X X_[(qo q :8J|*f>!6]<&j$?gKN@=3 2`T*CPS!wf6ȿݨjD h#XE0[k/ tV‰E.:&i V Ș{ ħ,l~&Õ֔\3pNA(ħ]LЎ,"5Tq@#z`#f=wlHUHyhdZdf"'PDPS6rb 3ph ;Ua㤧xբKr\#:*1u-*>mQY4$m!$7 ݕd$h)&XɀSBAO4P\ m]ꀂF%.շJǑHK #^$V˵Q1: fcV`,.vxc-bu4T8"K8U%$嚒v X REUR)hu#tnT+/]aԶ\+k[:Qd$׏HXx"r?fɗh9a.0Yk-r=&ݍhF h/ 7<%ͤ#G]ܡx^ܡ | v?Y"*l4L\!b՛G3яi`oי 1Ă_ר H W]a4A=Z&I|,7ϻ6$MuW 㦥R2 N;8m%>r[*z%Wbܥe&,[&ǁѳ\F|*"x: KTw!cBno~KR3_/ ~WF%l4 E}AB8`>t&["ħn6L| 40b1baU`з<݇%>f]`_,W8ih|&I!TbF TMb^OvX*oxzEr&ۻ.s] oSms&jd;vdl jO>~] u O]IQ11F~)*&.^\ٵ"W\Pu2t@.ZonhmyCJ5Z_c hR")5~S^ZEXypȂrms:_>x="qvDbħB^bO6".SN*oN~cG{bqpG|40ȹ`'`'J|*)2dTIb&G'Wur#E6K'‹(Px^8Q>ʚlW?zh1euiD l$a&3PcW&3ZTZHV!|q e)gF˝9VVp-P\/"M`:dq[}$I,$ARige`-O)Z7non( 祥0z<󜘑,S#Bp,+B%Rasq"bCħB^X15[[K ,X 1*$sZD^C-%>T]zNzύq7x4VvTSRq&im--PL$6N3BGFObz^=cnia%}aRREYS $7AEkf?JNxIeuf1"'$[J0-oL)GzzOes2J|Sm'0݀Y,ZGLj~3`M 8`ޯ_ۏ4AkB[dzC~1ch )fޏ ƛmE|3`a+TS,(H VXůVVt(f_i_J+-oUaHDMd)#$FΆ8 z} } 梭C]%@?,}G 22ɚW%B[5Xc@΄ C~b,b\Sjfyf[^ִs-K(akJ !x\Q.1Au\T{ 6Z'feY [J^ ,G-6[.8mS~h Pʅ^|3On0b8l4d7j>pM@7;U>5q%*ф#ʨuAWoѿE*Iw-'gH0y:abuݤRܭ>X|>~a^^M6qP:dNLg  ;~22kks\^NʸRxSK "O ̐,˔ztp95]ˇiJp5 <'ʏΞ C^ .08iHBLU(WW W^:ZI֜d$#I?3'0d=1W0lpHz:úF'.ǧ$^ 9,>ֻ1xp4j*׻>@ax|6)=VDdޟ=34}rgR*!;cVTAc֗J[UK?og @O<8;|Rf=J3_ g K7;4z mp:0J/ڗk郳&M;-P>es{bHFw4y-e耋 @n:yNxYzlx> ZÈi} #xG 7yie6 f-0L[PC}gFKp&⡕d*d HhXu;뒂A(bmQD/(vƽ20ZzAȾF5}E7nDIPquhךL 1ҭ]1JrnQe Ve1쐬ôE<\wɧEDÉsR>4(ma7'T& f ԥٸ1[7CTxHN)_@Oֆz)S1v6b1Y%c4de0ڕ,@LzSd4l;̈RK1RN/!^Rth6!n):̣?ʈQkl2ͭ|2fnR/#^s sbzx, ڥ!H1FQED%Y.HZ84Nsp&DJԫ3&8UDɆkbthPocގļ!vcmDܱc4qԕȉ<xFs8c~(hDꍧz:b, *tsTMGk`Zb"~%) W&! utciMLӈ ʽ[#kb( 1BH 4 gIsw8 [78 f1@44b RTtH8Un(OdNGU*G#PoEkcmQ)cH ]VhPl6Ԇ̤O1p*?׆zn `6bb[ PhY2"DSn_amAUN9)2>#^Eǘ7T%ZqMreDY qrKb~ Sf`CI&c`#(V|VC[Bio-LȹOiqJ C*~1S砢~.l_^Q>b2D<Ƶ#i"B4Aʔ+pQI2cZdRE!v?TIimJ_}?7~^r~fA>- ӻPo.TSl?gHl9$ކJZ57}-w+;/oMofX$6~JѽtF|K4VW*4o&Xm'żQNq_ Oߘڎ曷uc;ljoƇ[)(V d],%faU^Y7KP>FWSJHǝրqtQP@<;Ph }Zۊ"`43p3G°t ~Χ%Lol֋+?v;sd~2['soˉށ/'OC'׺vRShP#֛q]͐xSl1X+eC\>{ y惡PEo|&kw$Dӷm΂u6{k?gW緫rz~~|7-CuԮbWE]2y^N-,"0ᐮ?*;Cƒ: ^΂D_yf͠ɽϮI-yoa\yBۭZ4fBתZo..(S O%(m5MА=!0Q>=|f<1!h3"x#8#^6ȴ.Ic ՜p KB }߯9īl"bW#'^F&1Ù͸s)Ԅ]qPKM~b(OW~mhF݆ c*Aˢ40"c,u$\ BKJ2A yF՞n4WꍵCk1Fv:D%R>Y7KIM!> #Wtm̤Q)d"z-UГ0al⌲^bZDJ"_ʔ4&Oa[),*dVM0$_ !jFQ4Y i~PD3N%s)skxOs?D@~Mu1@'$WM~W.|6]Vᘸ$vjǯއO?NVF* 6c$)g& 9ӫߟa؄6Փ`a\?3,~~Gw~:_E*I?HƙUAMNz󇬃A4qYutWd{1 u[b8h2ǸiN|a HxdىUƬJ}m~wv: KRx[ݮU ! :0a.5{h[006ܾ{5rr7E\ppY}W=啥h} 9e.o6գ_xG`sĭ9^Fn$˚u|o P㾜,e*\5M U`zj&Nɯン|gM1<p[@&iܔ07 Lݛ}m_sǡ9{܋lsj͏6p5V@%;}l $?rK=J|շM:n9I-Ѱ(7Z¬3zd W饦Ag[ Iah<;X C#IA2' BSYx{>/PG.v䢈 #ݛ9撵YĞ;. rN5d"3nWjX܌嚥 R)P96"%+Xo"AUf:!G^O"xEiʄ͈C9˴)\C?pi=D\%,VfU7[n?W 4r|`.>HpO"P-UL;RQCer;\ )]o7=G%:jzS4=_w]{~^]|ɏd:64)ӆ5:'-w Ռsi`&DG-켎:. y/~uRƶVee%^5ָrš磌*C~Ye ,=C6ȉ XEGNAxoOCh/N5STcp^ M%UiT41+c >9}Ka&e2^,%(qmIvaie5*b:WlTT v!˂=V RD('2{~_~w~խ‡n_C7u?+?9*~L 2נg++_M!=4 ;fq<\wMu9D-˔!0Lǃ&W~ȥ4fl\d%4ʖi1ܲq[?xnHImo2x2p7*q=U}ƫ$e[[FsV%#'X EԧE#Wm+;;$fL 0)X׽;S'(g7LY))fU[ ͺN8Ο_f 9R cI8w^X d86hKKr\z+HX%0ϣbfWt!?Br-E*Ccu2ϲH/Mm#(cq*0$ "WF`wO05pde|s۷軘QQ+zٮhei^5EMc80A3h2*,rV@^(dVsi +A,Eqg h.3$I;!D4rC"V+eD \AȺ%b8:+\#ѐh'7 Rho1Q:/S0+n|{gZbʳI `VjdzJ Ed $!ޭtL0`YmDAOʛٟӕü0-* b֐l߂kk9H4tuc4\-@V|<c AbWhX_*% 99Q'cҚQak4EWk"PPTb"r: wBdjcdEE4I9 %V6-UT5d/s<99 x"v(N1 9]c17W})Іjq֤v jS0㘤y%sAe-SWN6L1X:2Z-"qz :cn?0{`q`'p6tFJ?)̌qo2>ML˟jd E첩`j8]\eVzh"6xbEkQ:E5aE$|E ZA4.{J8g8`Kܑ8y\>ymn A^)!ߍp9hgQJn B B?*ڝ"'$UN?hInG(lu7vDc\2vc^+_[=^ְYTq=yet kz-}4558XKM`KYAVLH5.0oFaosԌ($kQ:"CjOKcԁf (*~s6o|JU n\kDuoMSK06rD P4ìF3!߾>Ƞ!4DJQ7p#N3Qp˛,m{C&3s ~z6Ov.1ښY(Y"=yG l]5u,_l̮6NCMd`JIl71gE;fŁCYO, w+;!2x)e>QU8]2 @4By+֤|y?f7o39V(JeY# O״WWZ69 弚B}3,Q$X`Ƕ%b͐&`~`P"Q.XヷMye7|{ees-ƛ5ҥH[rdH(^"\tO[Ƌ(gBW뤹a$ҏL/}"0+nM-9;t6S6#Ƿ O5bZģ1.B'QN cim  P2|۹JtͲ⠼P< eRu 3*:t b 11f1Ƹo.l := l QmjIyn$MfH'" 20dN+ސ'F~Ɂ:u&V>u$,gbfmhIɌ"N:Js&\ܪDm^k@׀>Dc L&f6A kO[ fp4DHiF e1;|qD#d&8*]5P6O^~a(joML^ZO jjcfD*ʈuJMVLpP-&|v<8i1! lqU`v0 Xa2a2=)C#q% tD|D Z(I*PDI#-0.GEm|5#R5cx> PI (% JP49.6C߶m^joE@Jgc"jvM_AWQk?OAvkb"Ҽ!L(bVBYm\&NhkUXcc+xt4h2/@J!LBJLP? .HZyAQ^0sc{AbTyJ#4촓_ڐh ;pG| (< wa16-Vk4bxc ! 6b]63tטc=GXQ ̢J`Ҿx6 m"quA?a~ŠiV:]m;bu6r<~s+m_ۏ1ěko'*ypgߦ D>d½#Tw~Znefo5d9>UAKV4,0än>̏.A|5YՀ i"7b[NWBTR]Y?{.T?n݁SQUQcF2J]Sc-[!r^mPP]aEo oщ&I i~ĺنo?8Rr$2$qΛx^HƙUqZ{yY%Ry*]'=v+i紨@ >>|PrE=F T;#%ūRLp nz}̤ۅ=@m3Re-Qn/=w~}۽:r}lz{Z(AL)ZXEN'^CR zw2 Npr:lX'= ng5JovM9훀~NXigHo f+ 8*qND>>baw؄s㵣#RUkbqٯ{'w%ևM=POc!|K9N!lޑGb4-FqfnHHǎQz|)B^*mfkݞ9R$dBՙc1ڨU.;r\5y{Ag,-+y g{ŮKSKrjmΣdGިzƚȮЃ*9GE!?0%9VƃbWaMlJA5!6=cI_$<3*ZmWF0-ode*- e˯NLg _ŹK:U,md|㡾Y4d;Ux.yIk٪feELJDs4qI6+ޝlM5CФ_ƋmsHEJZsC (%a5\IҺ|XPVj<[/<˸l.ZH@'ip  Pvn2 ݕm6e㫱z1ssmap A2āAҹyv ̉>Cϸ{Sf~Ҕ7n֥.{?MMOk{V+ߚ/أgUa1ϗq2"a4%ey .I &8(D%m)pF1NV!Jw)Z#6n;:t9]y*=#DzrRw tвxql8vbBw$'L 'ke"wىh'_a<9;<[dp?Ͽ/'YV=c ͮ!yrqޑi$:O۾=㱃'hI.=zƏ?46aMs~e=#ǹmz|N>3#wᮌ׸+iPF*,A5Noh&9B(ʣJBu%{^.9 F0A%+P ΡuY:&sxe dE\F"٤g ))C_2qI9u{F p}4f{՞qA-&(Y#Y ’((߀Ʒq+wU /k{ڢP"qlWP[ eJٍ,sÙ'I݀kRwL$$QD)t20 HOL> Sxk+\ݴF? W}?`ٸ cGd=#aY8eidGGTjG!Cd2c A8$t}@}Os-txѐSfN=.͇:)n ? V݌EJMl@hĬHlX,\AAS*=6(e{vK#08g$9)he$a .)UbzZL͏*\^^+oTum;r3t\MɀUf*Կam?]9aSA2gO(4 LXÓY#@j8)):BJ4́hC#08r|[)e[ 1Ƃ!qNd|p*<[9~'#!{TvЭwQL( u~p`$<8 'Sx\$3XY=I@ZŠI2 +&q!2I$*- TD:e }bTxCFXp(:]f\ǵŵ┘px(nr cEKN fkt1z%GHߙj_;E?i80Z[XUlYyix(x{W%f^Uz@Km9F<X^d{hCIn#tA+b h&\(U$`2paqaH^~)/}yj z7XfYY7j  M1^“ Q %~#E :#>m<4X:1nKoN*LjzA>p#y$xrArKj 6YRda )!_9]C#08tC#08>.#'GH 8Nɦ9. 6$ pDCϦ͚ʼn& Iغ1HM0R74#NW֊Ewt@`X$Bc&>u*F|=4˨.xy9j}x`8<.:ngS`e{ -fxP`'Өxk/`%NŎUX{񾹞Te!I+)s]+]\)Ŏ ~t&DDeDkf+։aք% N2T1. C#48kFS=n70tXݪr Ue.U3$_` ;Us{hC`%z0k0G*1vx{q]NcuX|&N\ pC#08J8ʉb6;BXl`U GJ8F)C#08CA~8% d ۹8[,rnC¿=%.+%1 oL1`PAGѼ ua9卺r-[htLAa{bϸ xhc ːv[DlqߎuQd2, ըT47wS3,@,wb&2E*˦8a@`uLٰjYͼ)Ʒէ_דPOTvHI 4,5f;Woؿ}\)Z`y:wyI4_4Re ' S@>}H6|<<x"重RUVwMxp-|N}dscSπwv"n-Y Q 5/ԵM4l۷DEk]UU:ߔ,%EWE?*yWm3$WkGhc_uuiMw8u<9l}mN͘k=P{yYtc$umͬSԛ`ăK[u@<[99'EQߊ q<X;yj؋q2z[G;^cMsyrhMc| _Gc_^ٶm}:wʂN26^EV~?{gTn~]qx}B].xe{ߡeRkX5vѺB4Uykh6snJ{< ^t-j_7Ze%efIfiOlrtW>_Qt2)V?eN.MxڝJxNRD@v=[ytr\|F'g2)FI'9peQ1x'[mfmr4᫑_Se"ĦLQG+2Ums$FiSL88GT0ܹ`lw!s\[ F6#dYVUt-'ZHLuzvQ a(Hv4Gg@vr nf ہSf#}~X,_~7tJRΡ|uo~|d:yi/'6^ jgv}| |"r 6L%"LTrb̖BsI\ fbujnv{)}цd-j_H'c'1e6rƯ+њkjks{ȗ-=C1̵}Mb :a aJQ /(XIH!D^,KlqRx|;ΛEXg&-6ߚ|5w9rhn)S/ttwe~~ESud=moȭSݼ;vF2ZؚeͳzXWuKc`F|}suv=F{:YW7mq7g?k`G}?l|UuY`Dj YtCCZwuBS3=,[>4}kjIZЁ6lrBvlme4&je==?OpzO2)%1_ Ūy6nk*]4gHN+X/#+VN|4Mt )褖XwTOYD;L4xI֬/Eţw I61t?a4(6Xc׫`]m=~' Ս:^5zW}5=CRx[%z^5wYGu!^.I@V `ZA3/m9 V99cD],,MeL1҄% "sUDڶ`&S{NeRhwqH_iKn#A+;qx pgA`!z^QbW잷G#MF) r4CU,F,}`rV5Z@,`;{w^x2޴>]wGGu>MIds͂kozlK:MOݤs,0"`r14W+^!U@ `zRHe iĂ6@|AM+hfg"΂mILF IEd!TµKo@4@2AgAC(K?LC+6F&TA$!Xe)Iu ,ɼ5b[3.AĿ$O5mH}$EjW<r4 Na/l`'h ƌr'${f, ܿGUaՈZ;~S37E?q/:c>}9+Wvpn~9ˮL~ğ,: As=γC2bZ{:X!Su.S1eESaw|Og@0`uTB0iQpiT~=9ֵL‹(A~0Թ:qM r賿FZD7i{;kѸ?:Z(qMh+4߂.|j5k W}TU.0rliLb?+WݯՃE`M7_^^ռz'Q6>ތ+)ߌ=?xj 4MݐnjfQX@i iXg]|8N]Ml$hsA64Vuɯz-WOu$vA|~^MalΆFZFcYƺqf+`9x⏟pquqӋ' GHjg~݉ tǭu6]SŶNmkoѯDm;#؂TlζJ(=>|=waU'I/ؓ37 jޟeIU?R'UT~}G)gH3>Do WG4b[e44ޫDwn$YH0s,`9.EPd$=a|{I%Z;?k>d,Tַ3뷖F3i,#:F /Ja^]!1#63'C;SYuAP¥N+AҏoF;oc6C ĒmfX ߉=hw~d`jN#B Q7ؿ_"TL*ܘŴ2ôW/ɲP toۜ5is)gu䬯Yoc$/!g ([ 4a-dZ;2pH9;&~ǀ?eK~C9Xa'2_t-uX)Ė9@'O`w_@ B9# Ƃ(q;&ܪ0„ K@ X|r B($2f4Ў9:,PFXp4Aw%qs9~cDUw`h= n+ÙH/x kQMضB'A@=nmy}X .m.:KUilo\b ^熺6%ںAji- "F@ZX /1WZ(~MPnm<% 5qccoum2_&8ɇBiKԦCyxpRV]t787o0JT :A*c2A*cFG[7|4{9X#eEWr& neNӖY`8]fkT1};U[]pe͸rꟂ͘ IcO in ZqzSF>xh3 tM6sɬ2zTGb1ᵍMߔ>,.Ga4cQrQ2{XwP?lwt,GO}x\o.Nޟ]`.Nqzqvcocv&םлN?~ܺkLp[]CӮb[tx0|~%js Qԧ2fvV Dy6P=G1[E~bikE܍+.WlP~e)h4QQ*Uh6](r1liU*wUM>1[¾hR ?ۓЯH"/8`j'*8XU`bM&S`A܊RW+3Ѻ:~ZOo>M6aoJNLWztӫ"m~i p=oq=zLI2Ht.7 o-^X]&9Qd6.Lm_$HI)G;P.R*HIT#%HI5RRT#%HI5R^]rۯ|';Fp&)"gB\ ]X<^OJO*]*LtEzy?0[ō*Džj"P.3"% 3Bs4<KY0G~a,+ 7@}(սe5Clo݊LW?h+'uM,` $4z 4f DZvwdV2>y R{;J"] )ӂ5ssiSC<_lo\S'wmmI >$lmvuKܘ|3Xz =1>l7 }?Ϣ3^B/p(:no/b{jhmA tC³Gws˸/gqقg9np1 9HIa)Wɳ:jC5 -^Iy,>ZyKuK!:ƌƞ1AC NVEX-Q^x٨hv4IKُ'zw'b2|y+QH/ļjQ-3aۢWf>gHJ7r,r`AӤBQQx`r6Y@2WiZKz{% v\^١v)vwMH0Haxg3Қus.`YG6V"ēQr c˭Rp 3<Z(Rg{S; )qb 1O$)Kb<yHi WzO938y1Wd'6/0sǿtڻF.^sZiNV-s:޵>q#e_rHx?T%T* bTHJ5fHj(< G 45_#wݹ0M>2\s(8 !FZWP[5{8һnJתҵt*]Jתҵt*]JתRHt*]/]Jתҵt*]Jתҵt*]Jתҵt*]Jתҵt*]Jתҵt*]Jתt%t* %At*]Jתҵt*h(JD~] ,A/kNY[dzh$I,)yJF)2c >"XG)IB<2`d,ji${j'q)#i/FZx?lR Kޡ-N@5lJcTDx aK MxO/$R= ӽ/J$;T2`=Rmc9^JlPo)JRMKHGtM/GU0Xn76K7gMpM5vnWdN{?*x%Xni7ŁY|Whw^.hkjϪǰXЄiu.gIu@.: v*Yj՟1S?SzOXi\m?!NJN:7 <9w?F H"zy\.G_V,WY}IjO?T%7i۪>˶0M 0$-NQ2Yo& 9t]aky)_|Ks0Vs 1mm!?0SŐsY}tȟ{;w nRa\ q dǵ~}{7+PAtgn_,=~P˗g:yX:66H0>F[Xz9>NHknX `ʏw h9'i)q_ϚӼ Η&:(Nċ%^;[I3$TV7Ne{.@o(۰J絁y5 -I O]^u^8vR5 ;~ "45/6wiAm}y3Kz1E W#.0JX^Z3إƂ:YcM/>D spr A(cq¥,R9G S\Gֺ9S%f#{Nmdg+hZOsMSh!ݑrBuv!=»+XK鷟PU.ٜ4 vfjTpM&~A#1cڋH U>8tTDGpnAc=NԵHo%n~j.?/q8p}u5)qv 0f͘Qq?h.΃? K ګ~w_[)1\KBl)>``tcl_n'xxlps9w?LxV;DԦ&:Lxk jVAq'{Hn%~o?7`٠^"|=CI0=C#O<>( O{.D)q%&#-l SꋸО>zE\R8 IhkVHrRR1ZO. Qp6i,52HL) *EՏ[T]*H܁ 1 bIm&40"c֢Dd=W!bO| ~bT(.R1Gf4ZNlM3U6-]xGw|0F X6=_kJЀ`p_[>!\FCP+@1fX6 a _7 i0I4;_\Qcpſ9S}x22dUm뫦|}t4=ִuHrW5J 7NyNtҶ4gѡ/(?d>0onk4#Ir4d[޺b;r6 U8RTau Y}S1`3(KLѿ~ƼHq r6`]{ӹUCFr#^Uׯ`07}_^ohul J>ކ}⏻]0OsXڅ#9Zv1gzd딪ߎon۪k'5ۆvǵR6yX]Ӊտ"rio+S#w=b?o6b.6sͺ]|lνd;W=̷wev1w]ݹeCZFS;]&W{b-c]flp^uMa(QOk>Q'JDI(%}^JDI(%}O>Q'JDIx X!r3鮓&juoFo/nEkhn0 .0oJu7'zp@ "Pu!BjO7YcK|l-%>ǖc >LNQ peWG`6'GRhkܺ4;5IbMG߆Dj^WW)-G 80>:.4y!|m PmF4qbAF|%z2['8Ν$Hxa,v ^k !: ^B9|95 ]]:(q*+ѽ{ƺ(>iO]ۛaX?HC`Ś5U8ysyƐԣ:#GÜT4Շt̲xHd$itH!{Ѓ쩬9$v}kZ6 L$> =Q='~"0#:r>{fi⦻haNE*[$thOFN>6I4Q?v$:P'X]q*ҠHQهO}⸙k {YaPIj_/0̏0rf6q2E PF-!WVs{ 9";“Lyˑ`8LUHn#+mH a<`ws,dqk}J)R)yU )g$H3"[auWuum.J0,=%h`0nC=T[dѓkcsJK0V8 |2)c=Ėu K"yA_]o_Dz|^RJmzc bޏG5;ЪŹJV6Oީ|XýG҃TE V:j6Gls%ic-aKa.*hBZ).F|j#^PWzF|G]Y;kG+k=H.s꽪D:E58jAxc,H/$yw-4Lh*"6FB)+g`"=wV[QF2$/\hXw:^gTone_U9A{2֪Bl~r&\ɏIW(fvv0,>aݐRܑ'.r1H E4B먅KK>j ؖ9:|AXILOT){& thE-L(3@q^]cPgqʶQdO=Y?dUykϸ$XIc`FK0l0# LaP'Q>hc-M{#@728#e$@b|n! I5Ӡ`Mzzqea`3x&lmmh4M^TT^m;ӯ8 -e8ޖr}-]45CItV'`b^gɧ7m/ƋC̱f1R|5M6JYUsj#a쪺w=YeqJNh]vJ\c6Q :˒x|3;%@p4 >rvq9l9ͪ 8la_n4-+U핡oz/BeB4}b;Ҍ˲~ӞpuIz7&$1wrGL"TODh#* 02F420ʉ88I۰y>;<;PꬒX_&%Uc1yJxP)Ir(N^Pȡ<./) PL'6pmA%<wZy0UUny,e%K9[\5,{e],݆2|,5x愔rW [37NCEK$2'|!@|` e0t=֊iǨ1"f%(iK< %r,e'UVX4VyUhloԅB .DnHj#LB9a#‹6;P(|( 3oH$ѐ05"-s`3߅B 桝jo3M$0n6Bףu(EbmN`C˜BK]ZMBPJ NC^4F }y (`⹷,Â"+Õ'qMDyҢ](`2H(-,9bDִ1߁B 9Q;B-8d|.,&U5.J0/0+}cIFs#Q cW&AXB%\ EWg0MDu#CJԖ +*J[{:yet0nN/mP(oo@#pMVgyPT&$BӖ:fDcVmt(ֻ]0lME,ւ}cjmX@+^x5S( xqݲхB}U'nh4Oݍ߮"gid _r4>0B7;UJ}ш LU5z/ ӆ)cz+1 c*]G#ZE怕[D/HpqYX'kǭ5}m'eV:넏F̂pޮ 'f{b` c8tE-5]/ZWZ0{s}]3 e¹6Sᳰ[ԏF<ֺV{Φ=gɶKw?o"Lm9JL ͑JC]@ mloxWR|x_.-mp}X1X= J;öJ]Zd҇s{;w5r z'7q,`+h*t1QJnROM\+-,P(m_dOw=BA MtͧN&kj̋Ịi4jD1 n$_x%~,˦dM} "*ˇ5s3gLcۂ6 2\lS? C@~L!`n4"!L;=XmcɎ ׃Zr4]P q>}XeME`m#VG a5\sH<Pb:$cI#q2 񘻠 6KsygLX nt~Z\&fk>9k_߳zT#)L; 9 RG"/z“W;d: 3V"C:8->9zfh$kbR:Lg ȭ^߂kwZmxpL qwǺW`Nj-7`j4;wkwEo<>.dm=7Shҵr?KUiqWbVy|0ʈ*\=uwjl C Τjʣ+g2 _p byBLeAJ4x&BaDrf耥ǹH 6wMPoD}7?7Wo|qөd>l0 {䣆fl!pb* igi%62nu4[mMr=)ZKfa$m kc\Wbo_IKҩ p Nٓ8eb{۳)x盐';vc0ԗur+0b9f#cCBNS}Q{ۊod9ކ']:DSki_{ f+t=}|8ltrl`s|.:?OU6F)X;M X s[ <pF%R 69k5I,}SAqMں \AZqTw ȱfM|zq;͌WKm "Jg]uYN튶Igi}.2 oWUIݼsؕ)Jsod囗ۉSyXO7y^_ąi&G`r62bItQyAŬid =!"}JlX{))qJ7Dd%0l\&pOM$yiFv՗f~piNHWc(6^=z-7۳4Hye8"jKHR :E™ȍlʛEy޲3)&5M/}=18eYނo{icBJ:bTw1}9ʃ+ZP&-qg;E9>N:-"EFbA1Yb>AhnTTDυ%gqSic2$0|YhQeCx&^ W`#bWUWcP:K;(ڱ{ՎGR,f(Y Y DuBi]V)Q0!S&ptek R5,4p^]WBWZPɍ+`i-RUd]`XF 9SD.ŕb.tnt" u1(RJ(XK ۇ+8} tqK/c0*N`,>,iσH)HFɶ,WZ7o$?*H"L ނ.oL OS?a#L>>^0XAzr1Ÿ_ՓogՁٖh*r0f7Wm~mI4G?#w ~K/ ==1]ݐa4.P0Q1ӧ>onofM&#A^wsɲi: Is'.5}? q$q/`\egzզ٨jfT=k?9û!|ǏoxǷ (0 .ݕ5¿>]O[tMk59g='x~%9n`5)77n DOoGKiߍo='8=UԐ`MW`ؠN9y>*M4[҄(#AyU^f+!Z` 3CVZ2?$G-ucvPy҅\\D2ghQoxYjiV#gIiz<>%<ͬ9kmwXw8bxj`oI]a-jыN *Yy Gԙ ]Ti˨ϓcc3;Kﲝ4I\jQc3>TYskȶ IN`dt;BptY8/Xo<:+\uwX$[EQv˦|i@mtOˊ{,W#I␠ik-I)w- ly9GYT@P /,h@aL )&}tiKă#\& "ъN u.D$Y-eɼ2e֊Ax8/;3)R|;9<g'>Ck6;?X]űt+N\6:D\|J- e~3xW}9efWt ~ &Y vwD;|`tBa}t6~RjkoPA沂8[_?]XaXch*Q㺸򿋫 h"`_aE$XLjA܋۪;5g|@SByå @3@cĜE NVF"&߯\Tnן`I)~uۆAJ<3a4{ =*Cyob0X6}X_hE4ۍ7%epV%zvio2_cyexQaE(VhP}˒΢VM*UX㊻W+ 03m,B0`G+,DB2""&ZH0<)c"9 y74~sDcmYu:_uNQ+gqW 7^r/h@~SvM\7n 9]Y5QZڂ. v0&YWi;3Ϋ ;B^z; \yc8|uB²H4H fޛYia@0)pį3@,/aQBƒEiCbZ3cB)7j<#"(R`X+\$JH!Q¨uA8p9g'=b/yW7:C╷/=;#IvJM>_q v!sFp&T((r`BB ]8Azt4C@\wLtXGCǀn7‭iU"bK(F[ :aFhNP'swvYم\֪?S*vzwCcc6G֟f9]'Qz4z"4J@*m\wdi;ؔ}ؔ5D7E)'L&Lk 5QX qMBAK b"sͰך<"ov-WgF7Ayir:.C0!B*Caj3jX$"﵌FMFS-Y ʃ]js] hW;K CRłQ.(qCvYi PTJ\9޹uZ|M>ߍov z̈́%|n79ؖ]9Ȫh-J)R>RL@Kn飶ěU`̙Yt\^jrDinUCZ/olUnom@hd0'ZK(Iv yI"R:"$nN6`mCہ~=# \HL3 r>0GdrmXcgcZadPg($^2"YobA% 9Ij$f.ˡc;kZi kA)RK ztY|wl7)}Пd_!V[K9ZިpJunJUjt퐿wS Z۝f-1  Z˖;M|ThIWaTLHTkľxf {{%eKR_zWo *gֆ~ n@vJB@aXx m~Ot)2p|Ovܼ'3dgD=< jyvq Pma"0)4]]pM':$Gr*V(Z~4j øGU7i(݆ۡgHےTBw 6MϢxfSڌЎFvm<>יėVҒ(~${|5w4ю>洂ztIv ۷uVihfOd[!.̼EYnjMGu`~3l~8Yywx5шda-ʔvRgpAކM{XP]Hiltuen`T|w LWkuAY!۽=-7~Xk([[2 >?>KWX[=3^̌Ʒf,LMFm4aզ'm ifܾ /f㸅tp;h{>)6l?OvsC ;!XKr;›@N~kIfhp)MW".phng'Ȫ(\R[fQ# eDćlDξe_5(xU}!.&0LO)UU\*W)}\ƇKst1'k?Yo6w՛<;"%a?{Wq`K>ʼn΃w B-9JUġ(ig VaU]z!Uxd!R&R/5eDDL ` XyPI+O'5HxhZ~x=j.uQQ^[% ,J /aβoZy|%T8 y Қ)RRqDpCK,s@)!0lioodA;5 Wsld@`Rznl&1C'\3h/_ _52f˘cͺ:92cLL17 <񀖗S I*;InԣDjӰkpH&θӓ6ǹ=^)boOu|;u]sa IHTTtZEcJ߰0U&a">7hM֫܄\~p3pe?{x+PM]sч%b2";kZ(ڜ-rm|1jw1ٺq#/W^k i8nu,~*yG߀1T)b+t@Hm < Xe|ʹuo`,%Wyb*FK~ l_7=_6dA> l %y 'k,[6,ry wWn}7?'<0?9l$@ ~Fo3ZkվՀ5DX¶EZGBZZ&;' +rӃu %7"JQXX[eXobUM5(9&r6ȹQjEw9lVJH?B"I QʁSt6Nzbm‘mcp+ޛ-;УX%z [G'8rsFp&T)r"gB\ ]8AZtw< aG8܎ chxu'x?1[ō8`1waʽ3ø)arv} _(ha.$7UC 2wIob ^qoy[exR !JFFT|@b7F@;gwJ0u6)Z^NMvFF/pX @I<.~o"p $gp$#ΥaA@Q-~ 0(_⢯P?|E5brqBt<,zn\nE@s0 7ZH|$H$#iKrŻ_>iWYL2o49?}9ֈ4^Wk/ {=]{ɱ6$N5-4aw{# -/wzcHن&l MX6s^YHg(m!Ց ƔAM9Yiyw 3 gvP9  } R?]]<'BCu3{=[np(M8F;# /"!%i 1cg$:AkqP+/5{בݞghgAtn]A9:>X< rM\L󗯟U/L7h+AИ-h#IQ&B g"7jj(T##Be֖ya z{ ;v'ٔ<_̯0ʼb45o0$77n A8 vp7g}'1TP6.+mzwszD;@DɔҳkkMic92$0|kQeCؐ琞Z=AxFG'Zxo_zN3|^z2œm acf1DqbPz? >P/(%!'}5בr@Xb&y$,pD:f2Rʃp-AX/5^# gG ؎aB4A]2}.68Gٖۛe; Y,2l+gkTPF!UD@`bT@Td4bQ-<:_u w"Z9PɆyHFǤ"9̰4 \+!H&k/bpǒRDElx4ff])ʃIH4N9 IQJ#dL Ʒ/M;ZCAZjԂhuh"j Mdc ,#9lCE"4 :sO H$p1K/A/<`<`lmָdHZ :ٯ՟,`|'o'&b]ei\t9>Cviv*d\zh "C'HeٛnWV اlELFz:X",l`)"!\HA)c"\}x =OM*a@UiAUB\:$UX?\|֭Koe$hsIZ7W5&^tQ<:@>\j f8 _6*T6*&~Ǚ/}]_}8DA|+0.;uCj!>ksט঺]Sv&g=Y)}[`R-O7wo¥0WXl.5!iL+0l`RSNVnD9Jfr׾l&]RqAبㅋ_PL- DE=`22%"Q$Y(Jxhg6Lϼd//JnybL_FA)옑FnATZE'|{ȥN]ǐomDc ĒlfXp[[(ċ}v&  խJQuhݕI'5uzG_Vtwq xP K eQҏp2'YI Cw^ׅٯeY Q\+}n@G2g8cVJ N).rANшeE/i6Jx8JC4iPʚ<'Lqg7tO;ߝ:[´MͰ jY[LZv ϵ9ߖxdҵϼ7X PvNKg,Y匶Li3탐 f`ѯ(ML}=»Ve$vK7˂NѲ]|wUd?)wR [8` ab3~x];;P?_L;gOsN'ۺp%foJQ*N gUR"G/AΕ--r/M$SX̵Q*g i[8uWRL6X߁0* `,RG+":\HR#Z: <NcʠZ6RfM^C->ZGK!3jg>OLj9+j$MgZ{ dLZrN2 3=ϗD^e[D;H邅K鋹`IB;vi*\AxtB.w)1NQ iU@ZFI Z/Xh };"5]'Aޕ'܀XGYΈb&G˸6[[MYI%#k\W,4*ˆ^Bxd!R')#"b1h#2 ~ucqKJZ<'?r9~,zg!ֳ&ͺ}i6^$4*k:Xm 4wFm6*GT KIw`iמ1SјX>{j kWWL!yv!3lb/IW*ԧ,Û^ibLϸd,Zrw'+{}6r" Ǡ) Ew^&Q+R346>)mdDkeB}%58.K{wmI 2RɮM 5~Zul{1 S`'e }HM~U|; bܙ%> Ww wyg tW!\ =QrG}K|dۋmyoߡ!cQ(n[EĖPƽuH ÌМ]WK|ܵ)zйxsJyBXDO|;~jNImB~pów/LzR)O)R)O)fJ>6e`T]`qRL@In#v+gYѼz$_ X O6 |Z"ovh]J4A) x$ )[X1c2b=6MVHKDgWhP\Ok>hޖ!&T $]BtGE`.2L&*Ű)f_l";|KZ6unWm)lӚAZj~*Hyw7h-JZR>RL@Kn飶ěU`̙ K1t(]sf|=I*Jӻ.}_AndTPE# |4 j-$* %^F@HHL2:+^VIXBgSEs`L1(h L> 'H !UC=rl4\be$cZadPg($^2 YobA% 9Ij$f.#BZKcJIK#E X-:6Bp]py,m_˻x*9?ЯS)_R)_L{'vM(2q{E,j0?~Lx%T!>L_'̮|6ĔM_vX ߿EϭF5 eY5(6|i*ٲkamSٺ?̞9 LCrͳ,Yydcrs1E4^JJ8%Rb ƑNnmdU@;-G면Xz˟vNN媡In0SOj }ns^oM{|>rY^2Hf/`O0Wvֽ'c{lZXzWՅItQ2]N 4L";]$U]|.z-܊ t6}i6e4[mxU ˘ _[X9[e>(I\x}sl4rǛiK$LRW,Pj;{oXwmn[Bb%2*3ǡ0Pm.;BNDtNz/VqviQ@yQ(jڟ8˨ &ʨR R1!0z4HR~|vB䌂*gRbr叫!Pah >6M4҄htJ‹H9gy !#!&xLPD56r#j%su#ЉR\)shvQL`lb=t6}s8OiwZ/_?{jJ>}lޠ =LDma 9Eo Dp&r㩦rA52B7.{V%̝W)I8M V.إ|osƏe%;3!,R]z-,&9Nu. }=Q(_o/7T[{ n'^.%!pvGEO@0*2A@F}J=/Dl:u%²e wbRʛC(G7tG@a+`3rY{4sJ1=~PQO6d\#āoiC}&^-]Wxˊ r#iUD]~c~Kᓚ n-X_BKuWw׺.m]f^ϋ]\í{nxCW+߁w*a$U[:&\G w1֬_7vٽ@Q?؟cfdeZȦm8C,AHlF0xZ0J,)kq6qW>)\o#SdxsZ ܘ+U_wf6G?rD e ӯf$F|V"ʶR+737pI! qrcE(o&`|;) +r[WF'i-[u.3>lz(״l+T>^doPXJ0C DD'?~Mj(wx;4C cg!,)‹/0mB0=%US}: edTPv@Fl2hlp Riș;_ozuQ^}W$лHSIxTM}~mK<{ev³!"hy]d~]ɯEzEO|gv;MK8{do1+7lKǡ]56gpcv7IM€t'YWw׌vB%&˿%| 1Hr _`b#DtcW`iVEߣV ztq~+j9֌ՎG+bI*b{]z@DQ,19JEi=Q0!S;(HǬQFYJy>%kYǁ ҜϟQ=f7@@dr0EMi^?nz[Дw,_x;r.,2l+gkTPF!UD`bT@TPF,jE ^r7M4|Oq#)TD2'TµK`@FDMXuGG#`dl8(I(d@AS,X=w#sS.M:@7(! D*ù> 60%[Aj2R.a0px g{ `@H-@"ф^zIzqk}) |]*',J9o~S sjZuFG&[W`>r9Uy:c_ WowŃ7ꃷM`Msmp\[kpeYVo0zj ۆ!:LYރ0 V0bQqlVt1z049ھmdۨms\c2uTB-#i`ĥo!,s|qѿV;1  F n߿!~ǻ7?D]ހ,&᷃Ї睇w54WMahƻu;+QSnb3;ps[ox3.5{;IIПqvəd_϶]EA_WD9U*D f:#8}OtvՒ'o!͟"6IDZ Ec0bkU( Jx7IOmu^^qx7IUN3t>?̇@ JAff 03(|}OBJDKX˸:`2|3okzdF%Q&YpBLZj\X@–UGFFS SF)a8)C` Z[r ZR[Åi'ɋ*E:&MB7[II(ƙ% `J6T+%5 mA9Zu<-x`(3*h2 `0֚|jDLHͻ$ wZbASsT1B"ka 8( E"(i1IE,7E{,$DX.uΝ{ϙ**3tP>q D3e.pW"P= ^m..9az*MV^2^!Si115 (_x@>0`-S!Mť lgdFN 噴UAŒ5P^{.&7W EQTcm̭ʤ7)CQAC\pl(ZIXgDg19ULA^% V`%B%L[qc֠\cdNm8EB$zP2 dܕ 0BDJC{c['-t:K& U RB EsTeJx̆6 4Vw*4v@:9.g !./50g JC+$d BvH |LcxFc1gG̨ naK$-0&lJ80)ЙVNecX"6VA[**dY){h*5c@(qH(NN0ċB(X5gHQ_Q@x[.HP&!EސgW&[`R"qߣ.%eZSQ$f^,.z~Պ!_U0Vr!FBI1UV  ѩ-C^b$Vc+:c"& Lg!!Xwa#eۊYUk1 E%eA;L'A<>ô0 R6ά(}rG**Z@L.H$􀁩p1;`Kj,(tD\@JPE"&LԴPyE,}R(T)VbmLmp_ 97VTlh .g,XN~T} yrx*fE <-`9%^VB0NFO CLm]LՐ"6mZ[V,(c#P. {^s@ QMހwK }Y]!jUW@ @v3 hcBQhv)0ְcFX8 9@Fg@!t9hs^u@XK4"hQip\p@Zi298kkj#:& dŊBi4@?x2ZF"v7H-28\ɂc,p0",0Ljp#΢bt",U@k6֜ß?Hcy[!ճpgi&0 ֨H `f)4ͫA*Um-=XuX@KPF039l@E}׌_NV%Tam \n F7 znu8i1yߦmRrsMc&o*`MW-d0L&Sas`*?|w`)jY:Z5$Zr9% on0l}a雏 3| g&%<%*`rH|( Z/5HNJTt A(P`AHP#F' \ X 7nkdXllFVG+lE]("\x>W+f)oQG]u6!۱j0 .TmBF8RŢ4xT##>j &z i,~ aQ0'#80`c 3725YH>9.J3)%&}*` PMIB1X>Y#9kP9%yіaL mMg TfW>F e֏:rfF8ae%3=f@D2/Lčr KFx>*̠!GXc(18c$@+Kn LacvNQc6DU˥71.ńE@ZĔ&0"pŒc8.Pklץ5HDXd3{&GA׳:>\%~Uuk32|8^9yX#Tqk]T M(jҭWlv|}Q;Ft\R޾iEA:.rMU;uŅ.ܓ\9zNo}Wk~m|M >ڽr@KBG 0L?@>QW1Bަ*V16oGf yr=.;!=rǧMe4Ж,OiѡN:ZA蝰 ?TIБ RȿG}8`2{ӇZpJsáuO?zu0Y=.T- ?_m $$&7IIMoxě$$&7IIMoxě$$&7IIMoxě$$&7IIMoxě$$&7IIMoxě$$&7IIM+6zDL| 7"&X+ݓ1JIDL@N @"%)H DJ R@"%)H DJ R@"%)H DJ R@"%)H DJ R@"%)H DJ R@"%)H |@Ρ 51(`e{?y%H(zJԏ֛wִ}gեniԳy;BZ]AWt]v*1 Mlqԣv_#xKz|<9z/' _s㵺[H,~[~x>  rC1H>ޕ$uC?VeV*x^dY9]+ )ϸʚ|eoZ8EXmtqtMvvv|q\8*gǹ/*Lz QMx_Xz[*drpr\7wWcfn6?Z'f}zF܆}=ܱW{[-{tyem}.z[F5goo3_4X0`f{s#_w3KPynQoL D$&7IMo}D$&7IMo}D$&7IMo}D$&7IMo}D$&7IMo}D$&7IMg+ṽ[lG}lxoX EJN[v=OK=.i9?N ,&'^Oo[8<vkҰ,twIs)d‹tA!QJ/~K_"HD%/~K_"HD%/~K_"HD%/~K_"HD%/~K_"HD%/~K_"3/=Zz+~k_;s;ltEp}[`Gns}=6|r+Eng4VMfvqn50Wև<;>}a>.I]P]ZuJ8yQX~ /b>Qߝcc\$|ogmG蓛nIUb!}[`_1'UZ?({j E;<9~h=t;7}]8[swk| u( izZ3TBJp2.LnX H-8ko1{ Lи~mopE ^ n _]7r=]^뷻7QU;>@rj.)|`vOSe[ۿZ=!{6>R:j|hGJm19Ƀ/hqC 1+U]t.j25muȘ>ېe֤=s^o(mտf;yGsP-|Y|~r!ukK>"~\^V2W9(*._wr"^P謓BI9v&]|㦒J0mժYSw;vruV2 )5:;B1"/)ܞjaN66L[= V%E 6̵i31x.jD8lz=~lnV|h:ҝbk3/3M,H_FZneP>Um=& E &e>G%YӺKocŻ#=hB.j/̘ qS&$˜&dj$ЎGL;X4kIB%1LEi}tHa%R!av"1a RX-ߜwpbV1YNf{Fg¿7,=rae%PdyY{Ka984KTѩ֚ZXKY;ݥPfSQYfHxV4٤VTm~N$pΔ v`N BƦW$ xS9ʒ\VNNŜC_+oNON-GfJt9١.(:Zγ6 f82  jsM y,<<",wqd{͈;!,M۟oS` e$j9amoilR&JfN5U9K`39vCI}׃q'Z϶&r-iƉ$s}שׯjtdE)uEz}цNm{z]~5e-[v>yxs絭Es-CJr9o`,lf;q ^gkN?i9]_e75\o]N1TNJ˄hsYuu:10-M+خ:J$",tlut!RJD`B5[_1yx@Z >-Δcfͫ EXIb)A܈C$ 1% s.<[gJ{jd.[lͦuOs Pb9z`-MߦlCtӣiD 2"fjjk2sjcLDywؔfQ2#SVe (WKMJrX4 "zqyֲFΖn*HӑXHHϽg18 XER8MREAFARNZ:-*_-W%c;-\Ƣ DX1`t* ssi 㺸_M ^A"A ˭."IA帝7%bS:l>ZGK!z05ԂvtTŃiZ+IwLqo MbY  >znLϋ秒^y'|V B\ ʥ$W %Q詣\f;D`A$J~z~zga|H-eEsXtfIX:J[S]NFXŨb8 =e ~|:%_NׯSunFT11O| c*7S&㨂eJce1;|.c ]u>;2wS)bXmòVLq9fy0;~_m dkKH1H _QeCGBGS^ι9xX+a_f3L,Ŭzv\ sPGQ| \`psZk?<2Ad<'{yRAqTEbrr4bQcgrpTxGDJNcRɜRafXTJ〃A$ B{Q?cjN SCT7(i*RV`PM2[Pa #^™"XQFYεripƹ ;HcF^E-DWE8Z8q_W#޸_¸L{G|%rWFJLJiskЩUDA2 }|},3:Jq'+00E]HA)ŒF.Fc.BOw%P0 *:m1J(XK On]w]ot~_&Uν)7E8܅.wT0)\۽ U+rQ|x6aΉ[8>b5,xœ֝yr@ïoՓƷh*Ĕ;ۏmTvI<[~B #Wr$Wt4 ia3A?%1O{,h6f;^79ڼu2ɦQ抹d:(]r9ϟC~ vҶq/pgP݂rMybתnzq[`ǟ`;O?J_?O1Q?8kDP;л񗭇547*Ќ9zW603lN|JcGn)@~C;@w 'R ٩Y䎟 b~WrvGWwQRe4lJȗt3#Ay=Y/o e Д-싵Ѣ6|߄D,SˀQ{"1gsɵ*oI^ r]zY)}lOq [ ^$zE`o!#X-;/A&~<XGCǠ*n[EĖPƽuH ÌМm<̓M#]R,1}>6R_7(?XNJE9WQ6z$mgtK&lnIb-IS%IT oI- lzV_10.Ĩh#ԍcCU+] h ʀZ ͦLiv4~'8 n40,loʪXcaxS6OHa7dnr&bQЦϷ3酫_G5M<}&s/LG eZ30{-#chnD4|TwծN7Z+fHb <*и#' q؅I0R ˽)7cg; JoV:hT\.gzJ`4a՝x~.y{Yh-J`R>RL@K^j飶ěU`̙ [9tm)]S)r5I2fxiՋ_5(L7ޟ{U420D%} yIEt$LEL0#'U{A|ʾ! \HlJS jj*+@u4H&$!C=riZuŒe,10BB2PPg($^2 YobA% 9Pg H!c N,iYҲSZ Іب8ܕ[VNy#f[JḿH DbHI>q/Rw>nB?ڱB f+IswB?_B?7E}_'Y_MڍJ2Gu[/a6s/'nPAM'#vFYkʦbٺƧgٞeYnܜ`"{&CD{fhp%J7bi1g ƑVmpgY/avJтIL<-L0c1e<,,+$3 f e%A.,XY(,k+%c(yy`j`mO6tn"EGjjsn&-mqIFϻ(]v%xVm`M2q$-uqx Yc~} EqR,V9X;46f30X3ng긿IQխ6igo'u)$fT23Oke/vU:цvj_Mm/-67E%I w o#o) rh >y?%3?$ΑGg\)9kn]7VBu}-n¸5qm䅳IQA)3ZXmDc2:+iW[NyC*:Yvg^jgp_۔攵mY3w.TV}~fMHD3Jy6, 7B7z%sTn?km#J#_;m]\&w3pI_6x#KI9~nInzYjYVA"XRE)QnHyTLyiH헳_i9f^[2c+zlte\tW 4o5iQa8!bO?Y]Fe^<S:L%کSQ͆J`4IR y\<|٤3tw[+OQ#nzs Jp"-9#4Hd鵦VDC>shp=Br9zx%s-|d_|tUYGk6d5=`f|`AA6IT 6' ѓg6Y=+.sY|oN7*^12lJl}Ob@2lV ";SM 0XW?43LJUG`gĉ@e<-臓]qj7}ϱ޸|)oSwTn=^naŝl!@r3q_-f5u 5Tǖzty􀗵9[^2% x|V=99,[ 0ԉMFJΈk=||ZWPFW!֎QnlAғ(J.)mLFGmQ{UV*+Շ.~Et5kpXS` '1nuo&%.$*2!V u hDIQ'[B(C$e}I3Y@)DK=Bsi͑df"ڷ|9Y4+7z>r\2>y}DD-j. ܓ*0aɵUKh@>|\ZAPKR9h!1(Ϟ|/b!V A2 M(?|'K9;zDՍYu_Z?rĦ57;qo|J nrt'` G o%9h L*/<(jJm&r#xېNt $ilJEV{lq`,3J瓤j%I3SԱ'if*4_aԄ ذ/B`etВ;"JK*'{.A_~B[I\UCXF`A[fs A !N$)zAGr0Z-*<*Yz?imlLO5MON͛nωpGJF9;d+lCG'H)˓s}ivYyhMw֗]_9UӺݺyAw/o|blr!--ootu&ݘ7ϱ5 fwQ~{rl𧯮nQuiw'j֗əPsJuCceuøCXy_YKz8,' .MB".nϽ)%2*#zOq[YOa~͖{u2)#bIsA43nDJq䮏BȀl6jl{L@ӟxXia 3bԚ@*#K濰1Phn ’{.mC ۚ^LɅǑ\~F9}nzp~4kzWBQpp&D'u sTƉ:W*ᵤU^!y&4J(%حUi3ҴՎ1m0  e^i SUSS8'4u,ȄUA= AP1Ny'2WQaeTIT6ʩ]NЦ;V "}$V(OoD<˕x*'t™2QSQTnX-\jE3"sE"ϗnF>q4!B9Ӆ߆p-8cEp&Rjj)-@ ؝}x *R%ϕ{j'mpUFM-2BLdP\@c 2kmnX̬l+vIQEi^N4 bu:e.+STkHXJӹb,6NA}1?7*Ic('>֍]Dž/(>{˯>w?wumKGSdݪ%U[[u/M)"mYN忧1$e҈8IU,s3=t4ppZ˿GT4L~os/p6w?ηJuk3S/|:1#XXs>=OCn@ZO~J2/>]V~59#>)+l6[DoտmZj㹷*}>x3#_ w (_<#P#&$境\QS,򅈠HSrHQdAs s%r&$gV1183zP18 8^Xq:ģƂ LLJR8 %Yj:,wP ; bi{ x@O0зwqU[n{żܵL/..??/uCĥ>^v ~uՓBu?/KsEXk;OOHJuA[_<N61P 6zXK&Q괴W4BJϮ$2:XC{P[/`N:SWJ,4GKHA$e9G #zJX/6? _1%Θ1 &0]>tтNK65e( i|QDQ#e֖SE8@D"8Q1I؜}3 uWU|_Z殬y~ufuEןmvdžCu3b]Y~ߚʌ{dԋM>.d1J9񻟤;EAjF"!i:c45bHL'i%a&4'hTRŸQVQ҈hf쎯,=3Gnwr8xM\!%I;tE@yPe*LbbFo V9s0E" –TvTƊ2l-h@90rGl 9Ԡl r:KO&C(GffΎ%bH5_oV'.ο,ngZ.$Q>?.Y^QJIBGm,9bb$2Bd!kG i̚XUjzpBiJ1͕d` hRK2/9%{eՁ"݃Yo..eh[&Β{$hacN=!!4ܦ۴rVnmZM+i6&ѴrVnM+i6ܦ۴rVnmZM+i6ܦ۴rVnmZM+i6ܦ۴rVn/ $x7y9|~Nj?q{|eq9q#}-n&NɷN ր"ZŷH0ٺy!ZiWhUUX9a A!_|e z4TPJ"0jHBA,A!IFUeIhl(Ix X4Jd G3g|L*Է)=φ~Tt~ Rgے2ߠQ;^q,ALë«4lp8OQE{R9Z ן>O?]ElmKc7L&EMc]A}`|n v'UOWynӞ]1rU꼂y]Yb|7 FGy"^=$k]tr9[S>z9S##J,UDԅd(1huQ'v6l=Oc+o=x+ן0}8~=m)>6"+\DL) U-jɅ0,Sq='S6k堒Lrl(:'|[xm,shpկ!n wwhۣgjl5Ut:ka qXz6y5K.mRM.f8~* F{UCu>h"&׆ E ,0rFF]!*`CONuiru";_&n\&+Rؑq4sGQAbh=㹱Ќ M/ o-6\W^=l`Qu{,-n ]L.'篿D{šRBǨM[,D + E{.R凓cd-'80Vu!Eo$p:0@LIgDY$6 Ԁ4ڻٽ뗖E8\?ٔA٬X9x7ȇ㻼,uҀy}&q [D:}5ǫbU/jO=2]~_ry˚49J\.$U C %sZv% $Jy*dHj |sM n-1KdF`rUk_:ðE|[~xn4k~y<~xfs昇s̺̍y_;c5xe>1|^_;0 t]G1:n3ICذF3x\K܏XN~u>m:rͻܓ:A_ؤ%nhYۻyW٢1pȍV%wƅJku5b tsײVch9Bt6PԬ84q2- HG-VZ*͛Z YE')s4 <6g7bB*c=Z8u)qaǪod {zݨ{ JPn5؊ʭ:q*BZ.V!P( B83~nyww>^ P.B\ zɝ4ȋH};i^F*rw6YBt`JM)qqRceb G7wN q)p & dC0(H*u5 9w͜&gULbrn_w=TG62VK?)K m÷+ Yu//`p#.!Ȉ|dt3I ZI0BtŀLVQ}Eؘ8'CR)u@G bGMfUÆHe!#h>ir< *ꬲ6!YD-Ut)Elb*eͮ-`oCnBꪖz#l"c$y$**HL+JUP I"t\V>ۯn|>VSAmĒ,$ٛ@3rc覲5Ka/||*Rxm]On7#Am0ӁBÔժiӳ9Wk{ggw͍98?濦M/dj=4qx=nǣ>{>0?P3Eדv5pxQ`atQ!yAAgzɑ_1i& Q@?u/ 2<ݥ\\ߠ[>tXr6ՙd `hgK9jeShrA l;'tP3R(˟2ak뎂Wɶ+9POGTL2ɲA(sg'UD @"+= mnJFln&EԼ3;$7v>Ubs2w MEaA3嬄a/x@$2tQzt6Ϟf¾d:ok>ts(F"ޫڝ Jkܱޫ= QVCt=>y6szpE k QlIuW.@:d Z(B-;2R.P eH!E^ i bE*ɸ.- J>I`%*ZQIιXT (]fwiBw,19s[)E+rHQt?_餏8b!XzOG#F#h_mmZ_|Vfo`$%U#% ".EA::=e/^*^3DYe BrHH$D0BzJJG\mHMl('\prNAK`C u %I_'ؠLcxL jG, %k^[ xgevM׳Ux@rhӬUW >!7?T:1qZpEs~x]6ԝ~uMgKy+}M7*Z3mtssui͝UFi5 tҳUqdҳUi5iuilS2WU`T'c{%/\Uil*w_ hr$OȨp3 89 oM&<vRulJx=_Ф${&8U`s:*5bYZ'>oR-iGќbQ1WU\Ib`\U)hz;*N$I酳B6y D@We{.hm})=21 .#e!)JEY |1F{\7! $X|oaq|*Lsc܅7߱yϩ~ݱ˻?!}\Bj͚QߎO#TڀAm\@DicLGFODllIZ"iaOÈ튤GAJQ A IHR&8%(J9b{XHlKLXxEI:z6[TibQL&O!6`0:-kϾfrPX\?j6ؚ?d%[lȔֵJ; ӂsy,VjVVG-LUAc} fAFȖ@"cC*Dd7[[Q=l>-^z{pRc5;B5H*JHVR)&i!n:u [r|kZiRNZv sKaIYTVqVzҁJd'\/J+3yG÷Y;;O%v뙚&Şs0 y.Ujp0`PA7^\T8ѭQ }ٓKPX2]3h9׿ z#oL IJhpQ+\@/ 2E*^|V;%ap75c'=z_' gpVW0W ig޳Ogso}=>fzQʽ$B^A+|y 0/jw6Q-n>rj질l 5W)#Ϻ%zu~`bJq1ēpaZ27]Z,;u)]'ԋְ*p^ds!-l\Y]YA{[=n;SDw"uyD{zX10c;a|,qDu[;~]}kEQ%Q HCwqztʮqoXlŋN[XW!Z}mf#7_Cղ E)o{o@/;KmŇuq9oQ1dHC!jL>P䒔9Pl3s`q6#&||s,aƩ'.i V5PO=վ%nH H`82q=l I 0%\^&%*b_P%\~#kH׼"\|9~X|}KbThKEtoӞLG>U_Rdl0g„6"8aQzqdz[D[ZfN!$=8'e:$G^jşb JF)XIk|p[m.vy[2H%%BT\_r翲QOyzv<+}\`٫\~^V9iHhylI? ]ywk%3 FŶw W9vw!,A;CZ|1pt~us[gظ</@dJ9 D&XX }[T4nUhٓ/eF*E))7VZ ʄ_VG%cN2dB 1:Dط$ b-MhQ($\,*VYcֻY Se->n=٩=P%4Jk|hKD7>^W/B j`s¸Ng];#LzԾArPO55sTLBZdy|賓D" aٕЄ6qT<~Զy۱w#v㾁X}KD2 H!ԦPk JrVB0 <bOrꕷϠ|4$KX}́Zz|>y!lgŧI5DF4OK:d bhRM8椚%K{$iKL BRHW/HZ(B}p*ɸ~.7[>&ۣź".uUھ1jN'$rVaj9K0sS|@ 3OsR?\ux l^v) 9mee/{xyR&*\@:Fb$!ҋTrT:2o_ƶՋm:  NR)h r}$R"$K0i /aYO/Y-/4鳂2*y* Ri*]w+Zde1O,Nv3ϡR[w 3leh>ݲvl]5Ozo"Լ煖h8\;yߨduCݥZxXǟ7]@aoBnnZϟ]xno~ imR(vϗ\q'V·¾;U/^G @l%g+{v]mBl8(TF*RXj`dR"jh}Qj(B][z'-eH$EI9;'P&3?ll7gǕ)~3\sH߳]JrIdw(.Cۛabބ-눍euZL!XIJ@2%59&lAmQ"Sh9K左!8:x]j,Æ,ɒH$XWSkhBIdMZ%*@`c š Uzmk4& ܄.Cu(L£u!JW| _!'ᔳB*'_0IZ:.)֫߯qq+*/1J*Y:M%I&J1tS :)Q Nutlc.+uMo5}V[oh;;@J"mfݎn %u]QYuDꈟ)=_'ɮ6K:MǧNik=g}´'s-YJ{HuH%J'T{".JR*-'ծRjjR9>XK>37_4\+]r#EWR !G}0GOy9`L/Z[^rHcX|)$~c 0C!izftH6FZYE҅Yd 2E*2k,]& }̼8;/5߾ ba;n!s#S?^NlTr.e( wmI_!eoq/ X'XఆOgTHʲrgH8$%ER%pWҥWΣ| CมHPRAh(, Wl"9P*"VG#$K3J1& gakg9pvhM%XJܦr 'G32S˲:0PbZiA,ЎJϘv1=%,vje͇?\n^j=D3:BNRp(ӂBEBYOڈ۠g|ƊLl5a!|~@:k.@5 2@FhpI*2XQ0\Pb$cW&qlC ݖ^ӯsmK%RQMWjG *d>:݅CNhjOm}8rϝsݩn.Ftsyr<0[BRΈ߻^\Vk9:Aכq8z7[^ԌWjG|aX0L3,PhCJ+W1~nУ1WeddI֍Z7W* eQG#Ne$,}6Q R'SzB;->T<~DzX_]%_P}xOO_(_˧\q9&[`YGCtCSFǛ5l0tԬ' 49q9)ٮmq;r>W9<N&Z N~F1"H=>(P>'1դ/l`_rUIL91#VDS)ZJFte/G[ ކ+J=/7UJhhMK69$'ՐJAɲ# v1uݵW{uT6m\{A'JgOGӬ*@/=)MaE2:j(יK[Ga9'*STY@ SP(?\I4s TBWxY].0Ilql m_"TY۱ ZWP R'ӨpANߡ9lx[%_]| ?`6sK /)Ats 9/d-ܓˍ;A pO&2e%$*MIJR$MhK3 2(2з Ild[5 ,!r"ݗJ^s>g/s4eh ZF,q$ہi3/]^> L Sbxb%vA21TH<>"jcyYGBPƹIB\$2A #:j>1eJ*k>Jc䬯^A[)x1BμNXeUV\z_pJ?SI`X8)bQtޝۂޟ ocH) #IDDRlnZxrq-<9?Z5|-CpCAkE5d* 8FQ(he&PDUZX?u o!'6W=޺xyUp^F*Qɻ.t}Y|z lz԰mےǼ&+èb1J&ijb (w[ %QMnpR9&\gZIB\IBT~3Iަ aHA۴ G">T:B" "%,XK3^ȠlD0Fsٲ@a(,(-Gp(7 -U56RD7Ezu[h}JK-yx-y{xxWΫV׷yr\ʰ\[\aK˹RZy}3>(jQ瑯E@!`&GLh1jFU !9AXtH2T=gO^>v(|1hCbTrKPI5#gf\R BѰ..ܩ.Q6Ɏ$U뛒F~dW7/Q~a@(& MBbHc&CI\f[Irg#9YV\*/I{DՃ)wpJ795Üqy0M%jmٰ֖n*6nQ}T;-(ȐZўG&D%L#r u~`]H"FuT6#g>,k~g|hQ5U[HЎHhF$AKY༆(Gh2P* >Ahk_svԮVZI;xI;j ہPWk"FxU8oqː @=1ˬ@Mg$㶍k$a,NCWY!5tR}4i?g8 >(3A^׫[Œ1`)kLHڅhbfÁDVHu mQ{bQ)PN[q{t´E}c1.(0sVnRnĥTL$А+fgE3),C)qT9* ]j=kBn Ȣwj"Gj.6B/s2"YՏPO^-&T[|ߊ٨ÑD2[ls`YZ#W{V(^X(CL<S Ko{5tQOR]@1lhmBex&V[pnNxnIY%UaaЊP^ Geϥ(' &럎Gkt-3'7I4f&wg5AC=a3<1G*6U`d{n@x2`'NPinD4NZ\5Yuyoj( E oC|o@.*gu iI2";j! ʾ *)MAJ3Ѵ)=Y`Oh5^fYtc,{8 _jwjJޛ3>`ܜi*OVp]sƧQ93'4gmszU[~#q\|&L; $Rp]ؐWm 6Mhp*=9c)#;l-RS1dDZS 1PKwQß3&H|2f:jheF,_ivע b7Ȑ5xzv<6.rm$^}n~n5aEPdy ysp1?zᮬٖ{[unnwna'0;C1^R%P;" $Vx0]tU,3:^DZXªjZF, Gl-R9ӥ߆R1 ਲ਼ M. Z^8weH2;NA2H~L` )R^Q$:my%(NG3Nw;hBjaA=܃o;g7?<33?HiJd$OFnX3Ȼ Zu]Nֹ\DDrQ%hPV)HK×JZH'CP,Xkd"HG9lcj0E81J!9b!Jx+Ti#F%tsCBZTߎkpNJxaSᶍ׳OfB].N&I:Վ뼬'P. @LRYhbNjNK ?`x ,^6) 9m!w𲃗Ux)x fE.tJHB#Tr:2o_vkXZ{4, P)hP r}ZgED* | ' iZfizW=v۬ x>t^5]IJ]s+z8x0?V'љfO7[̳^y''̃smdp9hovL>uƚ_p 6@AӥZuU,Z]ʹCw}܌PϟN?[mu3`Q1p+5c3mpfo>ôB6>ݩj:RcT.RIz|-nkil-|-*NpdxRa)q52 fD mO8J5Pb_DaDxK2a1()gSip sв9+5;VlgꜸϥziTJN#4W~ ^ \o͛^7[ICuZ@t,tIi1` j̭0!*t3H)[oM(wA)̶2lD"y5r4L݅ h-$8TY SQkQe@IA 6 PJ70V(_hf"tB/hp9[Yx.DpG9भn%rN:[/$sXdݢj0~-WcZPh Ci* (iȍс0F7/ :1euyWWm}V0ug2*٣̒`{YkdԒ̒L% e|%ǯ>Yp>v88˳|^ܳKdxXp%kHM@M(CԬf9 ,w'lgHg‡\m*'*22t)g. R;%﮸iQ77>tB}AKȬw% t;4S޹'̩S'm+gsQYp^7Ent!ESR !LL]I犳A&A-m4VR"T'A(%1 S R|+V򥮑hX?E )3)ɒd2{2$L"Y<2\\>y &x^]!w<؆[poSF'g֪i>;~^*QJBT&U!2BdtV:(GY(Xc65-)\#\d2").F[>8>\Zn&ؾqnEc $p"S"t.; XW/0* ]vmX*ʤ(ZT%j@{]1lu͐=3B6 +Xmd [ 8}ǹ'.3̺ƍěLqAZ).֤>ŨRyX+w,˿ ص ݼ KB-j|̋u1P*btM#HzVCXSiz͝2Ab0`l;"k`Vl@ko^C>G{UUcjԵ:1s5ۛ{Y!OpR ʬrzX9މ'~sojϪOAHRApuz&V %-EھVcEJȃ qplRy8r0kU&}߻1ӽMue|㇫E`v1s?dn]킑]OYXYړ{:_֍\ލtKY\,iSgb/^yz~:Z58|Xlz]6V@9uX9Gj/Xޠ8 & vF, n?&'{℗X}폿T?/~y/?'Qr mL7"6;] muͺVNѵ6G=5!/>ce4η1)$O~i\U\S/>d3OKN8mbL_nSErRqLUE'H&wZJ 0wJ֭ 3׌.Z7f[ɳ`(>:ĻƼ1D :RXeDAMuZ9ņwic aiGx0p=y򆛸_z%TO=}A:xTZphRcֺ4k؄|FͰ*]`LWyKK92(xAA8%CƔbEXRсl Zlu`[If5ժ#A:;ɿ@pIcq&(ZI0N-^H2?qm7E{X<<ÇSpL 3~0~_cIeYۣVLfo2ZUrKF+@zFJ4]FWJiH\UqU]H#:K쾸T"t+]M4`/ f_UV]WJMz 5 O q"ЮJRzʠbG\Ur%틸ԢuqU\H߉#x u=ӥLOzo s =G?~]r0n4i&*jo*ftv6(:U%mSqk{#*foPe7WL% ݉W(>J%kF\UjiJWR\)\7+.ʟ'z_G esf0cO .cQQM!G-8*},YIs'GLRsC[H4J@rNgfYb)0xGL5`]㯃|&F/ɏƧ c_L˰ .as~ <ݛ卧tӕK?Np1 If68Σ34*mF)4Si %E:^TI ikݓtUnɲ{ ]\FͶy("|ӋQ]abx^exJ x!MA68h &Ko+pVo;pVo-pfT H%STX  (!!bt2Vg"HXSs1Ԁ=*L\RD@ιXdKkPE89ޖB/!x3y1ZZmWDyOBlZ3޿ vЧMCkB:hV fDpS x'zڸ(M1 ]Dlly UŒ=U]܉ IB0hBb1&QE^+ƃ~m FU(HKJʡw2)d &4y pkoɨ2l5rL.!o,0O7xn%[:hďu_Erm%~]-T `&vEUx*[ǫ:d@t>zʤ=tJyڟH>ty[,uw2yu?|i޴Jڨ{='wms9, P)hP!Ed)-J_' AGf/KZrw+;e!79tNx&דY,P@@R{I?{Hnނy"&M2n)E?H{}۷y'gm0}Ž= dd͝mVjXj K$ڕ,U"!eA.DqfrE, ՃDv dI s>[3nR qSu U _T(*6.$ol˼*fכ'ޠhW,ؤ]G֤nr(P>Ŭ9ώ Ylu1Ⱥh$ AHQԦ&j.L. t.ک>٭d֜ˣ*T;ue{[8> 9(%c.eƓqF-tR?L\Uf*!eȁ(ѐIȢ(,QA`)|HY%ɩF}e}XuKѫaE#V㏧jDSY#^#q?`m 7F ɀEτ"Qs# (B]m95"\%OՋ^^Q+7uBBGպ;Kobc }=o&է(O )]qDоNY>m.B$b c|ٯ6Sc9"Nd4"PqԌ&ggE!u|l.&_oZ1{p|iQ-< d R@oI>\d %!d3; Bg.z+td֓v27m0/g4 >ߢ7dҜ.wW.ݮ2^~2;>fMrQ 7)Op%cP$JRhII&pH~ܔBbyJvXy']2B6/L"n1AV{GЪ3jE%+e:g;,lh/lx45BD 뤷j}-fnaaʻ7eEMJ @. &MهmE JZjVi"DFkes .V=2b엘3y8<2x㭃5^Wu=O?LSjj͕ECbL㨕0GUY=+k?;ɠ{fif=+V>6L4>dр ]%fAf*g^$ ?.\IWwJ]κO=wqFf YHDdi.ʙ$8ȹR:,T>l];G]"Efdd'퍓s<&rB)NTBJ$<&+Brt=-0lw=T  sme, #=i0e1bcCj:άcy݋fdD7*Z-B0TV(cGB0[6HV(glɸRmhjy KH ܊ QT$DD*H.10cA2,R%h!)dlkJ P6kWƆmJDtL k L>čNr2F31X6FxƤҚ;-8IK r#MVrh/ ]y/cG$cWNZ~߶}]Vٴ w/{aղz8pRПiq̝k9OB`ҠkfW\!(!:&  HGDkWxm{~l>hp9/8O*3i9ă{i-籄 c<ˀ}M2ED&(7 6<:ͮ+NzfS*H{mzۍ8V&EsfԮ_H'ls7NߦIn}:OɗqG2Hwq醲^aӍɛeR%+Z3/ØڰK(ﳔ&aL*(b&՝rjBS I:L&H͙0 J)SJy[OY~[s6P<{WjDN:KdH!F\cѣ"GgR99V%xURGZ0`/35WN\;~WW¢c}.`3! $gtKc^;LVޅ-,.v7y:ߵ2Ϲct6طs [g}!e3gDXVQh4N4@\c<(}@{E`_#yS2rM:e,sQP^hd&hx6(u*ŅGRXV*l3d8U,6fB)-6L6b5rvۈ;Etkoe6k[^vcqƂܑ@l=75y SJtus^tQj\+A;.#UTY0MA ޼޼ll0Bdm,mfsAyRRfHKP*S 8F`Hg8V, dpHA^R#NHMTW.Fg4= :rx^d h\ EVs*dfXw[NUSj^^|7+[śQ3?]}kzR!n,xWfO0'zpgV`')\`z8J(dapϑ,iAn()&8㒉Kvrf\?EݜLg/NpŽoz N9XLVtC6c  'LO3ҝ~›oHۢ}mZBޝ..O/5Rrҧ$Պ\$?Q }+։sZ>UM0Lփyb?P?w~jop=;_\xnm IW\q s+ܮY{=jrG801 .9:BBpdר]ʕb5lbRF/X`Em>w93N5[{j;qxyNRG|~,?O\؏_~V? 0.}6 x9s`迾yെCK { S/&|q > k:øZr[?ӟ9۲KS6nWM*zi"b,?⢋ܧ +EjUFtGv*b?T_l]7$  `$ CasJ 2W} dH(Mފe+)w^=ڰyM+>c ` #[d$fAu2FH2gbh ynI/@z`rvI7+M>_抿pN4x~G>ήq"WE~=lۭt.ut(P5["؂z3l J5B;jDcfV;/N'#\rR>g*A,}g:U r |mTpMJ"Smԯd[ɢ#5)g1RS-xQ !t\Aɂc\+523kZtr7Pݢs^|i', o?i[wHX{@S ^xNL ߻@V՗l<C dMN9ډ1z̰c Vg2Do0z&RXR(d:{xH-U4[AZ %+,}cPIvKmrKi3Hd˩߯13_"43 tht+⥦0+Cʘt3ll4sh{[x`ןLwe)󅹙|4A1FTu:O9BAQ(Tcs1IsJU ݘi>Y؍8pʯ5~չ<"|6[a44?IOf":~Խ@JGuO/ŃIJHT::m!V uȶ&|r Ur4 m 2^qiAӂ>r||\vrw[%7^g2ֵn(e_DCIԘϽ%J{ɱf.Q DN)bxЄ^,bSq.`9VLF^l,&R||[7qfW5{.amRjJ~0jx`R8Sf l귳zyQTS@ oIwVzFd+[7#'3mu^Cց3xX_+'_LNAhl:K(^g_ww%;9ϹIy|NNzɸw9Oapp]:(WXS?r~s JPP}Cߋ)ѭs Yc f[ u }jݤLږ8lR%ޏ-ǻ~Fe>Tngyy#{P6 ɷ:M$'K"9Z13:'wbQZv6z); p^1\`B 2g\sܝe?q1D[drF^.)iWytGpm nQ'SƓK䒳S$*j=)<' X;9HFJҳIe*Q)Y W/MMU.}@Ճ7?㻿jZ,Xz`:PaUJ&ZD2ЋN# Ab4( K(Dh'/=zJ#UT1g0Io.S)CO Isi0eJ0LE=qa?1WM닟+% [31x)݁4|,Dgjҧ_Ӓ7b.Fs7}&Nŕiq\Pfio{2KO*A鶫Vh .fΖ,-C52䶴fI0p{0/aT^`by s`X(XXRZp-ӭ(Uyd ^ȋH=2BbɱX"Q3G ۀ !)sxJqTQܦb:2,"1rbrX+?RFQEG6r֗xBϠ!x1NVIwc{Pf3Z@s8+Ya0?Ӈhވ^eXT$Fw8S9LPPD>!ch=rm-ؾj9@PLsO!v(Fǔ"0 HFj h*ō8`1wahCJf/hl f#g)X)Y|; pr=ZB v ~>_nnsem=ԓ7`}ګGW`ͪ6J?%&f hZV"ק\禥[)U=;=@E'LG eZ30{-#chnD<Uzϩz3GlS^;0`h"`BGQdب1F}`E#g 'H (K>r8AK~pFn:MwOǶ(3#lEUύ014qP `(k (HR,18 #$$8pFBb%#17EXFx҄Ij$f.ejcS%ʌWqJZ\. fmPgCf5oލ'ԑkwlud"Q>n׆C)2e!YCFI(H4ڱB fEWh d޵Q/5 =($Nh+z Vqp4rW٣Gש*K3|>[qc*yyanI rH#+CWjʈ|1h#2:[:)v^7nOۭ4Խfڲ+lϓ[U`XM+7')7XH>dhp%J7bi1g Ƒr|+ ~$* ަ1azN ,uRn0,V dLnƟ7=S^cx )qujP{1G(:܌\(8L0+&a/fhY[DH(򅆱N)bxЄކ?ʬ//ﰎϷkWګkk>೶lLޙ[]J[H&OX}&[*Nv$ 76[#M oK5SSхG?n7; ~LYoFa7r8;Q𞩝QIԽ& %eQ,}T,Ts̬ʷUeL] xP$ "?UnM m+^MlZ^O58׼8Jl~\WգoKש:X^~`Pӷ9fr3+t޽S~܌LQ'umg8^7vgG[Q~񯭗#,ڛ8C.>-nG_<9`b U`ha)1H[ܷ.9Thr`&GSJVGh4 !/~sHH ;&i \󈃊~ |Q::2d,^[3p?_~"zлc+?\-4Hye8Ac&TDM!(DnDqBGigGz"6H%\G/T"%cB)7#x92J1E8#aREbm}D6b6rۈBHS["m)R_{E[5܁ y$0Ȟ˷?} '"^de)$yٚyy%000"^ ̝F 18*hq Ӟ-Wh0X5oϺ5{ Kur,ֱtӦ; Ѧb{k?\bZH9ٻF$WJ褽MA>"_H;=ڙtV:m < 6kf"lc .B-TV9#2"W:6Q0|^\é+KMFM)yL3/OlldK;dYUQ`qBUTl SUS6S9Yª B` fp2wDIULXURp * T6ixЎՂHJ;OuNLNq1ʹJ&bI3^,*W(GW}Y<&8%% )тEn28jn(ДHɪK8Y07mlԙ_hdٮSѹwB)#})ѸBG\YRi%ꔢtQIETBv<]˦2"G5":dGD$'H`L{Q\(@T,UZ c7+IZ9sg>8sz&n|re7%׸9v iG(aO61ĩrTF}+7$”G||\߳qwOP #)%O\#y!Qs7TDwRA!l6\A䴳FFC4RXV[\.it6BLx-2¶(9sR=qNkKM.ꮌ嘢Nj|!dii22Nf_崱hZb+đY.FBUEߓ[o<j>D39PSʴԢW1Dā^H41뉗R|;^lw)3b| e9nI rIh #Z$NXQP+`NioHƮ!p-PoܛeLȔb,6.-RIrK%/Ύq9?ϟ;>~!e?~w 1M$H6 EYgnu+wMmko5װB dWW6-0>ٖ[1矋/{:`pz+Ov8]{Y0}F6? ?ޤ03iDf#f4#AJl'!$y|q'>Zp{9n[ r b:$2BHzGU 02F,[8b%TD@ҳ[2M8Iw;vW(/*$A(YRQV|Vm ̂ tٯv*H/-CD)袭$!-ɭq8ٴ|;5cuZ,jlHqԗU[wQ*qEM[:WZc/?uy.a0(WcqhlFp wW ׊lB '[#2@E\ej%lT*݉7(H+$UWH.-*S+馋LzJTL0W\IE\!lv%*b'ޠM+I)tkU&Wm  v +EA E 6l{L.lvjURN\Eqń͹3p!?E&Ks :ѫ_+D Ox_?|[ րy/g()-)d_ỳ=nM|1)´q1dr1U3Zub i(E*ֈL"Z zUN\EqQ`O'x? gٟ,'_ '%hR{T k_;`;\P'Q )mLFGmU~|Ptlގ-PX.Zi ڂeʁI(-UZ& fl R3 !ۛ~9&*N c.70Cix%8ɗ&24Z9r@dZƨuy/Xo;KP[³\{雝ьF;Qv1kՖnN@6DuJ]_W鯫U*u_W[]_W鯫U*uJ]_W鯫U*uJ]_W鯫U*uڪj%TIӗ2n^(vվn|)8ĕ j]Qm$@Jqʚ%x!F8=8h;bVY1\cy֪#Wh e3klٌˆH qDa1F1uHDRժL=!&wygLDգ5OqIJ/MK-b_”G>y ]lKI\!75 wKb^9 $jjNY*B9dc=+h"9P*"VG#$K3J1& Ga[ȹY85Nڶ>*vc=w wYJ3dNгG}4\iGq\^(-nU&sJ@]R<P|UŅ4{j-k\OعBhXvЮ*yr28{Fw{4Ҵ4ӄ=}ClWب'8/qXI_b%,xTU,Jg+IO'=q`:k +;-AE&)1<RQ%S9& ɀg@]1Iw%mH:$tD&0}SOϥc^Ԣ)=Q1}\$J& \dö** L|H|WLHPYE1&$k,ZڢJ$%zSQ1:ujj&k+ DV><2ojh1偷N4ߐ~x 'H!Qgb);ɻISqIm=u CId+=)>O=<sy Y[ P4fƺdzƃJ0cPKP%*od$L"# 6`(ABi žש]>%#EdyEpz}vBI4rи-)CkX+} `qE#UmR# [ԕ'ͲSCZzIWG) D"A)ĈRk %Pb@huQ赙8[fٮ5'lgCVIb8 =4B‹P¥QQAePUkE$Sm:L[`(16@Y@X0(<@!6J%LIRT`uU81b5 !4:6R-c3q6[vX-l62v-4-ܫ-|>CEE1L?>Nd;ڠ [l٢UpuJֲ) TP9Ƃb,(EEm)i m)Ca:{!+1*Pluhuv + Ei6hKe2MηΞi%foFy<rVcWm[IY]KBW mDk>,dF9tZ$x<2 KS{Xu!Z=dFRۚL"YE3j8b.&N5zl3q6ÉS_k.]թXfcW[D["&B {2!A6wJqGL:f>e0u&jTihXVڙc{Z1aOZiI*d뤚Vlį:]w3:XgQ]bo׍x #N΂D@'@fsna" Cv^`4vqvTaѱ=w{0aO|+2d ড়7pL(T0ce?iHe0>L*RJ`^@_Fh۞Wash[bQWCMrdss?k>e&7s[UunCkHnxsqQWL5`Ҭ^ۢ<8 |f㱒VzwzhGŦɺJ5>ë"8k0̘[jB=乺jHړ{iӏL>޷k `z$F:]"̕V{s (kb-Ei4_J2$Bo+kH1լ[6f.7dF`r+DU/SaD|`Cxѷn]3tr{\by<|YBΏXRX)CN4_NL+WiNzӔ6_rolU5|Wʝ{-2% a5~b՚9`ys]XGzdĽFtzչh/I`ma*ЬEe}WtЬ!b&o+\t6:=57OdzN.*-LUG_YrjF}BQ~a(p 6 |b(!f[J1276u?dZ,}"ݕ7sdqKLHy==IGk NVWB*su3H;՞U<]|n}N.Ģ:] ; T}֯Ngj<7N95ot#㿳2v txmduSI0R)5VN$16LL!1I D*nmpC%y/d)+Feoe%flA:|WۛU~!XDXn!]%J_.ъMo1bG;=6tj1΢G"cp0hϠI9DWv:8Umiv g@$|[oMCt >`Vt bvXYDnM'L-3 h-*)\6'C2@Ұ;o"H]JQ8*+e*e%WMa7_|Uu^t>b6/!I k r-JPL.D4贤Y&1u8`Bk,Y6aȍ"kZ5Lԙq]QSE~}zrC=UhFp("Jpp1.Jy2J.)]w35j5vX1;vc"iC~.y!T:cIs<қcFy9ޅ`ca"ZgeחfVbVup ".':y>UoX ac )T du4 V rtRٓ)E>l75EyᣗH>y$U9+JĀD뢲fl?zB~>"o/M~BkV}6nl*$? V)l o0ug,Ku7mݐ>r;FÕqƾ]ύ[PίrxYstqŘgΚg= n8dzn9ymw&!ږ?}8>xp'[J>讀ѝV^wHt%Mb_iNu&z[4EYPJ$ rڣ@2"٦4{~dq\OWXadU-4y؎m\C/#)wrf.{؍ӏmM$ Ar1~l @G@鑽;ʱJht[[DZB)-b%ФH2/sJd!RS\٬Kq/u"4{:a_$XHܧ+zH%j!~&U]JxQ1e'wGJё 6)#D‰{Y;ɷW/kG! PCǧ(1_'%Ry#u!E(X<UH"Yє">98Y*Q'߷mXd^duh'ob$F*:)P$5k69)gI9g$r:A,8zWɴIR`E jvٓ8e˿ t;KO\YI<SSM"F،) "PtOQ_,ԊE1t  ,[ԢzC1s ʎkn"T-FAMl*6&2Fݤt`NYIX׆Ul_vMX QŐkTmH_Ey7a~Z9w7yʐ xI\e~/u;y͗?߃0&i0gV`gJ%`G0O@ oWd}.U}&=;50%:gHHga #2i1ym6 }r~|PV-WgiV|z4gFh#xy⒧FgS6xLփib_'Lmr?/g̖`0_ßWgz_2/Si^oE l yȶղ#tad˗dew=LSE,wsɓnl۵Ct5{Mrp'i-0m -}kF7#lo3 -ZLh(]t:UŦalsNW%GZҰY,U4 l_iw ܬTܮTӜ{Wxڽ M??P}~Û\ط7?~G;Οh)>}}!cpyM.qUMЮa5am8^n|7.{*E|{S.vM҈DžpW?!i?磸/7rJ?` !VV.b8 L$΄(7qIs]ب=$_?/BI&H%aqjX6#"dr1pQN,z+)$]J +kOn4}I=:DOD}$HşQ xi,mq9{.61'(RJ7$mHiS55S2F YZ,֍ӂ SIĶeL,'f/N\J3eRYy3HBE DUJ<F NŷkZŤ/L^DS8оR{%,D i u`V&ꦼ%sMb\XKG 7Ds>HFjlGz\VbjƒbT‹Œ_-RrOg{LN/&?}d`v&DT0'-mʚ8ːEVz#'*$C*ʞGdBR!h@#L. xp:7_-q#v8⚊y4fǾ+P{`f|r4QJƼJi]njqF-0jce !Wì1`̤!3@Zt`bQ$,Ѩ R |ERsTg;J} $3XMeD4"i+s"C4G}^ l <^)`-Tרrƴҍ]ֆK4 r!R:ind"B#KF5َ'>u \7uLgfɾh+pqM3^K K,H KY"3{4(Q̥WD\<. V⡫n?<<]z1d~/҇v)_#lIךQt ֏Od:@>ӏf],7<`X2Piq2W$q0zRᘳ;j\7dDHf\U&z*sXTZ;$ʒ% %Zսo/s WZ8]HUnKgmU8aeo;>sxwʈrRm4hk\(&<h"j˓%lNB2xJ(T3cJ2ϲ!`K\` OJ @:"I!G BcP8 Zw$RD@8;x3 ܍Դ|ҏsbd\TݵA%6WSV{DY6Sӏ=~lV<6Xњ :x (c֑6Ei Hp$Wn^VFepŻͻib;B&ӆVc`[ŽN]2FG5/ˑ[my~zo Cf9k}VG1".]ҍI]#0CMN4-\ /hGҥG뿦AnQ~{sݷbÿvp㴻߅ܝ^xZO_^ӋU#^Lf%}_kdn[m]v`ZOĴf |S l+.R1shHɰqԷ&Z+JNWzKimeY:2ЪVmU7n3Q=[@RQk՝OϿ T*e0LM73aT!TJaQiA ܳigJB(jY*!@̹,OȣOITf$a\I7n=cխ%8_Yz\ׄsz<[*SDW8:@QH'`L2}V^*-%oŻdo/a.eo -=E߶9X5M~;7OHnX}W!9 5&~bhN '\c<$''C>u^d9%#wQb2G.I ]Bӣ "$óFMR*eZϲRdC:&p!Ʃ`1JPaa2 k}LlQ-yiyle?O.,?fKPuG5ob>4%\/-ۇ 4g5/>au.?[LEg)WГyӣӇ٧ \_Nia&|&.W8ämä5(-h]i&I(H߮vϼ&%-ꛎ}%>=(/$k.^'|e[Zz _˟W򚶼hwկoFt_^ʿYMj`VoG]^|6w 50zO|KFe{Qv3_s~fRrVM<-}?='\uK8jwfDM+hAs6j< Dk4hÃ6J"D륃$hc,h=Ü(& Xڰ)VYXf.% e(a0\f>ީM6\M/Ӄ ǒ:7s^[gٗ얪UKx%FѺmEmk{;Ġ\T1ViJ+]10nrzGVHc^UUV5<f[h9wĶOksYѓ=yzdk. ۦ+un]#3?Nԋ;'>q팡6%>Qu*ez7yCum0IwwE-SReݺ!5[[fSm!7lA1+-7^s!Uδ#SL+B1fnԣȲCRYdFtt7NeLZ8I']MvfGd?gwAkEqb{Y\{zn&6ꝲ.Z#&:adx!s%f)7*Z]9 !Je25@"o RDV(g64gt Z&ΎS)ȰRg0DUޠB࣒+ `w-mI ː~|11~JXJȋkgcQJb\:6i(Xn` 8W Aɋ[3ӱ4Ӣ߷k:܇ R`P|p gHJԢB0mMfaKH[2 Z*YPre>BҬEt5iO"tKi[*etPr3+I"k0 7W\%m4,D9/SlÚ,-ɪ`VFp:@4/_ߝ}M'Q sXs*{j"n&hU i+l M'L6JYϒ$hMpT2*=+UA56|{QU$đxMNi 10d ]^+EǸGq~D^}0֮4 \()C _8>+>_|-0 qFu?__᠃;NYՓ߾Op"Nb0Jzy LqYs˭#:F QvA>3(4ֺE &TƣIpnGЊO(exzX6}0DsUD[*tP*7CW'=Ow`FqpC+hY86tE:zkMtKJpk ]ZP*ĤgHWD t K*U-thM!]QAF-EUZ Z &ut %n]`2JhERwt m]`Ac]% 'NW f]=CH&J9n ]%_uItJtJ(UOYg{r7n)mSNTkh:%qZ]Bٴ:*4$Et%ƺ*խYUB|HWZ۴Đ\'毈&};tEzCS̞͟|qpqhjq(eì+z]ю:X)i DZCW S~ZxJ(q"UU[*}*ujքk*U-tE5Gs:]%M]utU t]tk&vtJ(ҕ@7`e(m]%ZNW eGWϒ$D)s\BW S qJ;ztE5,8gl~{:ڛDKM?A⸆m#l9/ЗҬ; #m|g#o4p ֜s)7]F9+l%G}ps5LvSW}Wf.]u{ͬϬDs1Hr/󔠐b#DtuDt(y]ʖ|OUP.wqa$\)k&y$t@Y | \KGg`6ZpPq* ؖMo0X\**Ngu,K/fO2tፙ]ʹmp$$C =_YŪ[= tC厪"En_'3&=òU*лtEFB^xJp㢐*"r1U>rr4bQ{tL|CBFb zP!0'ӥHaR TD2'TZI%hD0AP4 |6ܱX yʲNl:GG#0De$$e#)RA$ 8C3TND7IךQ$ `FRITs}*8l4aKSKAA)RTaz9?(EK0D8L<3=c@H-݁D 73ʽt\a yB8//xX\ܭFR1E5Wƅd wU-M]L@0FOz ^;* )2O w]aS~oNn˪?8/{r$3H3 F"һBs)(LQK;3 ^|qo(R\% H ˥CRpkm>,>$ Y>1>Ej\}DeTfV/:z%{B}PҙՔ[*ﳟ_a&plry5ra-23eY[Q6*M%(>-+okkrYpo=ο?ڶ&vH2n' ?D3wZ{18edKEVyH2Pszayy>41XtI7zX[M}^΢l G.bL_FY!1#6:1im NnQE"W'z(N 4&[tpas5$4Qix`,(3xn230$dBsKL}wf8rt@Н 1f;eܑ]ܫTtpʮ[OK~!FA`¨@ ,H!Hi1Zs!"Ij,C,H I[chCz^ wE%࿀b5INMH qw[/rVj?XVaZ*/ӏz}aVlw~]${\ۻʊ8n9b EOeCfk }!vwr쬣mM5m6]KPk&X7}nBIQMnUl~ang!;OvJ9u8uS8xW>&~Up;of$p3jg>NLj9+j$~GylVS+hAs{\]}F|9ؐ3(l%h݇2˲Kwom }56?'=w 1=|U g[>fI[<1Kf~uTWJ7.m!=`<'L n]wW:|7ৎς_SJ #,gD\eMlkMךҦ~P=naM*Ѭ6RgJ#>#PnQz!UJ# be}(^jʈhA #(H8H׼Qsָj:|Ӥ%vݿ?-@|첱,2zh{g39kGQxCr*l :i( YI9W.-6XSj`1VYE 63hQ9b0XJBԀRA+L\5x t2}ɻ{կlp'kOQ:wgD%X=nr+ƌr7. En#cʭ <KT&G2}J;:i%Cq(~#B"FS$ [0GWEe2%L-b :9 ^9*C0!B*CZ30{-#cVFs+%fs68:wRvlӅ>ha^RؖRȤXX͸@\#ă7 Fp IMW 9N`./1I(J1̽)Ww6k]pWp>-, fW7M TRLzSNq飶|`a &`̙@Ru3cmpglڸ6x(򚹐w\\B6'˸Ru -pUݝ_=hr=o SLye0sN&~FP8j"F%,\t1Zvh !9{D%$ڤB:`^e) z޵6r#EO@&ŗef`Yr$y&Wld!ږ==@2nYf}%p,hSdL̽e5;ڪ2j{mPB04HT.3ਵB+@d'\5r^gYX\4q{#K a5KYD2 =gwQG- ]5$saX<>Vw?kv9vW?g7HٺQ,w;ޏHս'ZFg,79Xj)")ƺ{wzVqUa?WB8,RCIk8&hE/.F Ny%rg6*A-?hfnuƟ7sGsqŷ0Ψs.U R(Y>Fb$AAJxd:?=ly^9![泄=CBQ " 1RH#O)t݅Q'7Rwt 8^9޾ + R$MSNN+2"h%0 Щaww)ujAg?jBv RT۲R2_EFA7$|Jwh,جKZF7r 3/Q{7H}?<#d*lvS_mmo@-3x"t #^nөx<ߛ.'ԯE39/(7 PjḹzSo&fQ]nhI8G1 ǂhzsOٹ~皕6*P|9MN{GpDTV|B)uRFeQJR),r =U2C>ifi~q/b[X=VR sR<čCw2!Ѫi W'i-Bm3kAgUc]MY\u^Hzp+[(wNy`裱̦R5) 4Q1E QJv1q0AUhwi w,=z%e֪٘|^fجq#4mp4(Va bX7yhU`mOBlG¿x'YiPM̊)͵^[|ɞxHd@uKA(uwX묭UƬe2x_rV\{87n;Ѹ񕶾xw4|;QɀM{Pˌ7ZP$;ˣ׀}R$*jP)XE`%FR:|&\:~~N8-ljɈ018:F&^D$[B;ImdE 3Lb3e{aY[)&g ",rlIDɐtI\al,UYkO6xco1YΚ侴Yl,~-||+YBO1>zOr\^zm$n,38uA^ro%T2Ab)K9c^1X)&촧VNJf0xmJ##&hu"GrJ{Pe5q+wW[|xėX>nnjrhXE";_ȅmݲL RZ7/GѼdخ?vzUIO^k}>L'-[ݝ@q=鯩g.Mk>9+[;M?zn}Ԧ_>]zlsTbuFb{цH{4gm1FAJ=vN˖K"-ۗl_my q%3’A` dΤq%.7$ΥTPg>nd%̺B`Y)$')g$_W6Y T8r?k'$kYs( ilw/a\޾7-*cdZRCc IݮP},{zhB>?.. ןǿm,7ZC> kk凃>'0tQ7qy29_ "&CFz_?pp|q$D+!w?qK<~i E=u֐Ե'iY$0g敳=L\`a"{fRjj}VGzP¨ k!>\2eJA,X s@뛖Y`|JX<=ã{ ߜ]/y$!+oj<`oȸ19:##9dBF+*H<*<>@^jq /ǹ=Uhg~qAB2ʒgaQEJc>+m2 C2!gb NHNV8f}7mdKV db61 GX]ܝSHIJȢT~xij~Z;]̴߭LE& `BIĴKG0*D&И"a U+c?_e<cD1-+LEےT*8ۘI; QӚHа^*&GӭPH4R'*p<:RFAd@h=J#(dQ^9\f)f}r.|*e]H9[YQu9V∆3?ZQύ- ʲ[tT9nK~Jo:&aL_ Gq $>aHYCn(e 89;,=EDݥF(V h>{.ϿjJ_4qA)-iF-kԵ,sOgUŝ7b~]~!V3pGÓ {5kDdy=M2'\ʻ{: ln'o1'->Q,d<\paKnuӻri9zԑұ?i($ l?{F#a{G50faAP|9ȖO8߯rGKMٲ@b-b}1:>F5[F5͉uM?O\~?~/o[._~)!}{!V[u--uWT֨C@GÚ~ WKn@R~ŧOE_zvz^VZwe BkQ\}&5?΅&7JR9%#k +~/1+yW&.gW2;Du{F0@*) SĒ90xn2De)yъ V"*h$=n V}>#|} m£k#C3Ft|əH.G,C*:8ڍjsFyaimGW&ّ%4e&,:琺 &q6Yr+e:wb"Ι+2Zpk%F8.{` S5gW$} Y7J3V=3r1#Mm|qI*p<RABDbMكB󧍵 \Z[[9hR@5,1Ϥ*1Qzcft+骛Bi>42,1(8 Bg."t`13Wy8;: =|>tK|> q{`o­zC׷Blb{f-ԣ˱s`+}NIgd%$ ^:nYwbdʑW[N'ǹv9hS-asBn*hT4 ]85NCpW~7U6ۦJe4%猱3{K юFH8d;np|o;o75 "$*[CZ+  Jkɀ%9~,LjGyvs(gJ?9srf W/G2eo;} ;eD\D/<`BԖ'Kx'|D=!B4a'&gp>`rYI, Ƴ+#t" )[Yw-qn T(+ &Xp(AOln#1 mųeɠ=r92dH!(JNk羅΂$t6O͗=3"KCcڐ(%!CҘA Y *!DH#TAB1L˼?:B-8~bxZD'0a) &CSNhp*0/94QHSO2(}שŀ?jB Ue7 !+ ^vtY|0В\#ׂg_l;fH+םUNfߟytk;=zo[D R<9}zhZ++wmv{&7h6ovi^ፎ.98Vyows>.}86n4Ԥ59އ]>tP+[@CB 5)\-[)C/ "ܐ `]\.48HZ_2 /fAݞEu*A>k:=vg7kSܳHl* >Jlni)oba%nED7x}w6sÚ[ar˓ .1jR5H6Vʘπ*YLlt,'BhcБ0$u%Fc^M vs}[:k|lU]كFkWR3EΓwa*Wf`'A<67\VyLyc!R#N:?U0p%|߿z[^.Ԯ[ ym0^>aUx=ë .\dzt:_>H`mUH <_=?=LZ]Et>LJL--I/Z4$YMk7Atg}M}}1O}|)]h#KB! $BAEBw^~[<v/iuIdņ-8~u\uVwTlсzT\t%2={nwG4.g+QR겧 R& ъ0 ]%4$-OщP`\f7N7cn 0c&DԳϕ\q\!ajZ *a{Y+ۢ3.{B{Fv4 0u0$Agcq{.a+}Ȇ!I8 o0Jj_ipr, )x70=S*X(@vVE|ыeu`h>"*/]^I\6pdYFp7NxA<}(bVpQV=Jq! ȲR5sΐ5 ֆ,7_m̵&h& ÃC8S]SPQ0|* ޽YjZ [Hw2Y,w~cP+65ֺ݀qZPIFj,"ol_IxA@F fp1RӉs[R$>6D%Te.rbUV0P;%XZT8$jqd|Xx!K~Wp)}Ä c!G(dqt $fOW-s|/t`Nۨn+sQP<XmR (LZ6d?X[-]^%K}ʙ3Fz%&J0\6fkF |pYUUjR”[R$%w qYd.sKvytkㅮ^c}@ȏ.|PQrb!s fʤ»t4ȴJg#|Z q9HUH*[I+^ E!'%1dIkD&fCeE&vE'5si^Bq^Jm_J`Ty9Mm>99l w65CRXMˑLUT_ zZ(5Z-I3h^HJ 4˼L.=21@"#cԃBZl̒ &X9|İH %,W@62Vg;2*հ̸/XF,{򦫴;ˋŒ"kmUfw48>N'Ȉ-"&BК&QqПD`)kγ,C>[bn0ie&R10 &Rؔ ^C) M&ێY<8C,[{8⚉y(]mvue#j vӌBC&R.eƣqF-0jc%tx6UB&fȁ !-D cV2AَQJSP/"ʈhFDqӌWr"C4#$J(;%n) $mRJrƴ]ֆh@3>@}ꤹHYD!SM-q#wjOX,=j侸h+qqM3^Kt{"KK KYF2 =gQ hG)tB@Ea}aq_U[A;܁W$ћ_h!zQORӶo.ť]WK @=C%HP!UmqҴQKo2$hPX['trGzUs)Mbhо2y2` jo/^R:g"(`GTѫJY n SDFqJVƁJ=JMMM ggK2RrJJd,j.E\2r;&'S7Aט 1NEKn4`9%Z(\fDl.pǗVp(D qzZKohɊ$3p~8tpr# ",rlID ]"">iKt2C* >'Vnܖ-Mۦ,6N4(M/U*7`?Eh6]y)ZqTxf!<1Obn;% yȩG'^n Uq qUC\θnzXx¬ہi=2Lㅫ)]\] k7BŊZ: DA܃eWyz SET#dU:aqE%*2{m6"!Ew>Q *( Bif_k}{qfs X-y|3Q˵Vށ"U;-4k?V;)ʤFTJ5VuXLL&Xr<;f<̈ he3L91-sRađlTv},7ɪ= fCvMZ즂K\>'x`Zf3QĒzHۗxE O8쇞`?t|]>_EΌG.avQ QCP5$٠4qI&[#N ; ų^{=H3B-JCxh=ڬiVkpGoOɡM6=}Fv9u` I"&W53EH,'fcWޢTLTt2'sF4 REV)\EWUԄL[`VyH If#o_19<$$-i= `[ΘHq(Js5B.8ԧJk% PY5qHz.ѧ9:͒CqTE3​fzZ5Z,1/,e ,EpEEUI" .͎CVC{d޵[#wz_69;xEEGLя`9գ7f-*Cp5+ XRCc]=I0=~Ax9YJTTJE!g)cYWО\j!u!2`DNLlՆTx7=Ayt$hܴûݪ:^=;,HDwebǕ\4:{Q(BVTKidVh!xt6酞y^^sB"g mteCBQJ(IH4 O)t݅qp\<3r| V"jf)8.IߧV`eDЎ(rCqJߣԢGg5C!C&RT۲R / l#FӖ_Wh,جyrBQ)Wl:td~ `i]x8G'!cϩ>7OK}4|G}5TRm] }u#~zXB%lX^ߒaESjޞ 3EXcT`DKlSI㙛m7],FEZZe{9Q.XXI&fq؛cw}{7ϝ5Q3one嫾l7=ڡ(OYL%O6wf*ld3 ǂhrD{fa4TKy5:]v>"fq*+>@J2.ilDΓa*Wf`ȗMhrgU/C'8p|LW?>`Y>zCV8P jB4mHhUHLcg^!s>nRfG IlP2#,FdKq2g8VRHj3H\ Xm$ӢEzd0,rocJIy{m1t!r,Q2U9>)4d/iRPxczgBDj j >j[RaX;T5q<+54|47EqJ${? E⒌}CLPRWTY" B`jTHIm|&i%Mxd|̀ٹl!Eey+w*yIy^ui&tL~:ǘJqk$9q6GNd# bnD )) gJAxIxノR(se~[׾|G>U[A;܁vv\.| 'ߎ@gE.b A4(?x7 ^iӛz'1ӫ(!>LgwOX٣ -gijs?iZ8oORgLsҋN2Iօ;= q=?_-{“wK]хH).:l{i3b/%߿߭;н{>7eÆ%}׏e}럺}?b򻴭VjuwJMH~˟GhDL7ཋF4Gwߋz=.FK7:+-^֐x.'"\v:؄w)oBJ lf> {6jɃM Kck1dNIQ٥ۙ~@2}7g\̥4]́t;?y-1cR{1a,qdzI#_[3QĀ!b:&p^{:Pbq:錯y82FDf)&$8E5&c*2NH2J̩*guj881"ڻ|r+5?{8 ?PpQk3^Ft\(LDt]NSb&'^^*TZ4UDFgS4 AZ2SI+%bMU,8HRV$zIT"g5P[M.*;1L-/J|Q~/Go;.omd-%p`>i6@"nx" o õC67ЀfF 17Ն'\' 3#tx7c|s 'y>\l Bu57_]=}Us۞*(ږ>t^>?;}5s揔#wI4G?Nnه Xv"A|y%EWTI)Gb7- 5OQsC۫׋5xm0z=7c292|{~ܓ2# 1Df}Kcͨf6,?hs #8 l~~9З7m_9=>J겓cU/um,m5l>>Ho/Κ8rv6gb+>B'5\Pw?|^>Rz!#pURc'O|_ur80Ehoٴfڢi2Sz1[Ĕ]ib>}2{Cno@Zݳyr \\*Q5Itvd:jrj]ŷa~lNkXu:c^]M]0v>0wDtp%CUNK?qz\Z*%LIl"`QDhۮomM:RӜ+ER$AQM>#'uH_ ͦϼ8]v_zE~9Jk[e {tHHgCGg'e z>MId|*&@I› " bM]UFD*{u*eDKO= $jWg- C,:LkZEzBZgks"7},Pl<&wp #bݭξUECǂzx tAU]/T t5i]MJ]`.c[y驰Bza^PL|A(Κ_ %夝@]Ն:RLWEt5H&X⾞Ğ*kOTYm%ꘂCԟ ʇ1{akQ\.;隆j-iNJj1פl[HH&!T'^ӘL-4]Đ:w,W$ǿ,}g7,C r^Y~r6D5 (c-E'3g#DaMES^\FȢaƪyhR N^9ozd7)p5+'ŵR&Y`TE؂'gY_7:yړ=Q}DMsi:!!i-Bm(BYUۯPD!=_`Kj}b`?L5mʙ\"C!;W^ۚBQ( @5*-yB1)Zi1NOtVcRJ[WR8ʪHGIQMB4=䲅.m![ckJYوKьiec1Zs#I(k!2@>>6h2'>E oIHvE6NR%]cZY->c9F .%rX"fUI:Qb%rHTןS%QJRo[TYTk䪗1(!; "luؾQp7H-rL),Ok6!qQLh&]`0Dle UthՋ"a> @ #+ʷ*!P$RrAtQVmSH*eU  YOS K."9h F'E]% No_mhkDӰX8zZu $Ɨ"Dhi"(R(6 je=t $2 a׊؄ʥ86Bi8Aɶm_%7/[H{=z)mƬGp|-c pQ|q"5hs|y}( N" ^+pdr v^Qըq$y~9X` /a=\7{hB:,_Ր(CRtˢB!33U_75 )@"`c2$ev C4J6yd v)01C8 2`:ѫgB.ds;dBnuZ:-mC8A N'e@Gh3cvX6NS'p3?Yb8QM$Z ej2ZdJ|yT9yeA;¡E*XN$JԚ5'o?c9ΐ6p; j,dJ 2Px!W ૷ ! +M03,Bf./WϒNr*a _ \F6ZxҘgoVS0:G]5-nVn}$h (K` Ե[\A骅 7x Fg`*aۦh ;5Yxj]ආ$]kp$CȪFf Ƹn`ʻq;~ڃMeA@x7$9!C>@H7<ВͶ~!頄(B0eHP D'p'(j@z-jF=L5zc4x ˉ B\D W}\A\Hb9iM'yMq7braB7B`Z `T( y&#>j =py??M[e#dFiâHYsxG1@sJm,q[#725 iU gFPq\!byPĤDžd:BWMI ܵ;k$8v w _śZPٗ9%U8:8k 6@27`93#̐Zk5@2|/4%8n<!KyܳŲ9mِCL|U.uM}٥ :FfJPn #Cdv]bAn\NF Q|Mh,?@5 ߮H!ީq1eTsJ:ōbm5V;ܚ \΀/Tw%g!9=嗵AKف5*w &j-n68px//'#{M$R\_ Z~$$$L D$@D D$@D D$@D D$@D D$@D D$@D D$@D D$@D D$@D D$@D D$@D D$KY ~5@8djH '=Ry@D D$@D D$@D D$@D D$@D D$@D D$@D D$@D D$@D D$@D D$@D D$@D}$xK+_ c@`.7k?z\HLqň"H ""H ""H ""H ""H ""H ""H ""H ""H ""H ""H ""%l r?Jލ-o [~;/ǫ>{ܥAq~r=Z\NqNgcGկCq[¡wڵԻn{έ<݃${j/W5{pMcl5hʇt%K潛\g.8!%~kgerW\SJ~{rrp{/r2E//'fkoOۭ_z/!q|9!_>8x~\bմCm9ret|y^ / g=I:B?xcb(M ߞ*ޥߕ7f1/r٫+Wy5ĠChV3w=Bjla>3(痟KϦnNG:'.ᾔY}@beJ s$] QtŸ$')km}'߈C/@" .dP߭E4uS+(kv3/2`,k pRIT24sǔDmӂ[٨m^؜jx6Ȉh@EG*"bv,%zb=y||@}7Y'f &gs$BG-ds;LQRDq@`2t%)cL獀K$ ZHl;?[Ap&S%֞{=y~Ǹ!|6˿ϭے\، WީoqoA$Rˣ/_l5`njm˅&nYۭҫ/WP>]ՅTOaܥɨ[4w"/D5OJ]ѝwU S^I#Wԫbǣ^zLEH)'^&@JeB6[:&_=6T٭^{ue]u#xKz}#"GCL 9.r:kô-8F= V%E*OdJ٩7-Hj8P{~diuq4 ]^|9X{*lSNUE(i4t4Jе>x|ߜ@\Sq&˧󛮽d-dz?0*^ 1^ cxn8 0aʱ(>&0K y3XglfnN4zơP7|[mKTƧMZ~ˣ?b<:@z\/><^[Q,ttIK6%̼Zneq8pM  'Lc !xU>FVjՒV-hYɊk+!e&XBdMjUU3BhܤEn#ת@%jPЎ 43g9H` @jJ>;nzGl? anu-\ gr#̿pن,S5Ip7&zsv}wJw :@g\Ϡ!yɑ|VF!imJ E\X.wQ/E!\iň=r/9^Ee/Vʆ-+)p)]0c%X#Jr6"DQ(i7&i b`?ZCz(jٍn@O0^jqz| _/܍aƟ)M1Z^4™FU9jDnbCd`/*g4]%#{]4r݇:0^SѤNge6?-5DXpZCZLg+K{%07Ӻ_pfO$/ܳ_O~*~ooGj'/U]ǹJ $9+-xۯ'boNlz~BeZכkk\ܯ'?bJܡs3߭mR쾃p>7lÿ }.~Ͷ/Bz5rƞ]i <9ۯnKCt|S=ϵ1hY} am:TOnPiXTm[-]^Kt~[xx\aYhWo+d#^xOU:aXJ֎}v8l2q<_Ko:ltjsՁl Avw]|Mܒtu>nMz`'Sk3bfC YݹPZGC?^I.صAIW U]\(RLBkd" ,9MNYpj:+p<ǽ­>jYsT#R,Jk_i#cR:0(Xc|eK^=0ݼ,./ln7^muW!?HG~L}21gIl$}zrJQF)T]k*)Ʈ":%YILt\."DS_{.&tKX?|?;=kOGŗhPS/$bodo'Z!23~OuOh3 @[ 6Uλ98 Cw8fOqPL]XFE5.rRm}a04hQ2ם\³o;뵟,R̢Y6C|Ѵ܀T,|:Xu)g3ZPꍺ1Z_[(UGRܛJ){>-?:Pbg̖_шЫRR֫rֺmu_C&h&LdB!A!CĘĘz5eC-ˆDNQvoAk\cƜg_Ýh@u_[6 qh!(A')hbdp"MnBrC4[xS݉ߛΦ*t"^lpaϚYb,k޾y,(f4(/QZf#G1@#臣*̹zR*$T?PN}mo..E9 G_X1 𿆈~QT<~m\A sE颂LU.zUo wʞٯͣS{ T.W|E[(tt٤jo[ԋ<[=( *0vyӴʇWԝ7 'UeffE]8\~,jv1yFEBW`J./DSR빣i&rK :C"e`T<jSnpR9&;V/;gŒ+6HM9;|Ǫ><t=«wݵks o.s+mVnO8el ߧݜ$Leݺt2[WJkM{y.JF4=/F 5ooO 8?ۑ~ꭋd+xIýe}Nr".nޔ8j}ZTleK{!,߳vΝm UQA %+)"%#BȀd6\|O hOWIolV0C}$&" R/l vllwƶ?@ fʦk/FdHN<4TӢQ6Wͯ=8u5 iDyԙ! D)\t@U5N% ^ˎ y.4JQ#e w ƃjǹHLj  1tΊ*HO)C)QϵUAz#P8eNys2nT6q eB4Tl=X-XT')JBǜf*G넏hU6M"PefAyg͢1i,),!\HIԤH5(R QsCՍE)DJV / ^giOcDcsŦj~O֞kӶgP|w #{%7sRh>G:R. qeEHZ8/Qm$"_jH>;qѣAZ #h"^$0R&Ͻ(CgwHQATuC ^bHO$S09z};sY5q2:C§p3[LmvH2{ԎPLQ*ڨuFaA} w%QhL3eb9̩{-ilE- 锶D|F$H; D锨\ Rpc/KD~sV{"Uv'pf~_.ߔ"e[VHG]od ۚ*..%|[dh*.JN*.ϰ 㜒{Hy0w_z>MT,ŧsE2q%0 a)$J\_VO75i34_d"Cfl@LD"厺3 @$SBDP2L,h5T"=&b:> EFE/O! +7_[M3FT_B˧>br\Yr{4/M)yҥW#} CมHPRAhH,X* K-B`u4[mMr=)ZcKnpճ,5{'ޔzMQg#a% i Q:ZiI,%ЎJϘv1GJp+NݫdeAfF)X eZPHQjW1Dic/6">ŭ73&֋[e :ߖ,@䪰.@5 2@FhpI*"XQP2\Pb$cN䙋˙1&qq ](s^I)vlCaq>UީKqn+&Rm1corvQr pBC]R+8Om@9q#jTXӴuQ7Ϯ5rQZy]br<}p]R΀9Oqv9R /&U!lqOi K # ڇauEdyD`X'\Y_<9;49޾ȧ2fOLV˓q]pxv|2hd+A(B d~VGڝ ?ꟗ2 **ٻ|1v:xITW%s`FhY&nA&HL91AZEN!HTh*AJj M%Dmw64,&: \ ] 0gc|њ l@kh?{:Jr0AOUʱMs;gė;#魂pϓ\I$ +y4IFE[\뽫{_m^V3Yr\;/Wߢwvq)$nUɫ8 B+'UĒLUUq:9oZȍu l .xݠ^{We/c}vg4UVGv|ߔu.QS*(鵕Ƀ7yMWFP]^{sl|+Q ,svY%{7\~As+KD)sAYmX?RSO &₉'W!WQA C >R* H*1r4M'Pޕ5#·}( Gr7*Z-R^&15t4 nE(P*oJ "J"Wqd10#dV ń#Crȼ26Dn#P$X#01+,>;r|V} 4{omVǟa͸Utx3-v 1:x^ȃZc9հ9, h@O184I1[c&N0Hri6QHΔ.p-t&NnF]?J jhp{8K9rf[?r:,K>Gr>_nGuy K_Tam<h սr믿 P I6L&H͙0:z((%e+iE /UN(B*餳T K bde!<&=Z.rt{F>mz, dcJ`0S V8N~BPRhόM 0E)|,\M[d'7ى'l )}&S;ݝ`ed8w6~D@Zh"o"qbF<:Dd:|NztkW*ʻ@o߃7i^d Wo *C;{(qΘeQIaa ^6Z&  n1r¯>la/s𜒑;nr(fe2]BGA&`ŠgÍr9oTÈYr! & &1 Z,!EJyٯ[]w# ;hZϡcC#AL1/oF|O3te䎋rSRxj< I9ւ^TᥐL dLK!H\EXЌ{2FQ$*{cKʂ`CL$Cp,%g29qw; BW34nO X>*xv 0՗`N˷ֱex 772eSy2*ՓY['Z.-VѦ { ͕/@h{եOvzFbП t**Pt3/ztG2۽0lͶB*Pf1[v>!{ԼVr|LԷ| ]qf];*N/'㇇ /r>辷:_f_??E_zH;Z?ny o'8kPxg~, =Qk\T6'!p*-D޴MʤLXc> i[n3iOr}r?qdP/deI'H)kѪm̙VC^*y)D.K01J dc9x֑ΚE4XR 0wpO95tОb3mwfmE& GN.3ڹʉy6VxZbr1]ّn|yQ94d{[>?;z߅$L0 Ad2ɐQN#eADHNa#|aiV{q{n l8a;h}5VFЮ.6J}`+ȍ6L5o䢿qRJ0@A`Yr%*i33Ҧ H8Ո7,n }'7۠e+?^W0NZd}n/(mVnͦsqёБʰo(k/Y Eܵ'H Գ6)6YҾ^t2¥/.(5V"(|L N*AC9yv *& %B ީ>6tNo7C.} b%ծ2Z[] .AB霹,ǸV9kdN}*rqARnϱ{M׻<>>+CTI,m!c=#yq| M]Pm?|A(Y?*`<C dMN9ډ該zS Vgߎk7fg)Y cI)2 <g ߫ -T̒ېDII:s [&#g*iTgϘcsu={"dE+vvrWf?7y5ڜROz|, Je4!~C+%u#WX9' T4:εs{[t*z]CQ،#ZzR*ꅯ٭qwDž-+wﮞ}5m#jS̙޳$K9E`z:8?VO*!{1d}4ƒC/=: {?6mE X k%A,J5ѰLYC@,8T=:poJ/9sND:D#&{!HOV%MC+juvikǛ)̫=DOi/_!>~H1 pT>Ql!_ycoD}w$FZ7N j"L6OZq S8V:8I!)Ǹ#W._",K+$<MII0RU(.qBUVy8JR-:y Aw!<*ȬSa2W.|&능0Zr=JfykgJ%0<$nB}g]%34GO~Ԧ?G\Op{tQcԑ5j@YxԲ΢d=zAl%a'ȇzLOR[#"d L s#IeoO>!m턴&b)*$#2_(f`QǼHx$ g#Bpxf(L\dQe+gO m2*la\[*BW+Jf7oip/ӸJy?MnGob `v$b&s 8aG0ZZ̚8ːEV0FY[ZI,ΞOlJm!12N¦W%n&b w\+[m[nQhD)F(u>f<9gXm,T20+m| %u6yȄ 9%5 Y%ėUbj4Ej춇S_\xx*xE4--boxcg8B{PǙh7mR$0)IKIV=(˗io w8sf7ߙ==Hڀ;r;))K=qjb,T)!RQ&IEyPR1eT$-i[E|jJo">jX uJgŴP\ԅqQ&D87,AX]zkVICSbq(xh#@ċj?{x ~%7LFѤ{?#*O&Õ4bW㢀ӣFHahxCpfeZtd1Χ|';y$zhMJ2h)E,/46Ҙ#+p)ʾoSBM+WZum&F}nsNҺ֝?鹜#ŗ;G0rRMh^m he\vaI y4j Yp$DȺܪq =008j=M`\I>%%I!WPT(.>Fa,Pd4&l >P1@Cγ !z8892\w*GD4X(ئ8;D3zX\6?һ پT]G]bs7ywKW&5nGrٷAJ RP6pǥ%BQ4R:]H38Ez)<& ==|midX| q+Fa9 IH֙ I?gG"ȓן$#!3ze8U"ÍP Es~SJ2#0e)\fU)E߱XS5+(A2&^e7rĆ!d!Tn}^V9+hc"#ڇD!XMq9"T{yʩ(gl@2P)0?⚆E=ThdZƨue.5nXGP:@3\FwvMteCp}HyBF4*~l|p_Lٕq~n厎SYuua9w]3q~j-hxl?D]O}IL~ĥ4ZOT/w c;ox`G[w(uxC==#MIҽddzK/9{XVC.۳Q6[Ԋ %:5.Iq49l oEc]0OЭP~G+[*/{[noS:a7Uqw)T7d@RW.Ӟ $jOWZoM-}>os.`6ݵdn1i"ѻԶڮ[";6ajD7Dtl6 z1!$IId FǙLDcŘT&$Ent=Yۡ7ߢ3}R?﷈Ui" ~Mx+^Gv,2|s~{p;/W/x__ $.sL漻I.j$(Ynz[3qs:ś&wm}_d(袭[m6'^ksow7x pEy _ƠMBIh9A-DC|R5|ӏvn aomwr8xEͧwAK~\2* >\s,-WYJ=\}pX(;#k~6p5gW(-*KYW \ ձ(\e5=~.pjNp •Ts2 \eq%9*~p=Jq"ӣ_y?Pg篠ZU֯&o߯Z_+43U6d0 o#inᴿCʆ z ӗ_>oi/iXZ\}.0a:K){aZ+ =#<⊳,:7YJ-z(Õ9#B鳁,.U&*KhW_ \$\'(WY-b-E~t}6 Uϝ R | Wxd0[qT(sLUr죓xK߷/M?|&Oo PTnPПծyAOtms[t:gJoÏiJJAs_z.yu5zѥCEa!/P=A˖ζad'j1cS`/PŌ; $2䜻!X#4n6T_lMN:.wiVkWZ됻t:vC:k֔z.hCL$]TwItOZ#U!2 7C\uyM|Y8E'> odv9Su`xk~gdz?鈫ɲY%whųή}-9BA;a9<5m9ӽ4"hԞ4{0-ĊSOiߠ6_fa"i a<7(VP‚GbŴH:ؐ$IObbLTe]_DfV: L SbxbJP sL:ϨdA4{/ͅsFA;$!JʘT&H@xF~|kޱc?DT0$q&0ji"RA2hd* 8F(42(*-v4RSLԡP'ZZFb<٥c_A5Ňْ y˛\«M+xV&ɠuo7 LsnJbzJ|US%^}/5yd7^9~+ 'dX B E%ٙ&XLXbq(X(z,#a޷ v,gGWxp'/9F(*:EQ{5qم&%<(RӨ2(2g#‘!r=0 4qAH4H) N,I r$Bq0b c"1iE*H\@D 9-0hLL p9bΣtb|#zt8՜OjrW!7ۗj6ݭlڻ%«v ugr7\#T_9۠ Z%R)(OҒ ().d$  l_NWJYYסE? mNkn AȎRF<+pSfe 7Q}B:{O~vMqg?pAf.trV|2Q&+O&8Mg0 Jx-E)]S"qՑ-E0U$o5\SЛf BA#W)X|O}WH+~NKAp~zjj МLVS&:i<rV1Uhg9dmDˈ [ߧ{?adžuNx.Go#؂nr쨙PP6usHiL-Aa\&VxM,inҎtvpp]g.mHWr6x j}Κ-Yw |[݌Vw@m %`Xlݷ:H{H^ӶxkyZh}CzK%7.uxl-H<|LOq3W[$jΕkbٖ:<\O|:X;uSEj|-n# PSjraY)oSc^f41F- -;ϝnTV,D.}wi[iF^?*HZ"))\4 '8DT)eR #PaMn6g_^tPGGd簙wMN{B$q*)<)-QF*VKN+e&R ކub 餜Pn\w67;nm3}Qפo'V4oE_D, -u}Il5ar^zl8$eI?-bzyJ)?[OTt:Y {AU ՠy2B- B"SB#J"NLbHnj Fy$sj`]TN[Q0KC$Ec(eJhhP3qvB]rz8 HoeG?K2 ӈ/a6hّ_v 7IJη/%xfp S(J(!(u)hF21`$Vc`p'Oܰ=Hv;3s mOϰN ɇ9?4bt{t*C`Y)pjy[j'r֙~C[ͦ>F~* 5Ƅļi O]cَBk"?Zg<ʄkc#/z>Җ5Լ >zDSKVGu 1d&gөw/BwG2KOde Ċ`zx^*d:S* $ɣq PTKW:&drv /5P21C7D/(9T,NH]Ԝ^iT3qvumR+݌|zYYX=1Idxk-[k7xUg2f޽ҭ_ yh5us56ycKyce` >; t0V5sH[W7j6\ͫLU>\$.ڴ~k}J~15OZVE^R*rd sԋJ2E8JVGWzVk,71o厾z];ůxAqdW=*ۇ6ihSt[n/[GPy9,Y㳫r\*X~.6o}tq2/Xzߙ_Ldzwy/]gcq9.EqdhE׫[ 4GvnuZǠDui%%0c'b7:m{n‶DnԖƤTPf$ 1_1@U0CTNe7[ִ'jy[i[^z.:w~IBhDEVFI'X `BJ2Mp\RbHv[ZWuiLkS0-NPD2Fgxe-PXLPbDxo$qI3Xr|DKf:D+Tf B7G._&<N&8)IY&XԸ(5x> V[TOW^Oޭ XR3u8Ni=^H1W p-w7IkKBnn鬭 3ͬNƓ>I7]G]9<ooFFnouɶVJl,5BGriU_̠"=Ť>m*ҕ*+Et}8w޿O8s Meᗽv4hCBvfwhW. QbE}*gvKn@,~vjEΓ6 _3zJQgGC"4UTRp]*ˬXJbeB  fK;Ҏs =irk9VɷZ=:k`d pTFQY~Fz4AH[O$)tP)^z[&&y GI/#mCz :sH@Ϲ+VGg)po:"gP%FT{rhq/ C;k_־pwQ"f~^=?dͰ"V=cZ= U~صBJ~3kj(̈2 e e'SeTYJJ[NCpVTքY8'Ga,s&丧>&@uID{UJp2 HErT*g8P 8x3qv zKBd+|:qUgo-7^eVhsbo^r%!`.EWyF"P{FDN(*ճ!rBq9υ D:J)艜A"'n\eWi\eq}. DRҞw[4WR32W(d\eq1WYZy*Kzt-+i}_s3J\UVR]oGWیMd]6>6cdDJ~!98RD#"ΰU]zǕn\  *a\] E<@sp7ٛAwWӞ ͝.\"ͼ8O6;BV۳o?n?.dw;g_7lbyڒ6՝ Ŕʹ%0e i]=˶oC mⲬy7J)4{=Jr }2XN?*$[ᾈ0GA ſxU׸g?Iq6JpIb%QK"* ?ٛ_/ Fgd ~@ βWN]caf4UqM^y Vx+揍ipg$ߩ!06}GTOᯠzcD _~S`]LO͒8ˉ |ZzR/&$k._]OΞ_nNP٦I! 27 `a/MsL/DK|m#=wmF`$<pؗ)&VlETI|D>D>Oi\"`$\;e(a_QgYim<⡸J?OL$P壾\20I*.}?0I-r[GM\]vX)`/Zo;+Sh$',CSP)pRU}W!3`Y*$|Ƥs:u^5xy^ ] \nl$9Ӕ9R t0JUٓQ'C(=5CmNAҨGj`7F`Byq 9PF­3E #$ 0RS yA)1RX+n'k\fZj4ܗ#-Ξx5#G [!TĞSkc z.Q2qIf$wJp*)II))!4@{cq0's Y$cUd;\.S RYU4.U&XC͔#^4&jciT<(櫲'cQJbV`81`tI{FAs-U #!H!e:O*LAk%nbi1g;Ja8q(Fg|gu: Vd˻j<mYj{>;%嫃"բSA#WS4f"jKH$)A"LSM jdļ0oy *f5|5_p|: :nt.h @.ڱ BЦ*~9x1"t}S[S13Bn6Ҟ)BzK'#}%`kkNoqUѸ=qMm)Ei.$N'i#nrO`@G;`|^h,) ܐ+Gn>!Z45VҰiڨSP՞rIF<`.)="t7aetFiFGQw"#K=CD$K/Ө;sP1G(rneZGEdXpx+\$JKҵ=];Q:f' }L;%Ź7Dtwhׄ$ zQtC?u鋹||0ѺԕyuӇ|t^Z' '"de)$'^cU/ ņTPzefR&0w!+6sī>SR`B<UX-\cƌL"^ˈiDk45[!- ehy6b-`rHThr"veztBUpE퍿'*OݘZJnh&EjKZ=k=ȵC'"y6o:Q`Ť59ya9 Yo{vj ›9Wھu{˛ Z\{s==$Oe܆懃.{AZ~uCބ`}'פHs:mS&._|.>?}^D[n'alsOej%ڎL< oLy8 =?T9F$㑅H>HԔ1тFQ4`pH7Kr"AdžEcLuqu;toɗfu=`ƩRj3Tä)mK~J$7U6zORK830Vic92}k|JHūW\ EOfa'?㤷u'gqp3{>0bo%L{{v!F"r6Bg怒Swfpe_8B_ճO%P0 *:m1J(XK +8uwUw x;[_H×ldCxضߪS|r5:S sD\iwFGԮ7 '>(jT ~ f sדk ~-\TO~(|>+f+\ϽA(ז֜#`8fR/f XY{{eO aUOWMݐnjN,i} V0bAt>{Wɺe$h}I64WuYua#cu{30f8g8^nTT6*&ޔFѻW/x}ݿw/^~zwwxOiB<_ve t˟[w ko5UE׌w9j[+QSnVc˜-ҷ7_ ޸Ww9:(ōGhKM{IJg~6U gjߖGIy[Q]QmΑfx)`? +-싵 -6>ZGa#i% B&2@T޳y Y,#\r-BE'Ά鑗 mr!HE)m4!i̼HL5HkKPtŸt:hGm 3X3^R;[Kr8 '=ﺨVv&l! L9W圿tQϕJc1F1sɭ`Z%)&tS0BQXdBjvc!#EuNOYS1G Em#"(-RJHH!$u6q6YGp\Ik^ՠ- VC]3F[=gԚat-{U5HjU-洙LT\g0P *҂@x2b)b'}EL J(D9=JPL܊0(.js >6!c07‭iU"b zy)a1M-h%Z\>$}G9 5x}7Oj{ mGVV jFP,1xC#hEc̏zVI n4i§$lՌw«7ڼ|7~OdWWMn'By@j'Oؿ8i5ޯGIe0yHFHjYX`.y֬węoXf^:#>hb9ჴ)`gfh]qm0`6P*:|gz5- z?-^׃w)a?Eӓhcft5:>S&\W}(?%?*[0 ]կ+{="G$= *W2A*jZoz>Rp1"$b=JR˰k$yS&-Q! u:rΎlLG  ?Sx|ǐ{)dR ,f\ .C#x?f(II&mCQG!0$5I9 <* s~]H$S̽)l58'j84EIeC3I42*d#Ϧ';D;C &̊L(oRLzK{) `BaA Ɯ *ٰ0f슅<3{¹Dmӫd.}\ ؜]A*3?;zb&L1兖`7 M$ *DqEJX9hKTRl8{D%$ؤB:`^e) x"Nљٌv4 .ZʥȌڢv`i#E'I8" |3)Rv0@ hJFm̊ iU32 hG9AQ`FYRhd#Nu&7Z6q6a.%`<D̦"̌ bAuύ01V~㨡މQ*e131AP J* |: Ą3[u  H\JD;083ryM5gL% `JUH6` {F)i(FP|vcLibnk01IE=, BfrxK_:N_81Uu clE1Ů{KƆ$dsO~0fw#PΛ&Y(ڸy)q۞Fqn^.]X#? FHzWY5IJo+Q oW?ظsoS'oS\,uv=^Qt;X&hWuĪr)piޥṗ ;d1Kƚdɪ/8ָjcՔtbn:[G|-tuTUmxs4;H16뚔9ϡor]p\$aZ?3%wD۹a̻쯥3O:4X[w/Y6Šڬ? wv+1[u&c聹YSm*lxRYp9v1.ͮ4]-d=]t}Q@biMjQh"Ǒ `!ф d 0Se<Ϧy}`AEe:"p a3*&*eD8G).u{,)+qzI]<0ʍlqlyƒǪ#}QJפO*sފV[ފV'i%ճzgNa0q9{VVIa@)aRHaJ j0Asaf|rPM*y5'Q-f7r x ~^i﵉"n^/0R`,o[F d_tGyǻ㯠įQ> 4>Sxʂ_?g?dS)Yd]`*s8—KgrB"lCFV0&{5@c4៯{!Z{6z;[^bC(ʿL:{PisC]X#ZA`N,I\eN ~ tRʲ@ 4c+3r4p="Iop\H[ہ+|݄|ph*o{%\%i9tJRjZ+F* U3WIJUo+j0n.E{~|Un߼2~'[߯X7Of[D )oml,oӄ[P}{vULts S:[;?~sA)AwFi U](}o߼c#tTb^J)+aRU|l*.tՅUBW] ]u!h)U U(tՅUBW]UBW] ]u.tՅUBW] ]u.tՅUBW] ]u.tՅUBW] ]u.tՅUBW] ]u~]]5S[#A&r4lhϠHoMo8t;o,] otؽʧ 5=avd{pP6Wīu|i$&*gR%ӍUz8 +e N:|)f9^ F,L^oZޥڊ-*m|ew1Y l}=_6_Hor|]zڵ 91T)b+t@HmojRz\Tk~XeWzk ͹5C2cάppz1 o &zQ2|< fl\yC{zw ܠ |ռj۱ux1]W83Iea''<`k$4SY6Ŷ-pf{Zy49폛4ue^fTl0L,/a + K*+L1&FNL{%`zYey|d`\GFzGFH,9RK>jvcRl-LMsxu'r 8M(rneZGEdQ0A[R.k%#RHDcllމBt&mxw0wH m6Ci4[g"XLg2}fHe7}P;i36T\gw *҂@% d %?p Kzpya;J}u*X~}z'8eɍI+D.'ȭj͑Tegi˥Ւ HC%%#_B$M& <)cZՌhEE@)C}޳N,-,CЄ9*`*5ckl֌*8㡺P E Zhfz-q*Ffu뉟''~cgGhA DuP1LG4`$1IJ$Ab.i̦\4 l`;8jK%8 GP &1@`d1t~mYc ,ACڭqClYkNkw:LB\(H1!̄eB2+uJsK$iA2ZE{,2fcu!@<ꨬmYFf}X&LV[㏇jDղFTF4:֧fA`mzn9 ;9Ov,ơvAk.2" 685I*ʃqySFIh fUߖ)ET߅rasDA-a[nU,G};m2FD5֑pK>[j@Sm4CAz>U!u%Qcg}9.Bֿ䈕7+B&<]nzkLHڅhp`2Qf$W(g_QVy`S;PȵA?EC7_(œ:Ү*"+KpR^ 'Jݡcj>;[%sV UQA +I)jA>R* (*1 >iO$DE0C}$s=7T_-l |9[blѰK N]Xv3%=DÀ/a2U ebH5oz_^`(o:"CwtAp甉R.SIג@\hDH)cUր;m<8vb˄Qh6L1"m_rW IH aT3#\X"|` qʜ%} RńQ%ARΗ8/* [pDHPyWBǜΙ$@넗hU6M"Pe谨|\UQNdL|2b.F jRL hƢBS"%/ ]gi'c$cX8>T˖ӃnXem5Avmjεx:3B@cG¦ sebI( Z3+a`E ey)B=y5[HWT+*->몈:̪XJ;L*\%Cr8VXd)DoM1j(zz_߯kc AEmTDJ4-!(UӈR:"\p%}|;cφ;YFYQmʍ]n/G=eɦRȭxyMp\)OG~|?o(nMN^ѽ8("ꓝCi43$"SUK棊[ZbBjUsz?G3u|\ZK^Ķ5L;tj3 ?)~嬸|g.%eWt0׈Et:.!ouA%Q;~aIn-ikKfVIJа`opprxD%ΰ-%%LhD`.0nnJh'ϟG~^mɚ3d !116jʒ7ZQP̯{~br6NCJJ=UwoEx=''xt4DECA&5(N.&FF'$l[.$7JC'lgKݩ?2/WX,?x_2xq{c{{7I>7s}~GUy+9w1*ݞ[{n_ it${sK [RIdȗ9 YgJe!;Av3 PpO U[m!&<$ypoK L3M*aRSBPƹ$ TRIψO̹\֎)OmpG ]Їd٩oMIpylsX]2GHfl]w"Jc~QƲ<6?[avXK\'WaY.FD1n=fv: S aPS4׌,KQ ډ {KHl/.sMp _Ywxq& mW'5 (u0 sJ(CbHD ɖ29ۓ3M΄$u,(\S$Jh R*v LhiZN,`8R,R"Ѭv̄ؾݚP5|i8-UYrO?J~̿59Yr1g?@):ݜM˛ L3"6A'|&F 7drŽGBI)v2}<<}S5$rUQJ$:N(O+_qz\g#ʷZ3ڝ:yHD^4ګrt|N2gz=[=]6ǂWN 3 61ExZ?-;]M.-fcsb.A.sbn6Gpմ:COn?^!+zڥޞ@nn^2A q0`żf'>{^AuX0{ HX m8I,jbXkۍJ4J_zv{wȎߠ::y7ߝ_^9ego^}+78x RL;`v;]ǭ5t5l56G=-U!HX |ɭŏW|7&uZmuZ"5Wl:7)ݦ0se|1AuUBh){_GWc$zΙ_Fb)9*t !@"EU PV*%'b}Qye%=q\M9$Ek*\.+gJ a$Ybc$y3Vt&qz5ڑ}W}x>za|N:8?\j/#b4d7ͪmivDe4wۙU<ϓy^{dX1x[p#!;۷^zQ+'?5^ye'oNE//^/ErVnJF6YvRӥS2IIV.zԪ29s\Nk\bB /%UD-袭4!޻OBgtYl'y>-i"I8g|㩵b\R]ԍԿ -Fdj<;(^PhWM T{0jQ`@n,Iq:[`?;z-{BGNn_(v 1bFD4VEmK'A@)b\ O>=bHP qDN|F1uHDR*['&g3Pgz9_!;IKu9uY$1uF]%z(Kr9ݔ(D 4Y}Wߩ:]/\q^|6Kǚuz;[\{c9]ƎpOFqI|U5! Lu>ųӐRNS9L`s*%\ҝuYvK(|ZBID` ܙ!N H5IzdnmT`Trm6~mVORxsE*q0 aT)$J\Hg{9l/NPɝ*zk !9 .6ΙD&u!eMB95ڇ D8.D%(ÁDVHu mm"&a1qi9l1>#Wlx eބ[7nz>T⪹g% V' .]{<!Qs7T)Әn*nHN;kd44'ԢVG#$K3J1&'–*^J3x#zx6,Rݵ_}-_6Emhn+z,KɌ)+tϷ$J[$Q'tpx_+[~TQÞ3*s֩띿-jn9x vV< {'&yQC ќB*0)yH94(#knuh doFW#נÛR{4md3 &bJs}BA-YӼcjju*A->8,%}P0Eq&W^ Ʈnp!<\[UZOwoWๆd c" :Mͭh%\bQQV3 Wr5~r5!'("{ťT`:Kˏ>0K  g,8Vŕ'WYZ}2KiXWo̞ޜFW l+cׅIZqR8.2+Փ_=Ҷj<L~]]qҏe6o5qկR?ùLxMu2 VTM<ŷFVwh?ۅ+D I2 ^;*y`$vD&A$@0t^0am+2xTK]8ٵX*D9E#%68Ga)l 9a @P A Q #YtEwF''ΰ ;|ztϜYK *odŴXrbVā /^: XN!W՞3lbe.Kʧҩ "D3ŪI?|Qo+F;JRNJ/+RVF2|'!J;} ء"wQD$! hCAEI%4 O騒#aȑRA,3"jG#u oo̳V&ù6Jʹvm}Lf6깆}(HQOݧEoCex8SJE|Guc.4*(k)dd%\/&%D[c"QRxh2ӘiIQנ/7lohs,90KR֛}㭧꧟n/jEN5ګ4)A1|9ABEfyB{D8"d]^/C / ZODÌ$Ē+(Gb*X#0(2 ZE*H\@D )-g''gRXQ娀KzgLh7x5'̑Su4mnc{3ywKW[:5Zb8ץ Z%R)(OԒ ().d$  l0j<[?0oy])6P YLŨ(yb2f/ͻKu=>y;{պXs,~.@m^ W(~jIALji ={Pci.I. 8ǖE&W8^[ ]M27w=JMjL AFHPؚeh=ן9Cy`z~VXwv;k/w4ԖrߍmFkk v. 6}?d̍Y4d>hubj\GRK_< YN_dBm om)~a}m28./ܴf꒻E\{wxx͇KΆ7-l,R#1'7y5oD7ٌ[4MgL*e "k 4/i>?2 8qgXl.$2I]! QZH<`j%4(^@.y\|Kbq£ǪN f(_'!^'#F8X0c⡿O௢P^]E)KrsιYRg~N{?O-gQ~Xٝ!O#סcZ5  4q01Axv!2$&xx&Nt r~QQ1h׀0 -!t|0xk }׫oOe7TmWZgZu[_F Yk[ڰ6F;۞i{5|X !Aj1R҅&@tvUbUI Z3+a`EOjt)keY:ӪVUkz55gvtbeca^HQ%ΰ:9?D GE,T(yu^[n}7y-_op|6ydj^O:/HTw\Q~>1F'iQUZa|q:n\f^VIiX|B&?v|y0.y_'Y䪰G?-? >ZbQ,5qfKI Sh t&fEיh)oV~JmdMdP~ĘĘz5eC-` ѨSHJ A:2nӟbN sF.L|먆!{v!@t H8IF##mᖃXibh6@(5*3.}ew/y,3'fo;s̞1Bo] Yu g96*(Ƀ R j:S*+;KFھ`i_3=5T%oxKLoe?r<4sQѤԭʈYR9X#J uLAe #xF rh [!;Hlbp|j Lsw{3 5-5iYz:?C0OY[^Ch\&]oXhZ9 wAI'vve`m%(V. pIu}7Dɱl/_PVIEkB]l{]0fg >/tM5ת".RޔB|z\sݦ^G:yew-ij2N'-hnܘ'(g!WR\.#Z)J 'lR''+76H:B E1(B)B l ,RX:mRdOA3h U~"rYސJOćM^E占΄(zh21PH|N^Ul5A-FUIQ#e OMpjlb  eZi 2ERR jxhV@U a=H|p2wDI/sXUROP 8 o?c"}$vd=ś1#b iM`3$kUS p ԤH0)P&G E7%DJV]^D@ҎgڡJ~_6h }"Te-KA9p?@Kn+5u=R#`^^ϒ@2)8S&ybd]<!!b}svt |m|g+()5T ܡt"tJDn$ 1ZExq_@)*d9s}>;`C)!llnNۓPo-ÍRh|v&/㣄?> iV(ӢD%)$'J\W8|ۇqLxZ50c$ !pr8sQGQv1_x&RsCeW-լ|Y)rr ??0`9 wK1rG ȅjڝrH,f1\u 䴳FFC4RX xI.Rg:EKb  Wai )Q9 t'ώ|eub,Yr#VٌۡlA hO$MK*s^)g.i[~r$l}.ϓRqOݮa[Ǿ/lMmU]x xzcK ře~xG*mq0VT0 dC2bSDp*X ΠN9ɑcUU%%| 7#j6hl2AQ`3Cȡ7ج4]`,n>NޞRζZiI,%J()#3]fO9^ #ȯ-EvQ'T, CDZc~u^=׿&ftMnz6l)b=a8<1 ?ya,l㌞Ң_; A%YA%į`Wqn!: W}*۷~[ۋ^}x w`}\FMq婛mLZ06Wl *|!J+" ?}ed.79rfD* \]B_AojհTaFL7b!ᄏ,*?OB!q1wIt PA):ZZgI&姷>/i*&C):7KT +# aî2&C ]!/]Y聜ɤ޻y}g[ z&@VgWZ)L U~E5"A݂@ݨ?8oR!kToJtRw'ģɹ@g{,]w}8h e'IGLuYuiT\Qx=΋-Q`!'g'~!q M"3 :U| \?j<%TZW+V&)zpq\Ma>̦j8F7S ՀCpF\wTWūs8v5ClvڢN;j2[UK$D 0i D .e(g >tڋP<"vށ-vUp,*Ckȡg\{4W~YDL@/4l3sJnRe*p*<$A%%W\J+ˋPl=PWBsFhA\c]7ZSVm4Jj~7k=0,W(Rpjl;PpuRRP *Q Pi`UW+`}M1x?nW %޻N i\Q*ƪD_E݊YbiIB\\CL)B̶WRWg+k.ʁrpru1`[mkm;@!\#^c{Z5kʕozj᪞J2ZWթEO-Ѭ$\`͋5 ڎ+TIU3X^@\\KUBt:C\Ao7U=+d)BBu:C\ k 3I}jv\Jc:\!ap(W(רRpjmdJAUI)\\JKeP%שׂCMK". j m毑n m2fbv_R/] V3[Z~S ᫄IEP ^jsS=75 wfÚi[J31ab,)fjOŸŀZm[? *]Z*Y[ ֢P[ TMq*mYP*_NUO+x>zroVѩqT.\bN-zJQ% TA–%Y)/TnpńjN]z ߥǎ+JB=.ZIҖ]zPB caKOtw5_=QM_uCsM5_S ZQjm7@M5IA6+.WV+v\JU3ĕ`ZQ@0cXW(3oje+TX3X*{ysq-W(WRpji;@mM^WYjKLbprYfDATix3ĕq^P*g jY Pm+Tp%n{zT\ sI̅F-eݻtpy-b0rE1V%E[=u<\\`KYb@Tҗ=<WҚW E/7mxYKk\4o|bf=o"ZOjTxS F,W xt׶ rY PmԬ ()W(Xbpr.Wx[ݥWO%pu+ \qa ' ZV @-5mZtz\ 5-W(ؖ3c]Zz\Jq%z  b+TZ+Te3ĕp) rZ PlTJs&ت!(ᗔ\r<[#݊1YҊ(洜%ˊq@ӨRvgiХ\?ǏT;xϗ7Xx+Hߛ$BwI~&) 8;>!ƀ?͠42?P8ۼ$V5{,וy|?q " O]zzϯUw&fp$X}?[?z8Rܼ"sI~:;!]MjUz J7ZNpoVqnߟP#D q NB*"u6j9 "Pm*j)ѬW6$>[Wyd̆gISLx?#@^kbB 19jSzf}?ӻg@xحý^=I_ƴ%8C_= md9d>OGi*ڰrv7 VȰC\T(WĄ#|]wG]T Bq\98e2B#`CUS"8'i&R(Dh9Z[͵bB(0(m8Fy.x#2XT(:51Xb(πQ1H va6&0A&fØm!bcT"ꆉ;"vD|Kr \8LHwrx* Gx$:(RDqSkUKU8?a+8hNpޏchit؛ȪYk\p:Y&5|UiX{>Njb XQ\A͂* Z 3%sh1oaސx"l?FCgx 08]7ޞ~wQ7e3h)avzZYP) M8+]i2͂Q e㝇vB;".q oO]YX!gɉ#XО(\rB$8X`IД j 1 24Ja>5g'gJ:S _,;` 190wMOqMkѣ]ӄܿ|8V޳GP{$?lT `8Γ4 :MxE e9X+7g #gF &^ ӻJMG eTRkbRwЊ g}qaj |luЃ駏09y J_>U9<1'DŽ %iyx$ыDhc&P9ﶒ8~_jmZB/SUܣS@w@jwUcj O?P>ckލz{O]xf1X6q;1 / ؔmvz'"Kqx*qqˈ}rY;}4uŃbvO6=mݶop`.|OrQ2ބubnGkcןiW=L sA_]C;R?28tY:rp1,Ș9tW/,a5rPvn>;0˥t]͊֫/,ikaOB?l"0ƻCF/ZW/]wRԣ&f1YSgiV99p!>uKe*eIp2yDw(D6b'!%^`>mR6h"Aa cӶnn~nvaYUG%&|R+vES%i|Wtk/#u[wl>zG+儭R;+W=%\jA^zp#W;P7@u4s`)xk"&DFE+9o^7Å~wD 竴u X1ާh}@^qoᰃ!\2 qGe?Oxw-T=]-Zg,ZK`WF_[ڲW;:vYZh_҂ofn1 ְyߍ _Ҭ/FcyՀt"aŌ`bγ>(*/lgb=[t̋,h_ʚ?zD'!m&PbyTG┦ZR-LyDFC't0C0,f2Υ o#YJ6 0@=Ƙx Yynz19'SM'N9Mғ _%yyCi~˾1C+gw'ؙ>odf-k8/fwl3?{'8%7n#+,)U}0yJv&,BbiF;{&94K1dW7U]%'^BJYɰ Or(UR=8{~-C $CA H=I cG]0ntxuOxް(9$ÿH$Nq +2r6͙RY k M9Aޡ8j{j(O<ēg^d". 2L%ǃ L3MٮiRZbD ].HCrY[ ]Lpp&.\ĔMuӦ G/&.dx8UOZ E;Kh/;xyqjLY$1r#Bf!NjIH3YU/i^\i>JHL451'(w9+3THY"!(OڔKXN*d5r̾T ;>tn U fxu *Έ\IZ-SW Q ϧV'Ceݺ /;kWM7ed湖à_݇324~5磊{6/~o.~iVF]<4'ß4]b}\~Ww- Zinn7捔ȓyQ+&PA5JVg+kBϜxU~HF xMu4RƖ6xcl氍BI3%chiA@18k4QNQbm QoX<)tj7gWW)y@G~jQͰ-m|{(-U@ &x\.hG>%9N:]WQxØT lu_!M'Zt=4 KӺz+Dl.^M5*Ք".R0#)%EyMx8HvDoܦ$UqBUT\{L>t޷! }SVBjQ}75﫤٬usRႂYOZ  lh1qccB/Դ.ٴLTwV7>6=.唘ke/]]~X-73~:NU, ,I=]/-/wȴ!|̳w~d>//GǗO##{ rYJP6uq#cW"SzB~muW_*PTG^}7q9 ߿!~?w?L_Ż'Q9Bf RoL¯`6;]57kFu3kHC^b}*eO܊8_4^rXGortWuZff]j>#ܓ0yF6_ ߼/2̈)U:T"Db8#Au<~'PK/VGƝ-ב> )Dy?RH") wT%@1:fQUB$ӑvnlohp1=*n1d*)EEk .ـn 3*J HʕH2tg: h|]-X;Ckkb@ֶ2̟OqAxsMNvG٠YkaU綳 ^dd?}}è,5\YBXK!.j[:)@ R. oh5#=(Ց 8&^1F1uH "*,U⤭%39O3 |oUwRw~^%\݅eNݼ?ͱCG{BG2;?m 61_ 8xTֵ;K `Fi| "TʩFGsAPRj"a ܙ BDNHm.m 鴼7QdC#f:+~/UO6ע~YsJN:|ݢh< k{?XYu FP?]fm$`Jjw W}WL ~ ?479f[כ?$|܂J=WT[Q&LҰH`aˉc~!Ƽj1Tr-Lj*20c$ !p ,wԅDVlֲ왜6rHc^T4No"gZ<ǺZ5۷&%)%O.]{ޏp?KУ>?/5Ly`>be-՝?j\Rφqqtmf%1k_uŸ:zc<_7,mJg9/q}4I\¥Ⱥ*-QঠJΚT|q:+%H7ۺ[ ɔ\HnUK%htZBn3θKi6k43_ٸ3ӝЁ5 }P cnN-K w䡊E;ꭢ͙!f'ZJQ+O6{Gb:0JHkf8Bn'J BާNgUA2țk#8s#ԬF[ ރF%k u}gtSg| juτ;uUE+FR-+wr26z=򰡥fnYB Q|n;*Ht=) ;*D{%#vtF|YS(P!6:SP[r=JNwh A^AkaFh^v݊#:CI0iD(QTzsPyUoܕA[,X]A8&_.+:)BfCwducП^=VFU,*ҿrIiژȆR_9ԠMEX m {Ze3@PԺ ,>b@?5:\vAjc4)@F&E0W CQ)Nq'j! V۞dWNc@!7CAdܡQ l;󭋠@zJB€(!2]T  OQU0uk)& AwŜI' sG!!.dLN QCFvu(9mL6 _ b7ڛZ(Sѝ)E4GRF U B!Ф=K*%d@BYoTPcݪ8wRQvO J*H{m(`3Xo $$ eFjuDtPL 6C Zb@$F=(aL! #Aw ԙKn R5xlPBr|\1fPTԃ,0F!NHem{ۏ? ]aPӦtdUm jmlAhqU +0똆'z$;vuM %tA\xO(z+R|$Q*LFȼbKÙ.Kc4-zOG3}dH֨Vx$ ;< cx6¤JrC#QhzBb9¶l^ A; IwͼȓB,UǑ+Ob `/)T*1wP:6h /&sP6DR& +e3nD4ePEx'@rFEK֬kdž@0,A ?r7ǩ ]e7؀q{ 7}y hY޺\==?:=W2b5 Pw]< pff=+[{ XQۖU-iVn6fY#e-6n(={4\}IeF٤#väDy ֵؓJC6G=T3ʍڛ-`'r5wn+2TJ`]PLMYځ DGCPCz[:T}f=. >[/`u|FQ:Ԇ&@u#']䍡"UA  ?0ˢ#ɡb$UcbY)t lU?<66 ڊ[Vjҡƚ4AUJFw;ۃd &rZx tmKzwz] %5Zi.C*mP zx7XAҤ5MD9 Z&PZ #Wfυiq#0YJF$=i9\OR- Z-1˱XTZI0Ppͻ %:)Z͘-, oH\g'~b!(E%蕲L0bӸ2aՋ7Bw2NӋk%Z\Ҳ:\emK!P/ -g8}(.~I=%?`]BфfZl>|~ϟӳ⋟Xd' qm{"b\|Nď $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@9=] b;Rxa'ah ݸ E''нCo}Ŗ8 4=ҵ^hN8s=C׈H@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N q=Ϝ@l2,i6N T6Ƞ$LzN 2ZqH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8^LiNN L٤"?'J` DA@ 2hqH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8^z o5V ]{o}z &~d\`c\Puz6%uvK@eظtc6]1fCWgCWkҋ5ҕ>4#G=&+h.tZ}+FE]F>8S,+k\ђwb]B`|5q.th(zt=PC̆5s+FKiQ ]J!F5+磮nR;M{OW@_]-ZzjkukxhK_!f_o3Y1Y|, -;tP[8;XjCocp6qK腢CVZ>ݏoik j|/7`Nzͪ?LPW.Q=c?-1/٧Jˋ jkcK 㗖L=x %y~z/|i ѨǏ4{ j 翞T ?-:zr$> }N9nşo:H-8]KY͒j˄XMv=IO9z}8Fvj~qX UxIajB0}0+ba>B\\e4TUZW 'SBOH(Qmgfiܖho:fQżƥ9ѦȆt7bvQ,ahmw(T{R$K)͈|*Kh.th}wb[;^]!fDWZx6tplK0BW(E 7ͦ$>^;]1J#k+'>p+;g!tb]B [7up ]-)wb&]F1l R+~Vz!w<͊sh4}7OiF鄦_#M'(̩]mo9+Bpv40]f`fl ƺȒV-,_[۶ݎĖ)HV=|,|VDQ\}썗{IkN?! ;jJ W , W+HK IyjsvjR\\>#Up.p5 4up |b Us,cc;"pvNpkgW(9T~Қw%͐B \eqٰ,W2BTW \ *:'B59g0 \*KUAT+0gWY`y>+Y\% AR bgW(ll,-?yvVp>$K{}hI)C nG`E \eqC_mR5\]/'L^s8v,~{i?iu(~R6fNp;:em-hP^Лi_V{mgW?i.vT)x^c)yIFEE/N*UFeǡ_ƕ.Kyw#ǰ.Iẗ%lLr-nr~jBǑe6FO/F˽[i>N,`vջU\.{XHoV4bZ-=N^w*Zjo1Tu9E+KeZx]ІHBb.*NwIt`s F8P!R[֧_,_kF(<<NLUMϳf8\ٹ\Ƣ=VE9$7tyqKcD wclowi45? S6x5~gemWY.hOu#X_ߗX]T=e&CN;.?Dqz;z橆Z; ;GúafӛE$=<71TwWlۇ7˅ZN[̱n0Úŝp0SۻOJE@b) 6haϹ,G0lR^@pig+1;@ʜlX6sAo~]> b.zWhWUP~ k6_-XêH^+/gϔcOEtA/.)\rp@x#|mU_:Yְ`5 >(VPrĊj͐bW8hAtMG^BC zB~A{ZfюƊ1⊊\bXb%)CNdc ^Ӊr&PƄXtAN%eL Ogxp`>Q9T%Y'jxB+hDc|"3LDLQ1٤{SUXD̚͜pl\]~SWPQߓk(me@?P*mQ }o ꆿߵwK F,0bHD*Ŏ;,]8Jڐ :ftǒ}*$2&n!`i"ROR\k-g}2TG| #C*4RYV;oÕ^̼@^,S=_lb yOўL>Lm(1ٶ{4J L=o$1_=_mWaC͞|sMYIm, ^zR8E6-c^Wsj앓a>{BiI40TqI liAQk!M4I*{U1HT4\9G`Ta'eښ8kŎ6B ۹*ݵχDFُG"HgDm+RP2(ѵNje(1ҁ9|G4-#wh$ <5GSh߲&vCլ38-`dO_hkVŐ|=yBVle%b-'U,ʚБVQSN':(6NΪ`,qPՙx.K2)ʫӖI%C$QE@#pM& ,9UƴȢRa Cb@zϓ`yUء!B3 )Hf AmdlMWi [ӌCPW¥DňE_Tts}gr} ;/?ruF!1{)nC~D#I$$D]C9gg%,oҠ6T^r'yPз#$L>&n;z-q#B<nM;Em2j;I=}esw/7iEc5ګVʸ+RC4Vg tJuy0tp`p`=$n\3|JKBPL0c} XCidfU`>b"B>kbTq2*hw#:_4YόnZg\pFΚb|ϓݖvSǽcvsSl&[zm1/`CFk3-SW<6bibVKg 'ѴB(FJg Fsdvg<]zBO<_0i24DnE+CbJah>s$$T I?CV'uY?>9;g?2rէȃaFpDo1Z)Z%Ser\ɩUAZ;G=X~>klvWqdhP{=. &-6@<<,EQ g_9C"i_zAHZ9+oG y)IWf98Fi<Ӳ9P$\>.\\HÐ]I]Q֑-z?k[œثyZ%\/4Y)l<@*;L]|]j|7,|ЯI,G ( ?!eyb|Mq#{Ŋ &UIUɕ?EQr𱼘k"rvR%p\קRܝkk d A㨊W[Ӭvarp)'?إ:2.ej4wU cSï˛^eb|ׅӼoζZI7:HmuJS ȝn7yuT[P ߛf4_P#_2w.]#1٨ܒ4WӅL۹[ _YƤ5 4O]Zfin;1 e&n]ԡP%Xe;]pmt3 ΀[n1'ꖡd(Ya/IJR$$nE6:t!/zS*|b#nB "oyS K!:MΓvZ[$.r!I ɜ"**)MbA@ǢɩКn“Ǫ3>)ባOLx; H$ht;* k.A&u0V8M!YLGȥ^5w‰ wP-JPQ 47 - wG!d@3T mI.U$4xXeU)̀@RQk"s)h@eښ8;܊0)M1Γ0ƌ~d౦;L}\3#xET9+gftM3!@!:A9e"2t&:0N [%-T<%6F` {3jc޵q$ۿBn]%7Y X;X`CH"~^ÇE5_ {Tw҉Vh6h4^c⬘^zi5e4E)&8/K(W;5 Z& k/6Ϭ h$M!gA55b>Elv.HeuR/)֫zbT/_:ekcLjc89r&JFHFkQ/^ګH~㭎푎]P}ۺ%ey[% @0 &(89̦ӗ},jZe"g] .rAAUITJX! B9\9:*hV\J̅:tK .:mnF*^gDңNik;G߽/zw[[ZdyrmO4fm\ԙWtaz~ܙ6PsE-pYFkGjuvR +@mͥRxO߀+Vlv@U3a ul]M\)ͶihՍ{6_LN@LFOjf%%LI0& .k%J'Ȝ[mDy'۷pbc,BW7,>yr}o?@0J's8M뱫yh]JTŕLL".RŵH+p߫)h>*B qD<.3~<6>uSz/G<~N9|YN 9&NڂWx1aM۷4~fevɣ'F3'Rf[B>(ژHW \m q2AB.a)b {j GoұC{77 ;}ܝχ~M.gۜpa '>A8qbJDqy[ӷxի륪ѲA>5ugJL!U,o+o쾡FlK}-YGR9:G3 Ϭ`0=s{P$urBsPVR|L.hQ*a2hS"9Cd&0p.3l7&ΊʻP WN{L1!W0n|d;śwo 29jki=Ocύd<#Ֆ BJ8<$ヾj-_6suMȁEUhhfZ Ԝ!SƜmB$& kAYV-pT$l.L<9Rd.`Y B㘜*O4>=2(ޯ+A>U ^XA >@(5^;2RfK(l Uc;w2[n}慠X"1Ȋb0hrkbi1f)pBhiM1y XH*Z6ٚLSԈ \DVZGI΀=IdQ9WRG0ѩ3p_OVxt98FikS6JwQG)ͯv@-ڍ +sO;Kh~CT_ÇW= 4f&̥J۹HaT37܋{[]Yؒt ,oHickD#XFŗ@^vGuN.FnuE.+8>S:,4ՑҰU(0L~q̏x?6ו54'~?__~ϯo~8?|7\髿:}F`B!\$EF&cp^r\𦚦MKP+4t Zd{G}SvlɭHw_?|:rj"^WHlRq]w GRT(Pͤ 6! ޕ ̎,ꍗLNxuفzLћA|w1F{KM`$aU'!&\6F@ $DN-Fz`U|ǀÞ;k:/\"!Z7C}\E3H.w N#{:{U9fAppAu66\wzV\[ͱZPF?]oPib:::J)+U(ȭ6ߤe-`:r 2#53`:{SNŠ19RfVBʣ ^1Fոٱ8a^[' O{8ub> zAS.ӱGϦh KB R5!G)p_v%ZS@f,TI*/kFPXp7J,wR@$ :H up0} 'ǭB#)5\=.ԩyiٵկ!1Řs1h厐mUܘ䍏`h@Qin1 3JEk++K5[$OF lpK:\3lew/[C6ؾm<8Kq(aY<uS*0Fƌ₡y˔tV20YlboI|: >ZV.DCBAJqZF~05NI\̡8-҂EJmgXTi.; sUV`UW*ZH 5W\iY*oh\qQ"i EJZs ͕цku@ x0檈PI+8wsU5W\Y^\m~̻d*WՓΫX]Q_0} 1;WHyA/g]q0yRQ(2VoodI``tW¡"{\4-~f쿹"1WE\\i*R٘+4 H^願".bbcvEJ՚o\ኯcs8q\=JZvtqRr_ a5Wkz87A/uQLg:w5,Ӓ.(-g/v?*JNIϣR[`Pʔ*Are&M;+4)QG`e17??yw's?S:?|J^trt_` A@26Rȴ2Pg潠MI^Y+ƕt1lo/`6c˻(>v^Тqڸ}`HإŏnG7azufRߚɥIR iŝ)1:g1E"ymsd+zRo<ثKԚFwZdqq){ ϻt'7&qХ:Օ} ]U;0S>sA\P]E_:zImeS4l<ui*Ӧ>S"#DFβk2Ơy&&eϙ*a& 68D0Fg. 4^FY覘Ϊ^e7m^e"S3 7 i>FpwkJۅO/z(a鈠.B |tCɳAYlx2ŭΓ"l"|; 5 ax:PtG]Y9k˖| +#"a;K@pXnw/AW[YҶx`%[Ւ,SĂE٬OUX#0(2 e3QKH R TNΤ Q娀K00g.#d}kL~gvUNjl#wcs3Yۻ%«բ }g_8XNHg 'QB(FJg bFGtA2 Փ;ݨlr2I}I~>#IW\@6՛d8FA\lxmI龦:Qb444N;˱=U y;ҽۥur/li9}wvp&Ӣɓ1`2A6#&,ܙ110m}N =!җ c@ .!ΆgL$:mnp܎Y;P[?8fߑz&b΋ގ:@w:CwC'!v˜y_5ag wÏMv(i{].n+(YK72<3> v?o0̸3h7oɕn͑wzzs6N_OtN}?N71n>6-O~}:ş e|O;x_~SorXK(+=T@1Do? 3S3Z~;u<*Faz51 ZQCX7qohHw\j?t=ž Rj~*1ynj}P@u]",iqOu9hH~`U+kD O[Ƴ-m޽p#].eK@,R7WaXo]ۗ͹-=a7]MYLi"{NGzCGKOæG^?nw_Czyy٥oSe[%%IIdr%"jg2r,cR)kЀ!qKeޓΎ, T}y" Dv}s`#`r6YvZ[$.ؤ΃QJJxЁAY8~żo-[zӋ)ھ1<߷{a^xXu̵as!{R.h01B_pSC2H(][U*򡄡=mYv YW|Ihs]ؐ=<P}V5Y1vNb@Uy}wͯI5 >tGZS 1PKwQO$с>i@j"NGa]w_sp>+ِ+B}_]` wr0k7G>_Ɠqʲe8Wݬ>rfS2f3gbZ1cZ7q۔ L SbxbJP sL:Ϩ#ь:5 eCq.@N%eL*$ 9xZĜS)Og^2>Hc~37K`Z*3U)J{ItH^%e2+a{RRX.sƠ\RŬA8o`ٷ'EQb1J&ijb (w9y3 RHF%S6L'c0{-&ΎI7B=WƟ}ẁHR$XbS IYG "%,2gAوU4)dQE@aHG%|Is,heDǵm<8MRHKc['{&z(Y\}|ߜ7' 6aRV<[.C".d[T巇@ӖK%GJJu-*4q̸G{9bO&0+' QYT<1K*Z!80G8@ RD H78I%!EϐEa˟6mff4؅wogrhm,]5mOˑϧ6_wWs6:n-\Ʋ;7z2buÚ)0z.mnWYusp\yV6Xauj9m%?x#MpUo?m(6Oۖg]1@'uS*HY<W/r-ؿO ";bZ}4WJLP߽'.$q o&-J)4)a%p^4\TeGo1;sx1G_F?/,Wm_dSqE`^[[ZNs?$L˟@m OlXNrͯ_5/mBtA-Z)f~Yg-boà%mEAB5+WzrY߈:'>5A{EW#;m6}_"ZHe+G 4w6渝L93(\PA,rzHk4QNQbm QoX<)t*[Ns< Bq) bTmAdHfEy:鉶6SXLc:@/G2}7HMg.҆R<06PJ+{-túF|ze9Uݯڑj*HL6OZGy}cj6UauU`r xSI !ݖ!:z^iD)Hu425'TEœFܠ1OgIʸH 5.;RV)JL8oOlQI2DKEN`"&)Ĩ5@ 10Zg v:z`'\qfIhRO>]/ow{ 󥛦w,Z%Dyԙ!|TR'wN(H"|N]*deϔh*)jcVSଶq."-F!mpA|g'[!5x40Ƚ*4=G*S(s ULXUR[4&vB. o,WDHPywBǜΙ{@_Ldm2,D:#*qI^5 ]qu,),!\HIԤH5(j !ucQ )5pYX}v+uv-u]l+{78zw"WE﫢o~oe9p5ʠwl=z&JQ;dÿ]*]i"uqf+I S)z**>UmtkOV]&h&'+L $pBIQS1.H4GCLSڳ}ڗj[C݀󽓴kC+m#I`<"gzn,0Ƽ8HHY,%xW_'@dzo H.2e￾\~7)u&~rW "`Jn/ΝHp-!g\2qtŎ΍pt{ߎ#RG_Yr@\zNK%+arsr7Zn8}A?*s?ׁ)vЗ;u)}\Zwԯj-trr<:^JJuGahMM#@Ԯӟ7O\gsU&|ܚZ_kտW~/t3|p 3b.n?㵕8_H?v3COfKؒґ@=(-],n\E]c'NBd_ w0yDM?RbY*\ZX@ gzd9>^2X?Yлr~G'H|o_$7J]&$l1B*'<72 uJFgEq2H2ɩH[w6LU^UK}t8l!$ K tdX˕.5DYU۰ MY#l,uR:JQUDq* SOU^UFrM}L@ v61&֨{!2ltJP.%Mm PFk4ky#D23Wek%9",^A=i L~ߧ>*` zu,ۢAZ;LK;+V^pA:q4w<:SW|n@VuPeaBXcFۚm˾V{Vgk3)8:3 tYJ}3)EitRjv~5zK' ]O3W} bnF{Vꩽ2~_o}90>$3FX/[$8r$ƢΡ4lP*{ꄜkDQlBu}> lN# Z:*kIhS<-W ׃(/ ޝ];Xj5=Ħ GֲܑeBN7=<OK230[^A4*+Kǵ[KyK6OபUR9æ*gecov#l8ǣngm'n} 7kh64SdS%& Zƌ“yCi%*  d#d >~X8;9mޔZ?Mފ|ж&nLi6 p6 Jf>lDfڰ0lZ*#;uU0v(PվBPՕCJ)QW\s0I1q%}VWDe~Lu-H&H]:uU=uUguETV] jx/عr5ɲ?.j?ݬ2J@%O/??~?[z0i5{rS.N>QpD{*[͵ͨMٽ3t_g@4.xD 2v\?);觥 rsZįExd$ OfuflcRQ) T^~Ƽ=0-@&ڕKiuNi=P?S~ĨL%W֔?7ksBsUZ0qb[BòJe7˲4͌{1ŃUITdS }TZTBNi_r/%u oyrke-yy+N_*kl9dCkVƵPȅq-j7],j}BҴ7?kZn [ЯMB5Z UW?BBB=tEn`l-uUvUr!UW crV ZΕT\aaKX8^7Rm~5ZQ3mQi PLoC.&{∆{ɢ @4[Z.,]Mg7,!AؤMxK <̃u.ƨF3^gPՒ˵/@on'-N>tNW;vË_r>KjsGެf24E`[{Ge;>BAOIwo Kx'8@TItZ !MN-Pն}Šmڋ4B*0.*d ^։sk8CFa!*IQFj bv#'Te'#W u}$zTYiTen 6F e zyz]^M|nMǛ=AMÆGq96n B}J0c"i*hj̏\'׫B;IVUlԼT<XmR(KZVm~pE?h2ܻJlhir&DGh ̆RZ8ǣŷUդR* ˬSD18$.2M2Gw(*a9+Te~p@Ef ,s Arfʤ"Ȝ$cP:ȴJg#|0\Vϊbr"2@$g aAm5_^gݾ.34ɭ@Pʓ^U$t٦}2ve;Q1}O*v`8eC9iT+?ݵ/'ojI2u$B3˄00jFBQ9ih1K.HP`Tؓ"1t qeRRMkyZ36UӅqKujXVnT(*sd|0Iu*v0L:>* wnYc `FCD@ZTP6eyF2d՞> @fgDC*ƞɒ^'ڔ ^C4d1Y9̂MG4EqwTk놵nv` 4`9#Һ@3Qr04Z08QM~ CNjPGvp!;ld1s7A̾B9hE1ςQxtFk%ZX)FRadp͞F`!U` @XY U)#҂FQ4`pHa1r6t[xؔx9{F1vÖF{@Ac% :Q9SY%M.8J$N:/ DFJLk>DؓDHkZƭ`hÑi E#Ƞ(+HgKʩ6#&c@8DEYKp3"ا\dペdY6thd/|Ql.!G_U3#D\q1(d6Nu8(˪ o'#UJ=H$w~K/ Jl_9MS9 N׳$wr~KQ) 33Ql!mݷp6 /fi҅ EϩGr:)o ]=,ty>LÏyua`ÒIrr.RCJ'~}(}?IWPt6`UsFcJGBN-q!@0f/gbt<g5wPsx>{ P/\pʥq2淂іB ʊ:Ed:Ew7+ZuN@'jc(p~S:*r_zbzȑO^,5TnY}VPPe4*Ƒ؍8`K`ByqYƜ l(#ҾYP`rHTiㅑJ Lc1>@e'^q< C }z6I%%F~]Oݰ'JzK gcj,QT>뮺(HT$ٯ>0RMZ tʵ SRT3ɌN]őr&La%5FҎYlK:eE[a}rp c`E,M\n;ƌ4`>+H x'"܈ \)6a+.sZPN:8. åhRTKQ25 D4V tXTnW'WwX<pRB`81`tI{Ft#Ap03qp0~4^S?m%+ cPk|"sR9rg;MyZMPv={Bg\DltX,gJ^jL"LE}·㫝aM>MgT?nr_jvRkv_P@|VlYX~UX˞uKK⳿W5P܃ i~@.` qh1myQNbZ>w{ <6\ȆJ]?W7{S^*8wλ՝?|Jwit6LJۅ7dҬt"P}w]8u2X RGqV!z_=A |=i;BM殃; s^ nOǣ~l C{ɱ6$Qs ϖ;Afxv)|-jn2KNZ0E6Al3-O NqHjlJaҩH |rVk/;4f,} *[>4Xv#7gJw懳pf}_IQgaM] }(<;a 9߽Խ|| /5"e15wj5us 7-~Î' rn勪l޷VjOB{T^JgbO 3 2;r -%& 澅R@)I8}e83 Ar$)A"LSM jdeFBoN]zfNjZ=,~D ՚?tΞ0IbO6Q0L 0 vnչ4,h65ZpMӜ:(-!:@N;C\3DA(:2b1%&r@p"VF :&"'w9lVJHH!McYAB$!E=ZL.E+rLtJK}_o'Z+8X.}B'2cfsIr)Y{4sJ1C /[xCb`r (DN2T3k);cp9UH n^LG eT%jX$"﵌FML%#VHKD1r6tͯ6l0_N5.e?S֧,vLodxXf>,(L ft]›w=M>Ktmf;D$G&>]Louy:/[7=sy/oڹu&2{rz+]mZ nYs\~uMU; K{WD,M7*OQis1CV?IEV͟F7sָn~'4o!OH>=Y;w cQ9.k+J+pG{gw&M{joEK;;0Gi ߿N$AtzdiDMpLX: EcD2Y+냉KM-a$EV 1醵P{[x4x?}u`2[he%jZ`WeGDѓ1{骇HQY7jݳb1Y:I040?ñK]&!,>ٰ>nx Xzvz3XqtW?_ˇW/~z:=^7ip)3^ BL»л>~ИআCS6& ƕGv} e[n@dg_^ g92١ `Hl/׬H]߯zFZR| iq,ͣQ5V:0jG<W j~]EL37/D4l]B9.@L<>(^ 0}YȒ .9GGMpk˼ H+YI =+0/q˂y.!X"q K-O N5R0m}Q s[ Ai̼, p#1 -A[idMU =.uȲPerfUm4vBB-b|z!ʖv9gMdXkT V0IR! Q0* P,RV]1ZQB$U`VKgbA2olN:Z҈l'=%Wux O9k=?Ko±.ŮuYG42XTIdp_ AbC#Zڞq0]s8H1p:3a˕Cyu)Z)W܂GI|l"@4 .f-:/'1bePS:Rt6Z | r B($]ch՞:]z4‚SpĽ=c ًf~~[!Λn1wi7ꙭ!=__0JTQQ˩EI@?EYR 6xw`5&quNSl71/F1ZlQAsgх U$;A (4٬+W1sn^~.j:ޯ<yL7vAY{|nO3.P s̕B<q-%E$rXk!OJD'&q9+v9ѼFgP8w=>AJ8OjYG 9Z)* ¼&BP_hbiJĨsFVO曐L8\JmVm79saN0DMi&Ɣ'mRSٶ mbݶ3ܶɑR$DWrH"\"SUD2?K1]`u2tR ]EUD)PGWgHWcAQBtt %s/%?Q::#bM)?ggn碜Orn0" 紏hG~l2Uzh!tcd!ڧ0O.޻ SR9M#\RVު(5h i:tG,!_@WT*Ut(;ztzCXSvz'x' 'za7LvCٶ%]ᎮzL(*! d*<h)i;]EvtutE@2% cʒT*tQvutu>tEJ'DWX1 ]EBWm+@ Ž8!u: e((i #]x#Xd j$R޺(k3l]] *۰=v'R3wpFUDK[?wQrUeBtt62ۏ"ZNWΔdZ}L@X3*ߡB,l0,+JqϦf2ب#l< k8Og_])>B#~*S[=Ue . &cͰ8^J^}ܾᆰO+~eή" UyWޙ~H̛lips#N.4fSԢO6)6gF\_im c򐞕t={]G)N^r H(ȝR 3"="Nzޭ/tu'U&h$^&lG\$41Ig9¥ɬ3G\}j!L9N-h f8]|.Mf_Dd*ݦCWdˮ'=đnS/S -PҖ%;j߮ǜb+L0I"\R"J;:C"+&+xT*tQκ:G\q+,M"\JSUD)yGWgHWL`FR"`WT*ղt(9R]!]qb &'Ȱ\*RUD)HGWgHWBpBtt*•ɬ FZF3+Ejߑ.!rj4QktPo1 46I#\̜]DKti:VDϒZDWJ2ҡ'3gъE;GRҜ'tq#\R"J&:z5tEzCR Rܓ7 톒lΎ@W}+ʈL"`WT rNW%]!]%P >}>2 ]E"ʶEE*$!+UD{ld3+T UәT*żtQ\"pL"ZANW]!] ^b9 F$~*+*Ch.UPoH-to3&r l5 )l<4xx 4RMK.n\.Dy ZYrm˃w4EF_n@VvAw ]3K0?MGÅ*Pք\`bis45AR鄡֏]6{;V+㯛:# 7΁SZAY|c[6,[hitʾ㵝У0V@q([V9P=RIUTve?Smdʧ rm#OaA (gqf#ҼU(/[s&Ye"="\iǁ P7]5h<ޏ0j~;1ե|ky"䛹oE o~ekh?m5g,{ h3?(:fͦJAy%׳ Y6.@(>oa߈WNa,O:M8 4r,/=!-`qϲpvG\Rv|;JÿyN/Ǜx<˝ Hrǂˋo^M߅"ce8l?o+_cx=)XM4ОsvZ8TsTba[ْNjuZܭn黁-++a*⊮8^l!y}һ}Zq1kuT](3/e⹰[o))Tqq@y|S/UaV:U);C¥.n „JdE:mBzdnenҳvҪ٣ڕE'o?-)j{_%9t_[no|[==ʨ|~9_rբ}&{r;33BN9_C> ,B|7gUCI1 wpF&}=C,ח~KqC V_#( u7c#;++?wϗGŭǻ_x^xi?Nn>u{IJJV,+my;gQBLs&‚SP3j\ pS#$5.qD,[%N{<$ awcw߱A^E_hr`)sVA[ꍶpoR0#4'(xA`j=1(_"øu p7&@旦0]A ى#g+_x0ϳ{*'U̮\ήo.:<μ+F{opЮ-lQ?\,}k ̥TF GfXAO/nU|x)+k[eq`=;[~D#`RE#f sK[ur ^M/a 5,k:e^|>-r~IJOpo* Uc9TDX wvPBqٸ{=_c_ PD47^4lSV3tnyUPSghlf-wuۜmlyg#B BIjxݝ}k>;+F]@E@ j2D`xPbK!B+Uia ( P~8!,($S@JkXQYQF$|ˁYںt⹗9 5j$]*Lu}Aoۏl|QI'ݷ~.;t୵Em()zBWA&QV~ل 0. F : 4Yr=()diЅ"EҁuVT}a |A1 .K_h `BaA ƜHifl fflVi }7̅£rQ6ZƗ!U7c3L/vefƯp8c^9cBK0LsN M$ *BcTٻ6#lddcI-ET %J)ydW;Ӝzrx{!x(j92Ll`;8f+pJL @`T1.&fƣD1jZmUjjv݌LJeAÄéhPOi“.I*mI2dDT g,%fxuS;ӪUcu%E(l]\7zD87,A 'xACoaCSLuTXG,~ȍg ~b}xvFQeok(XKYZ FLYGٜ1Vÿͱ:o5Q]Dmx&W糳|T#Vr)Vx8w+ZQG?mfsz< =$DMJ*[c@)(KrJ ^0-FP)-"Um` 2Y)<f 2s-C>^Џ)Njb,ˋ'xN%~ ]WoC_|=P+jI#xڸL^h h4aO!v7 0 ZO0.HČ$Ē+hGbTj.>Fa,PD4&ࢨaZ$|BD x+u3%-CsT;*EDŒa(&\pXm7AyxʟrHޟmVCնGLnl`C.;gtY-/I(g)hOʣj)(.D$>1<,YNs>'KƲ(Leh) ɨ4,ל" :"HQ!,pazpr<OQÍB[/5R!j~SZ1#`R&@%AY{SNow(mhѡ!MB=fvn80Y/0YA/'i.T8{fD ј yIbАՁHArmC+-c77㻄NB֡iwOy.ti;Ll~q:K&O?GZbUh<9sYɫl19~ɲV}X]K!w} }io/ͭZ_ӏgrq7MnJOխ[iG_Fٍn}n 6`_ǣ8mA?ebsٴNSiRf|r>\ nZ >Q;S3ǼtOf￟N?ً忦?vwF5O/xr6ox~Sg占G5W^TxI9I؄eK9!_pw=G\|-|~o ~bLVBOw{BB -~NO7Nߍ?嘳Bq~M;w]R(SѣCdP}-A޾PWw]qJFN.BTkB1#!wEM65ӳ?%#r 'hvFq;Mxh.We_Q~=lӊ?bӪپG6:N kv=Z:Պ)ڹH@HK jH/fyM+;'Y#b!qU(s'LN%hZgQʗ+QHxJ *@TKa)ϨH-c{`lL'2Υ>ʣ믮I qf+E W42 1 =bO ؓ 0LdMn *!)!A" .!F,yc%\h AD8~خU,vM8TwHOO dWt1O AFJB%8I4')hbdpMnRqL+omKݙ?i.cֱi<"icB<+moo83MMOvf<ėT~Drx0+08+TڊhxƂY7ZsG1CR=5T'o(CLW'|oI d.)Җe^Ԓ\FEֳvډ淚:mHߵ x S?D{.U͖6cx|Z~s7yuնv7u:9TM6h.~ڞy[\`gMIl>M{7| 75HuPYї|r-j#C<_kOO xm%bhG!/ӋiI֑p+!}5P0@<흍q6 sFٸۓq'Y;TKA%P*G4ehFHKlpzRIllNv~Xۋճn邖LAj-) H<8x`' 62&AINT˻N>hyH Χ9Eu/@EUv siww`a6O:W?̵.M K&o*cb DjfuT7}Su4v $ q˔Pr 2)JQ uby|ijEQ'DxuclP9F EMRHR*G@(HQj1q6:lIQv %9GJ=u |6o_}5[atY ѹU3!@ ;M ri>Xia w&J)co x,Xpˈjˤ\І`Hu8[eC^SGS sTv>;kl SJp'r)F!HB"ppP0Jhm$4rTcylt7T&$b֨h(UXjQ Ho% RTkƤWܢRS!٬+q \ uկ=[ץM^|7偘}[|ENgbJ4MZu*tq=C)cf V_'R^{9oQOyc;:Re^~هD<"HSED0L)Rn*=epFZ)X{ANJՖO#AR$$֕ڞ( dO8ﳎ,Pc-AB,+("3.Fsd#(7{Wqm~IF|`<8[m N^06%Rj%)c%JZb c.sfsϝ |W?E9lEhdƭ-s SɖBJ#Ŝ24uT +c GN9@5nM}}W-홖O=CйQB'eI'J)kUhc $}(1X F/gQqD.sH0: c,EM1b;kq̳#d-UCZ p:?vN<]͡u}M2a(>Xv.( <Dz;mS;nh# 3vQ-\q@ .ms  hxYqao#s(.F(MOhDG.O^RIB h\΢zs*dٕKR74||_j>?i:KݖXפ4w:]J_ް0o`ޭ5?uΏf}nf8[E'`J0t&S!iѯk>1pD.0LFKe(NAG;mϮeO]HX>32o1nxЮGT iϹ~N|ʩթ9-ho,;=:]h[jRNW;ۉ+} Mk(F_t$ qG>9QơRfnץ5]ln0Z/q9}׍tc8Ҍ'Φ>o1yi Wz["usKG}͈fmfX^iSOZ~QQ|2><EaVk\Zi]Ts +ϟ&'47T;tM5[j/FrG`_^E_>sD0٤>>O3ך+7ͻ6C4MKK+4Mj^|v =^sMK&Iqnt?؟su"j"#UUz4 `͏W N8EOWQ9h*BxS@.x_^ Dxj߷E}so9;V#  Zqd-)%*\cPRF9s[Q$R%(979'_lSv=~ 1=C H&瑾XU(>!Bg 3w^(8%S H.M̹ X:m L4Oώie0M\Zg`+2,?S^ϥdQ8ӋzAefٶc Zf.n+v:D} iDeykId\l"<>=Gytφ?- 63A`YJuWӺDc.H gb FP >Zٞ g}ƋqO -W _^ο}WʇP}~;*mϥwǍ:L A棩#a7n[ h")6ox)h} <R"sA]!&ۍJ}TQ58hplJB-?D,K2|d1tY(kSYFx~5XLJv>tݩ*z`ۗ?ڎa[Vx+DRъR˔"z'`:]z)I3`gdZCVg[g[K`(F|499`?:WAgg'p'Qh5vF_4_4@V[A 8u B>Ow:eLV*^fS]P쫗`w؆y1,E]d-7_{ͫt7ߛ]y_m:Fog'ӎ|@th hw'QvNkl'g' '_GuCL<#g UESۂ/gd+%ֆ*\6ꪢP/KS__6UMna#~!dzd]zvf5naZj/ѿ+^=Z~UpRcqʍL^-]mw2郪L>/̈́4KSjc:i6Ogךk꾶L8mudkw_l͜a1]"qPiq$oZ+OgG d*uBӖowkӼT>eܫOW*9\ZÙ`pjOѳ?jcS]4co3qӯwu,9M[a2mhˡi'fDCWOhEٽ;+ԫ&%'tu?} B+V>)]%btAWjCWz.n U+غUE<*ʧlCWUB]gp? ]SOLoKWp2wfjU+ߙ^mtx8_u:ޛJ_eww}jU'ؖ%aJ]]TL斱й}:,W>2Z:]+*ZzUvQV@O*2(yObNZkܰJ^i%:RUգfVۖ9]w3uSaUP ,Dr0,=~4 0v~7ggc2Qjh-sd\l^Bْ2yœ\GZyGȽ$\=?w|VxOI؇Yݹ󝓜Z,liG?עxs( pay%E2Vj :(sOY–oS0N,Yrp !QJClJ(CH"&)bX>>_;+!E}!yfP:D&`a΅.c1 _X&pE9ED%g9hM>gX] $C lj2FrpY%x'SA t/Z I!ףgcH>[*E8g(*2W2@IlȩgƊsIW $̘"hFE062DQJ Pt)RdٺLdthc&GmLʩ<o|a<$nw PH^1IBšDu0`0 J`!fL4JIJDW'g" fsa-8 +G@chD9聤}OrY}Ks"*i;c57 k`,Wچ@.,g} (D2>kV^zUPH(s,#Df,+Z ƣ5u!0: F;}mv*zAcK )>Ʃ!ŸrиX_!-dcH33 i5/H\ c1S/Xtd9E DGk#/&eDxAv`L\0o/BјidXH;S L@laxLqLvI*e(:-h?R|T*AI3X(QrRPyUNu(u Rc!a¤1#uP6 CKHrW4:%a'((&'Jz _, Rt `[jU,*I&3UR9|VE #, #҄\#mu$.nHCF\ZH+U2%x"fH57]6ǐ?31$7* Dbȇ+!$( ,*ZBWBE(XqXD@0 I6S#Ks4,z0A0`𐬖{)-ԭux#$nAd6&"3*Yȩ Տ/XF~ S,0mF6b$e9v,pwݭ^?3u"N̼26Dn ]{ #@xuɍNKy1h$!3i]%@_[0YTA^IQ3 I-`kf*@FgGFKAFXՎ f8`i@ 6 lTѤhN GPPr#شTcwv1$q0E&5"eEYME-u7Fd>UuԹUu֨4

_nJ5vU5Fen  Bxo}?^48k$M0Il^:{H7}c$^{2rW|`HN ^r8E@@rhKN r@"'9 DN r@"'9 DN r@"'9 DN r@"'9 DN r@"'9 DN r@"':lP pg+{jKNt ȝ=#'9 DN r@"'9 DN r@"'9 DN r@"'9 DN r@"'9 DN r@"'9 DN r`@A9n0N +P@V;:D'sr@"'9 DN r@"'9 DN r@"'9 DN r@"'9 DN r@"'9 DN r@"'9 $ Ő@ 1'fփy'վN @i9 79r@"'9 DN r@"'9 DN r@"'9 DN r@"'9 DN r@"'9 DN r@"'9 twk_}'ћ3jnח w?rRb~3L J@ɸqI `KV$An5T &&` *7;]!JY$]AΠ^ ])C+Dk{OW3+ӽz@t ]!\;d'3Itute f0tp`h;]!J+mCZz"5<\Zr!fǣVBUU 'Jh1=r C3o͗S3gp0?b#oJu[!̀h$W26F\0DӇHx~Mɴe6hf9kct3,AEV[/F*FE< grWIIG߲ͳ 7.)<49Ϛ:]7Co[*FT<;֧@KAݚed|,! 9ߍ&N^_op5RبMC >;'ԜhD,yz^̧WU4NGjR7'Y! }N>+7qmx :79JQh[Ss~5plgyFI$D㎞Ȓ{j. \ie]e4ڨ|u]Vh{㴗pfBL(U\=FQ})7~eSBKb+P ޯ3zZЕydӛ#f4j?j?0Ck_ic?g f2DWOmznC+l` Z=BVMtu8t%,Nм-` oNWR;+u+` :3ɾsGtu8tU=|`-` +DkzOWz+hw7(u ]!\ˇBWL] ]'2N2.CWW T4wute҂ׁ1#Ƹ=RCRX,1 \>$ъo@}{%4<~Ճ+Dw<fHWs/17BDWlz{ļRx`ck^y~h+=5n/o{Е%zjso+` ?BgJx|K K;F]WoJ+w|0tp%SC+Dl P*4SV6l0tp_{{~h;]!ʾY^  ÙjGr0 սjQz+˙BT;OJiczgc%ǚI*&;wQncޝdr &zί c&C|^+&M'I{?%^hUQ{S(Kό}b8rZ>( ZHcDWXq1Bb0̈VTp+ȦwG{_ҵ`jOpZ_}R=[q{Е#zjsU6؉f(th5}+D)4&yg ^h%OG ތڿb}CBߌVM:tyg,Irvҋ!.~kjWp JD\" :=[Uo_~ M8x'O,q:Zތn%ӌ{V'^-?{]e:LJ7e[}91 xՕ/KkZ> xW'= ?s?/,Vww3`N~/y)w}x}O\?ccu_[-^c㷫\pZZw6]L'YK%p~o mɊg+}j-\ok9O~j/Vhy<Y΀¢4|9:b Iڌ !H8OL<m0eN*Ugב\D.nӂyl@_;< ɟ'38&yun& NYg齍IZOinGXF9̣2-<tPѡ淡[|6զ_٧b8macWq1nGŪ[~${&ПVE[Uძ3̧m31,O69ΎA/J 8NJHwnan,h&s:Y+MW<;t~|)nqC.ǛZlۇ7kh;lኣw30]ɩ<ܳǣtēӨstuM h&|ˣ P(Wx# ,1>Vc/r -})gzgz4:~/A֛h.r^(a[RHZz \#܎+3%~MZXY& `BU^c.A9k՚sB%A+%F S)k>sPUNy1s|xSy911g kJ6v,;u)>)@4|zs7>ޭp*'`_u[wc=Å?q;T[κ1|w&@l1ڭ 멵n5qQbt)`Vv"+UյmWJ=ifyqċlX<%X]э7T @)HRHW7X듰t^9rƳ4ʨx@ R٨|0fgZܞѝW[J(*sq_JH2b\Ee9{[hlcV=H=*g>GDNfOM3e*X`^InW0m QςUI6نU)ZM(f`DnTNox**B'I@ZR38_Y; 5q2 gM^>pU}TJ|EٱuuДg.*Aj|@L.M𭝗\+a^>_\,t}]X\=TD HsYL:C')Ǣ,l"s0_wʌj5tA*) 'pU8B0XPVIu͌:3vU:"\;BM\Mp˶M2D$,iX2v%fD|QET}6KvE0​fZr YbAJgY , ܣE"Y&0Caٱ+xv=狓]#7~Qw(Z?-'~$vFm-ٴl}F-‡ N6.ɔ@ЧM:= QHҟp>L:AC }NcPɓ#C)TlPg;EbQL(6T"BF}&~L:"!`TJHdDU<gE WZ(]GxVy zsGޟ<>N[,MtNF#DBqaꨬ6  >\^N`x`G f,ǜdeK\Pq]JH: ȉIADDeSfȱT Z)4)F{Az \DƗ^}NlSM-nGNS- Zg bpNfonZbeKDgeknjzs?ӳf台`feAee&jKP1ɀH J:ySxfCzҐG)|COZF8$Yi% DօV$C_4 O)tIon$# 8xb>'N:Q[RB!NshHSO/7QԢGg5C!C)mR2w_E>h%,9,鄉YMP5KhlW#"@ 2e#A B0~w5s^?ck4 oDz1G['xg3]q.njB]ÿq}z76~_K幡#Ćy8Eko7ˡ7_"f~:}k=zFDK!Φ{.իџyiWVez{0JѰEW&{林ng<vL^I׊诧wFћ4-|ow]1a6?y|vz&d~rƿ7)Pf>nElQ,TJ:W;742&ih&iœx֡x[1FOOp y=`d.Djd12 .~MƣoO7w_,j%-{GJpX_ F5P)Sts >mf9[=A] H~ k?qpC/K!mM 󢸘;8.s,tztxO\bi^u]#Tzթt<ܝ&lˆ,2Ll3@VQNyW,^n@XL?__v^s6̕"DaggםȭVNzmo=lΞlV^')Ha23X^ aCV8,W0lh_G<*V~gt}@2o1$uF:9yk9O^1i]}Ү67> uycgR}P*՟' OI-~=Bl5+hh|ȢQN&#S*$e?!# r#XTnYNTUHVfPtn"2"[ s&c*`LkmfC38KU#Ȭ,huҒ`G,cJLsVF  ^VgE))q<`zWv$G!,fzvW7'|ؓ(ǽVA*d0[I!+6!s21tRl*ouSU,)q!uk)ErYXļ"#0t }v" _RWhӸ_:r Ϙp8(69TQ71]tPHAH)}e~_}}l h;ہG͔(.V8M.6'2,Oq#3J)XDzMĸYdO18J*f0H[LG}&ͧatOl=t%aN]Ycj_>9uwk6'n}7 X܃ '^е uuLݲDjtTODQ2W2BuF^e"^#q6t*aHNg/uKK^ґ?O?5%~9#xC}:OWG~xXiWT~R&h%3%Ȳ텖HY\\U%z%1-%esU$REnembXZK G Kewvu/І:X=QF99O7/u](;G4Gggr-gK&|0 rpl񟡂嗲V(f@M73aSr>zLy_}pۿ紌qgĔe"]0腡CAbDHg˭vj,K.>p$dY`%SqU&SG"@"K?d#B7zآܝ-59iMR|l'ѕhZwTS1P')-+`%RRzn< Ȝ4ˁ^r_饐+"$KYK/ A*utøgS* kW^ڔ|JFg pe(URJ%.2Ct6s\JFzA8[/5lz2?o}z\<~ 98>ߜKرX=-k緸,]6mH/ ^_zv|^s!ݘkE2c_SoaĠ [r?ngnбw4?۸ql^v~r52Ȯ|0\ouvsbϦyȚK.w~åнrk}.WM*OQm9`?}Db_K?Oc[!nH;}#zs<Ԕ`a]G8RtLM7P,z]ޕ7}{ }'Y[RZqͼr[߿ČDf"0+#y G'r nc+(B( Q tV)KJ̅3+< Uޫ>>6ه=j̏[hm%buwק!/ ZZlimGTFH&k@T[*Y+LL{9Ўu-x_f]+*0PUujTٝ[lW0+V?W/i`Uދ?+e g٩V Κ$B@H=Qf=|k;X5dE \q@D1a"*jES 庙ylh:$}lK<-oiV8̑C i:dc!3`9&3"X^jS{T\o&)tda$a9* ªhHp.L50@ aPNXk9t?3E?b~oFHC0U"ɻDRM\%Jn mN+)֠4UzbN=>.]ꉪ؞-~`tvqtR *.VOA}IMҿ8#MNW(]ZOѴV&w~n|V[nWv8G?^*/-?y=lH lLI/t)9f6&^Y@i4_"Ň^ gweoܝ"RݔW*-eThO6 K>ߏ'_}5MqCl<}|"kYRueQ}gP8:}^7ON|sJ9}o~x=Ο R@G$($ n%?/~:i辒Is#HZ}zHW}fyCkͧò˭z?^}z> uEjZV&̈8B\b/.dC3] 0E,b*\ !#K X#ܗ$ŝ󣯺fێ5޼PH~݆B}$DYXgL!O^ )`X9eyΥa]*3"SAAʙDx! wXCH%RxD$-zqs`9_'7k'{Ë㽏|tbݜ\t{(ڭ{b9vÌ탍bpTRQG(0 B}T.KGDa. /-J&yL7J6BnXkrCX!:2DGgiF鲤R<a (Yh^. a7ָo >Hvi}xv-m}E 4c_p Zvܻ'%1 \SDQr) (,K(ct2v~S7U t Ίj5a֕0x9C,X봡>+|rur28q;Ufn[􅮛B,~f+oӿov,@Q/SA3FIH2\$.Iyuny zx6iqSvX\$'vm-)2 ʕ, mUZY\8aW(Xlpr W T*B;\!4 [ @.% Rv\J֭g%^-~mťz׻n Wb~lu{6&KWcu1H> Oz~4ׄ~,(-8=dy_YQfoÍT9a+4}QNFRӠRLji+c +L3١\.)Thq*5pnYOBWNfr<,5Zbhrɼ= pe;\ݷQmJ K PV TZ;\!$p%W\ Pfm4qqYNT3D+TTq*puR23De+x.BWRWG+)$6#\G/\\EsղB]#dj++X.BKgw\WG+-acYAIfXN8cvVL6p`ϊ(ʕ٬/1n81ӫS WyeR{oFя?-HƧ:>y>T}s x6%C~4Άa尫6/x- ce =Lfq8}^?HSF T{ϣg٢}'7Zwm翼 JT>8ΪºXXԻ&kVW(٭|-s}FSJ)| )PB ܩRMjTlk4x_,Bp Qc2NǪFKvw>A$wXZ 3MQpFQfjJ>F[g>)!UVbP 2͑.TZoDvVpVEOI(qP::sC==׸f* iKJUl=LpW(W\pji;PW m:#\` s][}o)PPW\+s]Cm*7 זOQ%FWLj+ajtM>B+Pkj;P%W0kXFlW(iմBk.);\F4lprWV~2*U7?n:^Lo_\e9=% UP Z~g_[ՑeEOUŪY+oyJPXW @ɭw5Oo'z$S!5I Ӭ-\fAa2*5CNhB4#Yye_uk'|ŭs)F*o?~\=l;;DZ`?GIyث>~}Ffw>-+-z U*3So`ɼbT/JM`-5a2,ƃO OuҼ4B-47+wC­!*B30PR8|g^2i*bt.O hxῧgI5 J\3uQ9maHIS0,4EFY`r5Fnn)t|>GM]cԍi ;dQSt|W TV|vGu#^.k7;[fψJrtkeq~( 4ox=ZwuzDLIk"!:K)pZXk[!. ZSe@a,olg<2Aej6--%Q칭Mʆ!HRR ybKQ!\PS@'-/)Ì.+-Kѕ2R_m@RyvUU-WjcX5spFR[&j-#d0rKJJYyI/"X$4ink lĥ xup T Ҙ8_ i8qDau60(Mk8P.L`I J|,Pmphoje޴(3_t"[!" <!R<7sC!$G¿F߭M E8Ɠ}Z8 M9N8sˋ|(`]q֓?~-}#{ΐr3a@9AkOot۸כ !kv_mH²ԴgKczCnk2T>8x_:EzPm8p\0}_ͽk=gXё_ź󴡮 #ՠ>!&jRÿygLP^xd!4BՂ:8S*V6|-a6}6K7`Nй}+4p~ZW?}7\xeu}l ge̸µ:K~5˺.n8NnyjȨ ,o(YGwY~|+[si ʦ/DF+$VBA*)r!K5b¾]RBD Ǚ-EЈFJؿv;}NC;wܳ灏v]$Mx%g2Mи zbLb@@YPK42Q3"qC:&hȡ0CkZHΔ&Q)b%+a3Qcq-7gdx3@bq[_W۵$"B@[DC$}9B%!hI4')hbdpMnRqC2KVywQz{ ;Yoi\sv1h=sVP Yo[~s ~0( '}DXҊa]8LGn,=y#L3{jNނQx+Lo~xsQӤ&il0"2K8IR:4rLAU xsϐ1bc<q.VVcBh15ԸTM契|ߪf U"N)| >Zy 0)G PW'GMWYZŏ=YR67 PM1YEyy! a ,0 ;" K.'Fxilpˆ[6|lqjadd$yˍp E78E%!EτgLV;X6c1*p@M ʑ[ "RHF'S69k5)XXgk,ӵxsoD&'ȠWVtFwpf<,oc.w$^Y'Tuf>fz o_p4Qٳ(Z9Slmε^Bybo2%XUsbZzs8Je6W=y4 d%lݾMOW>#+[ԼTr|?|oyD]GGӯ>Q,3Нl]zԜw^t۫x5_yIs5'9q:d2skM2%NkoKڼa.F};&h]E]z:Rnb!` -{{ٻޟ5b{bϒyΥ ^ *FU9h)sD3Bްy6Aw&9P`Z$b\6 L%ʌbVhG (ODL1qdcg}1?ewwhַܸf;/F9-7[b*&xpTNVTm.')dLF}Tˆ#|N 5{|!ן˾Sz#|ϡ_ƑHUf3WLW94$V宇ᮆdu;kφb:0 y0"2KEpxlʬoNCyĩ`vtneʫn2  IV8iq<hdsP9ahA"S{A"B7K6$#mrrag"N&΄ zj ,hSj*RXJZ%DV`#ShRH8q#K,TptLhXLʭn @M3\-G<9N:ZYr ~*!x{CoZSEƞ0o4{ws3_sae8O:aD2/q׽~Z?8:eO/ HA0J('dDNεrb2^I9$M1{6VJ?Dg#1I퉆^\%ǭ*,Ϡ8d)Ucס 1t,me^ƻVHUzmWBըQ *U>:[&?'I6_>qGè7&Jf;ٴLTLX=zr120[#J-{_Lǖ]2tOד*~{AHMK%!׷tU oAm3˓#*ma>Q̻pпtyf79^?lnuɺVJljq#aOñU_3Llt RERsz.p:>|_<gهpp$ȳN;N]QMSFjۛ7AlѴMz6[I]i(>Ŗ[):?]n ՉG>h52 t+(6]YTz(aF,WZB}bKBO* ȏ [ϑцziEyRH"BwT'F X`*!H/nl/yE=_p psV+a,*]n+gTȜc#ɫО4rSXr46'-3b?dmd׹"*jvCqG_}A!U&pme"BYpJHJs&mFgDF e!TG"S+(<>&9$ "6'zo9n,8FI?u0h^CB.>+R@!;D)Q) UЦ!1 a{ Xn3`[?O1b$B 2boށsfZ:zo{,g%oox๦`E` ")RHNH sm.6nj7X:PI*[#*WkŒMC,rG]Hf6W/ \2nfBO ['"h[ s +5{ ={>!׽3ۍ8KUSb3sWkTalHEkSLݲ7:zP 59^P_U3ހ+T%ii]Z7iiF>^o43,l% &>-GE\cI)O0-b'^xKx: iRO9Q Vi%u:C!s@]1IQ==RRƹIR|$AUI尽xs34(=*kE3sEB JWg-/M>> $]@Y\au境:Z/f5NW7Z4SᏪH^vYHqn~xt1sRn ?kʆs.=_-%)u,pe*yYKILa[=3 y{! NdE+T+⮶1-}MkWlM٠(.]A :\1I>++DYr(7ZX{\{sH5/úgaSzk9.~$/e}x,w{f&m~!:ӯ0 F^^UaKh1ym=X 8=}D8qtŌ8?q\+.ᧀD5 FWTT J 2 bLP3L)d#jenq:&&_n"jRB)lvE,0b@T@Wf56jz oCK@:޾r?&{TT-ʩЪGOWJhn|9Pĵ*ձmmEOT@K RiɊΑ@ZRYJ[Q%e1'hO?6Ao{=3cΫ0ggUPEj3;n%~͵wj~|y!7WVޝءe#.rpA$rm(fENE,Q9,^Qs҃ \POImDQ>D  Pffl޸7ؗ M\h.|P.|=y i}qawe$y a:eD3#ځFMMH#fe ɈSu.H\2> ;068GT@)$q1+[lI31$}raXOZ`RwPM_pv3WE: /.:ޤd_^t=xq$1(%OYUb:0:@6c1#Vt[G)1^|^<>M:C3()w˳zF <HwN]u !GzG4uя7#JYVseGvްO4Xʘ򩦱-iz|At\BZ|:}|VPDǠ O%8Š2_+gSM{l*袒a1PJ9RȦmCMբWЏIp7 )5.~lUYÝw_/76=}t۳GŊ8.'g}l!PIdu)2@, @dz"bxٛXnU_uT2n!-nMnDruxٛE<3'b JDنrb+TTO?3_tR< 3_9 `nI\Tc]&b.<\i}@?1sB٣7D6$cUI],ko2i@"lw(UӢCgja84ch{E}{v'q7@i { FB.-gNܲ5f3BEɨkY!E*rΌ]?<]-s4sK -gwy{Ig-odoF,Jk= p'\9v~rKY5eծ/R1_ZGZ-Gڠ^^_~z#a$]Q|錁Z?N'ؔ-j󢸹;y[`yNٸ! ڸ[w]͖;y=.Nlf|!m]`MI|7ܱE/b= ntŽemaggߝ]b?}Y7^mõZǪV65=[)B*elmfbPaE/j=<9l.`c!fs+ ŏr1&2ٯe%4^/$0՗.⡜M6=Covs(' '?Ij/'/z?F2ӖiZ˟LP:sXeZeZi(Ӻԃ'U6he>jhӕdztEGDWXW .cUCe`tt芔2F] `V|4tqZ2NW q]=A2 GDW?]5ZԇNW `]=IKp4t" ]5JzteA飲a<S骡vHWF7W[y(W>W[,7ziDߔW_Qќ\~ƊG&5#owD4-:npgϮ%8tn( 4iڃqj|,]5 #]5 ]5|oDJ;b|q3]OLWc~=t{Еjߩ Z#ؚЕEeJ ]iѕvGCW BW -UCyh]=]y6<3Е%: OW@WO:a\J|Jڤbۓtz=^t&"}:aڲ [)9N/s%S2b9}%lNb>=&IaǓ˚)trw}jιmoرkw޾] Vz% ,_̥]ePX'iFፈLݔdC7{V#,z!Drz1,zx#Nsϱ>|+쥮LVْmfKKQW  d'n9!"CbSI.9g[ϛ='?$%hF(a6w RB)")RV)/z *1f DQs'ZZ#W],jLQ#ju5E (iL E褈\ K d62Ukf1Iel0&]kbP5l# ha&O,![hQ+1gg S-B(ҤHACUJFw@\LE&Z`ȵ:[6eyf6IkJmo= ۏcԕЈӞ:,~)gӐGVκ@'(nbB>GaIa#uZD6fwMXUb%9P5T$2۪w2 U1hYv]|<*<  7) :.˺,Z ch>TBŵ4z'X1 ɒbVɊ}I"!>Q."¹h-hي V u1aJHe$CҪ*.FPCE}% cF-)EבUeuT""KPe0@`erU,ĴbO”}%emQb)25k|ֲ[Kԉ́\WUl;mq8+KŒXy"#6ҙgA*ᘃVU9.+W6hW;'P !2ْY3.X\XcT0E_eY s1*gp.H@r9,[<ARXwOUIxK:ɨjT aP $H{'&I܄" \df'- skH_W;P]nЋr>؎M?iCui p wNA" NWe}YU+R,`ŨQAQS}SzУ4 dd^j+R1r^lW:fLiWNc YWӀ uWl%We ºwuDa)"0@%D&h@F~_%Θc,ɨ[s6C W,,8:L;*%@TBBB)'8TR dg"UT H~Av134jp\ T4gvJ!\`cRW,B,x'D)lHGL_(4_m2 `VW>c*AŽr"ͺ@[981cyŜ`!DR157EtPR쀴5>!2H\i$[}AՊX1A3OdPgFv& _nݩoŌTU%^1)arbBU8p0@DŽ9;Ovߘº\|iW돯g! FkΎe[wcGx@ppi?T v:@CV^MR2:hrUac*J1y aXP|AŢΈ Ay@* `E&Jy C0<ӥb)^i(^04_6R֨k#[[1#fchcb-XJX4~F5%wN52@6TB] !Hb"5 K wO]毁r4fvt#Ha*2dEx:K䶤0 h vePsNJ<ٰ ICWBjFO,q;hL=ǖ:j|!Ŧe/ #۰_,f9($ԁ*J/i@ʀD;35\F?"MbzTBLjTBPmˉ␭<"nռBzЖEs6iP55h@eVNPT+[U UD^H-a;hېV~E4YT_NJf__oXoi7ힹ+Y jA0$.Gnnt^f9li) l_iǖ'Jgdi/knh֡kbf,Ac ]fHƴ;-7\Aӌ4&Uy % x[#Rrlz /Pn$VcqPD{AY at0Xg$SR,=hzG:T}vLc_P|_w?B+IkM)Qf5#0t7(Vdp!1\ "-RI10Y/xϯ%9hU?=VNmT61mF rc#7+ڟ$࠳ԴcF&_6LZ$4T VP-G\<e՘: /Csƪ;[*X;T@ ZxX"Z70Lp6gkBz``x_HmO3rJaU2'֞f+JO(TH,J M\ѓd[d-Pibt1;! V呶˥1D 0r,`v*٤3|* a?PbCO hZH, քr5']wT BN>z# zA?]yٹJ+5׃u #LJ#gۛ߼bON \ϝ@PhN ^@KtyB-+N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'r@Wzry^ tڗpN t$N%:P= N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'r@~`:ysGMVoݯ3L ؕq ǸQwc\4RԈ1.-'>7:#mW^Fg?w\JW ĕ|g9#\਺}|pZq*IpD\鴧 mGٲ\[,XmslT Owz8 ^=P^;_-rgΞ)?7VWoh tϸkV:9wF4>_@7Up_]5o~ݎ^?\S:W:VK P|w'Krs}s^s+oшdʗay TĿn k;HzY1?9['_w+w+2+7w'|"QyMC4qagu/hE_Z:L4ri>'d$!{~_76bѺiTTZgz4I&h֎7W/g=W慟݇ $N\u@y/ȔSƾ~KoͬɪF:F\KV'祓OYm2ƨTCX1嚭s >6d\J,-Z澴*w2YZX{^u`t7bZ+V/K~>X;WUJ(X0nprW6WRW]OUG$Iӑq5MjZ{}i*if~K'\` \+Vjb4pwQMk}/bNWkqEE;):% XUF#Z 1p/r@\\վ>q*w6Wʾ1 rS/bi TR*hWl^bCW~Uzɮ+uxSwԧ?w],8n0 I^0jJVieaN:)dw+uQVkff*m\}5 Op =:֯8uGb64$ZY Ү׆w X-U Z \c: \ 5w\JW CG㿉b XڧRUpE: \MTZ%q,_oXpnpQq*w^F"Z Uvł}?;,o6U&)@|#{GɮX/60w\J#"qt";fJ34F>Xgڃܼ;+\=9^xa\MiR/G-=(jiaSJ6)=ҳsO騉dѣ--@pԶk\/K f*l,qi!!#\AӪ\\czoܰJWĮbg$8>ir@6)2 WqK^;mu+&C Xs%8mO >irc7ڤf+Vj; N\A=0w\JW m >&WP;X ѡ#\`O1r}7b/}q*\-W{=K4AdWPij $Rt?c5)$TVutMxluJQ S\{4nf!Hkp|V?; T7I%5Y\-W){ZhY.^pjODz`W_ >0M;,#A|x%i*w0f4WIp]Ct3nݷW,T/bnbQ + = uGb{ Qp@\hSWd&+VKvbs{UpE(RO` dWTWքUZɮ+E;MWƹUj; v\mv3jW2q/j5+eδ9?aQ{c{owoICv%`;}8e }0)6^A ˻ǁ:hR''/e#tf{NǓǶZO_z^s+2~^m'~#V_ܬ>yoOߟ8 oϯVH6l  pr_?~iE'{Gh9Ppn?)?v >~q}uY =bpfZ{}O1-dT2;csue7nѱ3Nݼy'gw`X;Pi'igO;57#ҫ7 hKFZc=WcVr%jr=TW,^pjIWq5`:•N X&bn =lu?5`4S\I1D\Ed+L/baXeCKU2A9Vu+Vq*M\}5Oz}G=: >#?Mn<5զ#jJ7{Ս+-Ү&:m=4vbW ĕW,8npN^pj WrfՂp63~ѭ+cΌ;LI3;$[7CϭW<ێ:Wn~ۿ 6r#" !mE@7  HkGjO~nIm<6E_Sbqz1ԧ={ߵZ`W^ZV OdbE0Fh\,l Y3>8CA6W?|_եE9l}Dq]@qT*Y$1Iq+BQ,!^ Li"}^f\zו^!3|5 ?ǓU'xIu2b}X \멯^nǾ`o7s{tͺt>+KR#oak>:ZF7'{ dpj3vmAU%lYǛyӜrXM|?}||?ےK\s?NqQ8]-m_\yM^Uj zs^@V#[#, #tei zx]l(0.y۴xc2U1sϜ~GIV?+7+ͨ{Yo>j(b:{nH_]I $eݽivfN禰> ]eVs S\4\+0. pwrC%6"}No-(bC!Vz!2_oVkY#L٤P.Ϛ|88 0^Æ}D.7|%QKj>3ֻk?G#dO;_FGgj&-$>֫nEa󋴺UkG|[CpPڦ ]m{9GښSo<'Bz*ah,NէϹ!хlix *̅^7{Ѱt0G{Z?~w~ \$k* -cPK}?BakVŝF/ 7f~xs,3NhͮNV`沯W\ʲ~JfzJ>![-xM|5%u.h6K@ʝ֍~$;Cj 'lW*zb "QmMVpK?D \G}6LiedgX]!KVb8g(hD@bL#"γ$*/b=bϢExeю5]MqEG!m(T@TKe)!Ah@}._Qf!&x RR4i ܺ$Iu&AUI !#zj1!15ϴ҄'mX[jBhjyE;,b^6&6?Jt6ikëgdn- 'U}Vr|4|LQɀ `C6/GDԇOUazuڌMpÏNpJ ebIPPJUV1"DFS< |tLO" rr_TB>|){l{*YSAp%P" $ a |0$L)H9rT&,PdUF:i~&ZGAY(QN8e%#O`1qvS.1dɄs7w}^rܤ&?\jxB9ApB骪6׍ɖ,7ϐ!ڭ 뭵-mݿfk褮7 LoNb|*9yˢg~ŵ򹸟RCw?C)G?œ/e^tqN7a;f_N@ߐ4HMV.*^%kqOGcmqƶ=4c[1M c \ͽB{aFC9s6I_E(Y1L 1=罥BD:QJg]LldqynoIGTy+"RǬ8ʼ RvY#MyUTQ MkB H6D o 4DQI4#Pd ^kbBaE-&nEMdhEi#7xr@Y :ɫn'+v>i\:<݊EN9]Pmыsr"v-qڛ A._0:;;q\98ꀲTQk La& #Q0mmQdDE@!ĐGCj4R(T J#c1qv#c9R ͌}PByG•DmyfƧEjfT`am ;̈̈́HLA)D2_(nbR&KIdP2l9`{8˰ɕJx%@3 :mGML `**yD4;ME`>>ٍg#.Bi,Ma\4G\<{-j4 =%uySN"i8caٱ/xhGWXU=s7Y?dl=uhܲwt~"gmŽj͙(K3FYGeT6+ o)TeQ CVKY=< =4mJ d2:8kR2QIJY9p4Ip%˾oca=J n`Z`EP2Um` 93}0p9|ZY8)&h0 .4ifAPFAxQO \^L#0 ShփThR84rIA[0,c@IИ Ej BDB , CN!9S1;ORD4XrQ:W1q|?"ft8OjtQ#gU56`^Lni`KF]TK } q`t;E퀇E*-H4FpD''#tvG鄞#|ii>ɒu |Z}Jr)3)@z~y O. V#'F) J! RDcN+f%ehF 4()GP֛ԢCg?JLv Ie7Č?AB|g^d[>p8rsdcD( D&oW,d@$ŕ53~O}\ӊ~&M>ϡ~Uo,aO'[s.r XĮ︕.G%/3b]{dYNk¾).r惡O|ij:jSmEtZ[=:zk!.F͝oG|{x5l9\")~(J'U A8Ǿ Úg]k=\I:/Ww|i|?w_ouO({zqY\fZʘ *ABI9I1~uHfNǮ.?'qơ1z{&$iEJXqB "(U2 7ikz9>>gO@iox5]*ԽRG].8fOq27m[ -ת\v{.){r.]Tkwn|$zB}5uYbQS: a`r;(N4ˡ3Dxl ZݥcQ̛y 0.h_d|dĖ'::hkqdbU)u_Ĵ=tW[*Z ">y݆YB8컳ʉtf:^mk5iHJJ&|6 'qsQ_3vC9.w$*qF{}`E(*Zkx;Zwy"oSI࣐Ңe> qrrZ)84zW 鷪E* ,gKkB0/-aiL*HOe1 \(ap#@"YEH2Kq: GVY,mY6ٰeV2!R<9LF@ɌTh_#/,/C+:㖂ǦES4lz79PtWgi@>n1 8*oc#Wv߽;/o~^t{KE}Q;q##U-ޘ]6|b ӦjD-F+$TB@*A9(5HuS^-?R1If]RDKǙ-%L"BMmu hs5$fI,x%g2R4DOc6m4%@-dѤ\k#Z&r$wu˛;>985γ$zCCvE>]&HJ %ќ`#Dy Rq VA 6ctwǵBVhcA?,#)%oBwx\uH^F*< jlxŰ v鬃R[GִJP9x @uր<ēg^Aȃ x梦IS-!:7Cc=K8֣ሁ$)N}LAU =NQ0Ĝi OMcy#.oY}B|}5ܓ}uX&{y/%yG_'lxl&f[*O8H׃=aI zNV7R4qzQI2HUJG_Y5gg_^Ը7cU}gwmxo ?[y,Ll6ݬ^'ŏ2qQd|秊>aR&w]]`/u;Q9/;}|Th,nz׃9VvyP__{O7.`MivB`9ް#2x#47zqNmԵlwCՖF3N@_*Y$1I BQ$,(/fmJ&fqB ]&hv@%Dy &g[Niǚ>Ә8+ԩGmqcCc->TI\%&䄐w"͇u'3Vz88dfp/?3~A3Mz?XRg$󹶮Ver$eKLD*u}%FR,oк0~(RS!6ѡEUjݢ/T]^!,AgunG5 1Yk -;d YDDJ7G!a Rr'QƿG//;WW1t2#ѝ*vT9MUƨδ n]KNՔ>:ؿy%=X%kj{K{xg`O&}$鎐Vfʾ͑(. n`s^MetDZq;-mͯ(l׼_}ypdr/=yxz|ora_ạ}[J֯>1]cӛ=rCsRɣ]V*X j6,`C^øŰ.^fBj*e; wUR ZMj#%p>- aCƥ ^ * Cu8"7 X0ýwz2 S$}/<;Ԛٹ&wp%r;e5N:0KB=.@ *@tIi#z gCKJvczٴ} i֒'/aZxŠd3Q1gˬZ½X'ij؍~6be]) b\&$c[VNZ4'^>`N*vd.$"t4T4d:sBmBR&zj 4) 5O*R^XåJZ%DV`!ShRH88ݑ%IL)\4S4BZIrRЌ"{y:)Gſ_8KΚyc'J~O/ %^ &Mǹ/m7`O^bǩB΂ #F$#U]w.CD|(1a>$AG drŽL+)Gv8Wh8Bxu/ߛkVJ2!H@9=&.WI>7`ui힝,>ôѳ!b }SʈTOTz{:99]/\ \PБ_),r$BYk'qucB(nitQ?Ckczt>v-RSaN7]v0k۹HuU9N6e,) ,-I%]-+-/!v{};[{˛ICgev:*Ǘ7##:RՕJlZ0DGrz?}=E/l#ZQEG_uw8߾.?ޟRfNlQ9u2X76iV.2TXޤhn E d B4Y%~>J̩Ovtʭןt÷a3!V:1_rz|'٧=嫟TK'=tİw'ҍRY'1,thP uĩܼ /--#E9J'I`NGӢArt`.iU:S0s郕ZǦaF QSR$BFZ Og@-mF],swl~yPm0÷Po7PE5[8ve9j,v8g%?}l55VKQ fD ,2䄴 kWL-oѾ9@w-j5a`$ G8g9)KQ[٤ m˼78!c!vUهpxף?l]I)yʥΣ~ ȅ@ w-]9TcusVG:]v}qNs U{ĕtknlFD { G?[ë3S fJE9/FRbQ9w_UFG)NscC9X*Rsfۈ|S /:>4^cܞSgM?OW:F**! FTv61No+;2RpJ(1LXQ'.NޤW2UlԝU۝[8`ڳ)Q,ܕ‚*TfF>gB3> #,bgqW{bgiR*ٺ/adp8AWy(*K~#Kɡ5W\I)g;7W]=N\1zlY㤜3{a#kպ]O%Bw=J\)ȡ+V -5W/\1I+mg@pdఓ`cȒcɞ8_5vۖ==A,"+-zEUW3]Taa*>H Ŧoe#a3Y(}T]Y zA|Xۈh"V\VL*qD~U/d64Ro,Q5;b-iJVfo!IU7 ̝O:C}U"O^'hR;SO?`#ZzjǼc`VDi iY;kJ1VlqʀVߑy54_>k.jyN!=}97T5~.j4iFٹ:fЏ|2Fnj#m4ZKM.O¦ZjhD@,@{"M߾8inOҽ/N(VŨ( =2 p g,c.zD6.c}7aE̳&{&Af}9Xm}pjlTZlimPЃ3>gBbJ%0!i6 1J2KHDnЇ\" #%S g'ޱw&JB])!zހ2dZx{ei:$@KؔV/2]VX-BkzAz~ﱖkmϼv`k-R $+v^4djr$#2JPXR7, \I1$EV)\bHxc,Hf^vdhcJHJJxDRuĹ.d4 WTxD 7hdH&$Gs [m_i{w[wY%+ۘddA Q M@(9F!Yɳ uF:j:FVPHpΨZC_M 6HСB%R[!ࣱeQv+[65e4EJ'L J$Pd+/{gF/Zŕdsؤ7F(]Xmun9bQ1sʠs_2Z)2hsrQE0[u헔֫f`ԅjtL}:1 lNM6KЍ'Ӛ72sy Źm{Ի_JӸ FGn@$C Zzt52kQ;23-^ VICOՃd"܀&:)82UouB<ĘUvpxL&1hcӻ6smfޣ?֖M*%1 , "`(lL{ zh0SQ5aE SRѠuV #cB9#bPD\],6IgUV%eЪ颡U$;ѱ7cwGXFi RAdN'da aT=DÿN_ XɍCY?uru:pݮl-7[ ͉;3csʼ&(!- Zj c=( G#<"ۣwvIևn7-:MNOIu1nTt1_c)cCW3=j˂ZC%xX$k6_@1ZfYd$om A:} C[mϥ^Ί&6 esk<'Z$ar1S5|O'iܐQf堪MsMsϿ/ɧq5TdTucm;׶-Snx.86|E-yG7O~S)FFՀڰ!۲6NKioJdGlZ_J9Fɉ j _*P1PjAgIRѣg})z> L)֓`3M3&.+N(-,sn]0Xٷm!|2^;fl}BUaMO)6FI5ч:b"Zͩ JڵwdEas|Oǧ8*AĔE HiXϡmC̗iINJ>.9V?v:jxI|S[N9lb͡#(&{jtmPY2@с J#CuPsXʡ?]Lj6c'Cr}#~Rg,(c 1Y̌X L! e2[hR}fc &qcى 72Q hIp$H%!i2Hg{!<<}H4?{ڇmW֝8߱IdKs]Ps24*D&02EO+TG]q,ts R1@{hp416f)\hhMѤyYH*CbQ(΄4Jt!:%6z݉g^PD5$ rvuqx>k χiVskg)wGu% VoMҋ²1(V?:qH_c~&S b>L>_ GqLU7Nϴ@2Cq` ?8cD8pi:&Ӄ4_,9I*[(K%fW&0iZmqAi<+U8q1Q}jS:s$<|?-kflyZ*+$(]*jŤuoNΏ0KsT]kjxqя;?47,.f[B08N}+o7#x2b~Goչ4Z2RiI5U۫Z`yOJxO%"d:s֑gE1b8d49YdsN'{:48nľq= ;;oi;uXyc;\Z˞,un^f]Vì aJR:ZJj^K% V?4އl+g*Ijx2.b9RfF`[d2]|OF܏nNHo$n#AplVg.2=-}x<~:6vsc^  2151T jTMG.rP' Tj3WɽIأ:^/ZMz k.i|, JZeF H,$m1N JfQHh%E#rɿ:[;k_4ȱ 1B"f__/ pnDx;={!9;u4uAݸ]:ߵt uJ֒-O1{$p=}BO uۂ>89j͸uY}hgBpNx13Me!iZ $HeDDc:;ɓC#Ӂsَ!gZ۸_anЍT[ɭfs]y~ri+E.IV\ocHJ%RJ=Jp3=iqڻЌÏ4Gr4r?ύ= oVBl%GGDX90)A>PxF2 ЃRtB9ƢLw[Xe;\Y# IVg>qM&'ǵB:=܃swg}x<NlcGۿ: kb6WtE+#ˋJX,GU!(56`Ժ1Fy&} L$GhJ<|iᳮwa_R4.OjaC羴9 duK$<k58Z=pbF+^q,y`Q",ۑ=h!|YҴdv%6Bۅ0O*MbV ~PbCKV & _a\4 UM-,j}CqZRـFЧKD:oJWW=?jnG͚ety'Ûrk36_f^~tEtq*-8`wcEYZJ%c$F5N뀧 7!ϗ'D.HRp$MߧVʈB;.hqNGDz +&Lo;*Xg5C!Cwu/RT;;wH'R /x~KЁf2 o ql,h$AQ,EmtudRHޚCv1FOt0~Ee{ZUZeXnHCI6%MǗ#iUh=b{ҊQqAgޏ%EQtQj\?+C5E?ynz&vݔ|ئ:[yG'wiܠ gsݖI?Z"B]|V_z>sx>>FnnU7PqMoi{QfvQ ^ټXSJڷ6KeIb8a<~~s,&}]qY٦aO#dv`8riKe),r`4q-BZƘ :)o4{P SޤqC WgwzI)KpgbmH xc1EOnww%lџ,9>oߟiw!jkWj\^[fAl|*a=JyQmLW&M7ksnpds> w??ci4󿷛4?8?6fB4͞M),ͪi;zt^AlRAZvkgWw\ΉuK.d&f^wFP~0Kň3ή3;[ݟoxHJK۬b9+ٌN&2S, vCV8.8 GMיhXF2*-UH h.;8>(#h$h$'EVe|h՗uBis43 Qe)͵^Lu!EwBHf!t\HYҗ6#W’cBkWhړg̹{s*2}>_K8=\fDo;Dԇ.g:iAd:|޾~wSI TJ5V E`E*)Sw'{bR̓Qc/b @q[ɜx`94e_60I%aRA2`7ɪ=fC,ag re*X͜mT!Z\gNߍLhyV3eoNeݬBdS㣯^ 'vu*->_-kl_إRcATT`7I6(M$tVi{peםl^9"s#(#> 0ATP{YӘ]*x>,ES*{)0 )E2 |Q?#&.2!GruVzPk5sTYDl{H׍ˉE[IJɜD* REV)񬅯[=!V0(xޏvb3N.'a:o  g?Ȉ-Рj"&CP&QBG"0 Lʊl9˘Fy2D[ d!`EhytP3v,RfA3̹<.ռcWVQ[u1MtR`@ )* ׌G N`63*f H!3HQt kbbA$ (/ c v2V3n`Z܍󄋓!uV]qTEb<^{#KXRHp@a, ] o '4t\<<⡭v7<ݡ9Sl;gb`fim6]d?*ɴ~K*jUR _R[;Eh~IS"RN9=z)^w B߾(Y0s?bY{nìVNҔEk*U!ڻ| m?k9W̾,y }-ZHcϝ t?g|,.FصLT-6g_>mILovȋh7X 7Ӌd}Aﻟk/f%z`cհYw9_:.{so'+eW\񣁫b1U; J+$-mpݺtz[ *OOZkw3~D [ \ Zd˪~5 KByDpU {aXXURB=\r="\shXWJi;b |6/pu?|bY՗upu?+ p\Ad\8*暣+>t*V.jW_\ F^e*eo5?iMtn)ݤM5=dN"=& z}mo>gD_?UybJ20},0]=|VY`kikR6.]ȍ'-`kI{sz]u4*彺Y%2V*RCc]I˥`3~ 8pˋ7yO\o>袟~:!l䮿:вSyt{7u&QBL}H<$ 6[>a5dQ\UzPJs{J+M8*vߜڿm0=dr2<=١Aw7\0c`_bTwi:> &:ϛ)&5T0#Jޖ֍1L63MT0f_oS` WRǛw!GC1|& .&E)a{QFPVI,P pt%íќkQGH d%|¸֡Jtf+2K YSwmܟkmȲ v ^Ǔ 0؉ Ĉ"5$eY CĦ(hNM]{[77d<ӌ%L@uҩ&) ʀ9 wd?niA2A@61m-bh6&e Pg`Q㞢P! ]Re.U*!tQHBȑ (G ,1ˤTpuLQ-)U'ˢ 4]|^+?z0~ F M8ǟW'L _c{¢^h8_a7y܎At#F$#M|G+Ĺ/'jPᢆalNg-5d aPN9:ӊsÑLn&S#z~mQUU%A. $u *oqtvW yu\X9:u|Օk@ym-5^B]W5WBZ.T(lu6xqz$nhc59GF+IuoLٴlLT?Ϯ\xs==vJ-;s>i?;-wsbOrr0N^imI-]5ڛa6p`żDO_ۧͭ.jX-[d:,d$7,qu4FʰqLh5Y𡊬 CU_r 黷./wO)3o)ں̋:`_2O[7M-4hEӳROjRr->%VħQfa6I7}rE#횳Zΐ@pO'm˥v?_2+ Nd#9xIHc6φw9Y?lY:ow'בZ%Rk L1De#@qZչێBn;~QK}݆Zʳi*õ D3ڃqWB US"Tudww9ڥyxd9or'"W+*&&^ZDGuR;qk%u4q1Qy "$ KqҥKug{? />gY/fW~^r_%o.S1V2h٭3n;z%vroh r4/WiHh>K^:ZE$։\y cI/aIt_1=^XD#[呆e5:P>@2D'R(?B8K{E锨G#,JrkAYm{6:+#{pwy 6(.Pusj=横'>CCH۫Ëfb}w;S|1(I~昙<(%if^߻^Ң\+RyͧmmwTfE.ֳ:w#?+y3kj̈2 u e>vo?1{c8?=~@1{T JTqh\ .qNX LQqO}L/*j$[zWNf#|AWy'Fq 7X |Ͽ߿\rR {ɕOcRcQӎJ1n˵(*R**kugf85c2(Ppg}r<ny:r'*ZtYv,_I4KI,t9w&2vWR$]νt9RtII9 J¹hg.8-tRX"%-4Tjo@֨8ځu._R=UrVQXFk{A=j@Z…&GNVɜQ`dATb byo@Y"1Z(O-wVz=SbH/HysiW%Q? !.GgS子H YF-q2%YD e9뛊ځ&|YC5-IbjA ϒO 4@N8'4#2z YKuiAqĈZ4kq^J$Puc$/ }86-^_/;i_{IYMK{\ZnE 8Qe@Y(׵(La`ÁXL[[EM)C`:&Qa:*(%X;O#c9R ]PBa^p٣e#ع Ys=,&W:(ᕈ٨vƒU)m #v<~4+*IJZ*ڪCnZ)4R!kBThm5׊ AhFis SQ&I@+R;Oa W~ Ǯ #CM+^i \ĝ8LH;)ic<#^#:(`(PBڤ4Q 2@QIjLh4!GzFbyo_G\s:$[gU+.¸h:\pqӊWYVDsn5#uQGYSeֈ$VǮxh aKϚNp[U:cC[ׯ s~gmOtqѧảlyGMi\O jj O)e *(9I`sWc_@Ay'ODa"2F)(RVyy.9L")"\9#ȹٔ)BrG| E0x5]ܷ9߄O/;~k#7/gGJr^CQi4AG0LAhtP0wv'Dky;`!IXeΆGRHIrH'Kn]gdxl TUQhH$ȈEb^Qk$ 'gJ:rT @%'p00I}GMS^N_ 2p]p|\Kƻxdio^ lSx?} q$wԋ<$PF(LjD3kGtr9]gVD[$I1*5z4k4'=S m5AVMq^?FzQGڥ9R~XPȭs9krq.-ijZZSzwC7C2#e'> \TqGY͔Am2Wk__%ps38 5'Owf 3e?w9C:z?GWϣޥI4o36''"}ݻŸNfF^T>V{?>~`>Y֡6&Z9)' ru̍/#ͯ׸ %{8WFuz}ݏwU7IҚ:kSNǠ枅@H6q8zfַvt}ZB˥f͟'mRKi&Gm.8 s`T}JD4cȷ)V7|xMN۱/YbLb so=vCj`}mϼFX/os!X耖 ~"a0'n''flY_VkvCC 5z`g&%;wUj5b+E^|y٦nIwtj&mII+&$UDT)E͋d '正=oBGVsؙpMN{B$q*)<)u\Mh"\o C: ݡD@[_{۲+W~(Q]+E)sn*x)Ά809^V>pQ:i9ZJiJhH9pVT Y^u)B=gǫ]H_d.twLnR ;Ǎ4ᣱB.]-~ ڨ<QJvƷ?xz'Zg?|֢rAl0i9 )ޗ7"4t騪'AAÍpt3l_6/2-^Xy%Y¿qaroBf﭅\hվvL땘$1FjWؤvb0(V҉ĊÑb_{YMPe^(K(֊nSzpy-jYT )@Q%SjI3sM 0C0,?{WFJ#_΢`/n3Lv_md#Y$[֫mɎԖe2wEvTU&GN'mLT%G9=Vu3qn]Q.mt/OtFkr[ y͋`>}4U}I'r6~xPkNR6aQQ ޹*x񃫂'w+YA(`lFB ? 2EIJ^@yOA0r뻅&YfJgH]O!CA5.<28O@YaĠx꺖tWܥB3/?ZIۛ.M{WxNZ(߼3 `ڸ2)xaB9^?̓yڮgj&pdOY@/h͙y'֎ME1G_vX{!Zǟ3AdwIZIYf;w; qXGѰkߺ(2PV \|r8ڊhxƂ,3-2W +#d(CL$Go3{ruµg߾|V7ڃojpVGq-ta UEd+=P#s3 9y]#_|7]BsH /mD=r6ns)#ϑw3]xoq fIS{1}MR/0[9;`* 6P釪5!<ĸ<ϒ=>᩾ ''Q5rCx <:,&[;= Gdmu 4-7ų8b^<-e޳PV/ 6 Z!51&i+eƨdb@\T*D,N@m #HccϢ3q:yfV(cHb 2^pɫYn'xg T^nGbd8]c<\lRg=Gۜ2}uM\hSSghAχVIF6կcpWaikU '_x8>d؋8q|$G)f>q{٢E(p2o7d47⿧Ñ])BY-J$evzL\'+^5lgV϶]weԡ![ #xUwJŔuWwG'rhm,]}aG{{6_ӬO~n*knnko^-Y_ ֙B+%=EҘ@u6ZHkG OK,~bqv,NCji+YqJb!gH ` QoXI!cY弼.jQW2ISTUtg*օ6-w} Up s El<^sn64au$›Z.QGXo\ӯK}~o>8:j@Hܠ1Og)ʸL -.WR^)NL:oOxSI2DsEN`"&)@$E)# dюQjg܎Rgb/I^Jw(͒Хdnm_&6[WţI Q"h9mbK.:~VKcNo^"6Ff5~5/ sl4a  c[L;eCb)c)BF9yAA^H%qʜHebҪRlbK::U4XDHtR %]Sa:޹:jI hmu)BSax _;Hax- 2^ѳBz%c#t]H@HeD؈GCN)#zϱ̽y+kWO<1k j-gV0 cff{\;=YpɠBRtN\%$l %o hT2;9q4<0 mgiJS;>~+I ECJq5hNRDy p˥XevbSg).¾io}52_g=3bW37yp2yDX3h1Y\nK8&K3Ke {5ϗ%͸A8jFʢ n=/80bHq٥U"y!M}>o؛k3|1DͩRN(l9<3U㏽ iZbXR'j;*a/9a\4iIAA=K8IR:4$z3⩣|bi2 O]cĹ# ]Їh1kP{\j*F^Yb{(ҍjB@̇oN ug?bgY\K38bgYJKgX x 2|8QNc-Zb0JȬ򌁋єǐӨOy6[:e`pTPAc@VyЗ;f=JɗC4gӒӸió5euyη~JB8K4P AKJC"`G)jȜoQ]2v4ej i֒'/40-N"A \%Ҙ2[@&V{lS(G,bJzz'hx!"1X|&Zi6E 8sPd( _2. Yb!h@;Q9hR:5LqMQhyH1n iAR%"M+) 4X) 8ݑ%K?R* :hh4A&坳7%ȕc泮/</ak%??D5>^J@O5p$ǹ{W}_caX{'L1"Y?r^?NJAQsFQ@G Ǐ rŽ.љVSx]'7_P$ZZ%Fb fq q?mjgY2Oz? m8B`$<|kfK5WB.T)lrk]H!k׉|F1W+}>] Qyx50z Jͅ9;Ў-7] Teo>8x)YYے7tkYVwh!#G1AlNj6{M뇑ͭ:u}ҴMg6>|~daqLh5l7$5՗*ԕ/Uh_+߻N߼wNߞz28}?O߽}+p. WDP$e 5?޹ihWMc{9;4-dդ.i(d>3%1 W?N~5m%Yf6 /VM{pz0+?}E#՟JeXHC3{7 yd=>hyм]⛼gMVѭi}$zǖĴ"u$aZ{8_ӽ/bry{cuM\[=3)FlJmb}bߴߝtWiUe8QhwG /j]3_ q2/!zĕ j]Qm$@ZY>Sq3#) #h"9*H<>& Q z;"zdsId._Qͬ\&bㄐۭC|YCLdzpYn9wسGbG4۝N0.Q36گ6! L}˓,I]xai)R S;L`s*%= iȕpE">!E(|µBID` ܙ!N H<*7F+B/\#WSsҕ9f&;{}j|sVW`el ozDעBؙpvGa׾6G')>~(j uj$PIBr 4Kj/N;oU@jMs\m#3L;B"x(- kH \@$SBDS2L,h5T"=&b /3g Lǟ>r8sLO?tF|a^[7BW|]{Β߃}p$ wKb^93AH" ;e 0fypEG7T$52J1Zhښ"uI{ST)Ƅ§Gp1s'0<8vkZP§h5~EY>g)%_V,κ7>L$L"I&gՎ=$]m$ǒo-\J5}]Iz<μߢ~蚍oo0)=^O:%q09)"a dC2bFXp.X ΠOJۑݻR4ڥi`m< cz)oQyGOx ~J687' 7yb\[rPmP~pD "i1ZJ~F$^Hi$:%< 1 $Yp)BfRڧ}{ʐ-ԭ/58/[ 8yl*.m+N﷞z(|8DCP,iʭ^m heӤżAj5BEfqB{D"|}M`\9F$Ē+(Gb*X#0(2 ec ݺ#\#fEќ_2yn3il܊8xtGWXGS `rB&6h@:Kl>ICK*ht(f4pD'Xz)<]g%ԐcX|!Ӏ8ĕJ 0iu& #Ota'z/9ԧp#B-\G̈LY@+' Ҳ;:F^ ٛA ֫bŘݔ  eԟ"E=Swxbn?ad!TNy CY-^^ щaCK]n,)Dx$sO?WcU#y:}zoH \i%zom]ͭiۿwbn\qݤxS{hjB-wcwiNaV WF|e3Mhv>ͭerbl{Sچegls]?>*ߘQ=?? q5:cVx2,1 1Hm4?WӦnY9[qG$ +5N?ެz@Pwv 7#)c,| F%}Yݤ|tkfݎMܾ5un`}Usk^=gwOkU A݉;6);-DW4PwO̭'ovy¦![ j415B|2WHW[Y+Ŏ]ЁTK}Y224/qW̹_q4hѳ -/" ukxѵ~.e 빱1$UM*$ 9™xZĜSSRJ3(|m}}/=R1'S[<5 .J}PD;9ͧ_4ۚᗏRԱ oIVRv~خ3^ԯ N5 M`! DH=IɃ d* 8Fn}Y*he&PdUZXh.P፧:CC5.="8O` y*L@LO S{~6Y0ط|x}[ٶmٺ` 5Un3,>_Ƙ aDKOjA(y"{KfJfa<{D \jP/]aCBѬH؄ Us72*ŰXϸ/X(z,ܹʅ<3jF`3?V7~{Bm/>0b3Р49R"$$*)Fr4J$撖5OhCΞ ,2lr'!Kз#$F&J)e݈ƣb jEmYe=]=<QL)!5p JŕdzҜ&Q)@t FW!*}UVr*| QA?2q(pp\+ Z_%J`…|S]u?\7ndP\b lYa¤o|VMm*`k"}oDM%ȁ0ϵ`R9%닧OS&s_E/#l^Qzyw|)Î{b ۛoƙx ~J66(/nlz펶xXʍZ;MΥo`IqnIh'8Y-扸-m@}}i9 |E#ͪmudIS8{YUzVILJRy"e]B K=Eᐤ"3D{0rHD`"R (-a$EV ՙPj@NM zQqnp6N`%nZ-қ}㭟K`'vKd"hO!Jj41p1It^(@F;<=" Xnq =0<0XlL[8f(9EE \AZ<\RNqd4!!*ڧΓutb :S <&Sx%#`b[E`j PVSJbpM5P_8S+T6x;|U{*q%1k!8^ l0T֟,T&:̮_ٽɻ3W&@l<{πOZ>4"LijhҼ Yغj/weF~biaT7؁98hK[6t ; Do E:~QnW杶ndUk})҉S#]q_7عHZ=6,.FB=2 t7@{´ !dqxU&4 tڡ7}U {VN/zC4tTLSe\K}G hto fV*YkJU|4D_וA\Mfc*/քgVK^.N@oI~DFVʷo|4M?>3goח-{kw.nMwyt1&}K6o7٤VHrqA]Qeapo%&6D5ʥi/uhnTػm+xɻȱoU wo%voJh(FGfxP&XJDD§,BF.Pڴ9ߜys:=Q先fiT^iM43Uʈ(qpMT\PZ5RW`A/ eY:lcsᙗ9HysQ 9P;vdž};M&NTp kʒxD!0*fp_Z o`EK/k_}l~oF+پ)`;Z8]m.,|4O<pXe&wK_kdynjB[3ŶgZĴ;hLk8f G*X(P,dKk,)8 ,=z Z;pؓhQWO^94"RXrl&+}6`CG`"M 9&r6ȹQjER:rbrX+/B"soe3g}:n9u6ikw '<(UcوNXky3QO}jPLgYl)~w J(Ԃ@% d =9*?qUzrҫ<{.[SAr 0&25B<P$(A5QAF9HKu4D{ Jq,z̝vX-"`7z{9A sge>2?5.gדZkHG|:3 ,mq%iqoF7w LoNbz<|! ~4;_a;'* d\㜂Q 3} 5q>5׃l^1"$b=JR˰Itf %1{n=W AHRVaMrvVc&ye4zlb6[!-k6s 'Dlۡ{oIo:u=uŀ2H3Ռ eHsT#cdxjDž&kXu&H I)\QQq!,L NUIZX&Xz+laÂih$P(T81g&HFlF|\%fb!ό£ ,wh.GlѻNl8 o~b&L1兖@LsNDE BGQdب/cD`CHΞ$HlR!`VR:Et!"'U\Fl;5TP;8Ef=j vۈwD褣)c$ a FkZR)c; TGhcV<\H☁ 3  D0 0G:Hc2a6sva/'`<D6>EDe="nܸ bxXOZ`ǬK\[@Aa6(B^t0BB2g($^2Xo) !I&9bT˜݈X 8]j^gQr(.̸z\qqۈh dyTK"Z<1x'Sosiqx*xmt:3C)QȝkX IG?>O#G^,[I㷩36P)RvQ)qU 7ح,RE+$ny"ɁL_#tE[WO23|t||c2Wir }ȣa[4+OaL4?u- g5c2udߋ& @O/ k8Ԑ6ֻ6 3\煉Dž\[ɄJUcCvvmFׯjS4/re@`W7y56{}шΛ9EKYRxMy{,,`4R#TUR?=D9Xm!* =\YK,C9p}%~Q'vp%TS,`w1+W)U5 MTGQDepkX sx$ͩWjf-;cp9UH _sU[n2Bo7m^a71 RXzr;Vg-VWaW}|rYm\*ͅG(\0%ӄH&y`1IIVlnGmx.[jOgPNU~PsZIkS .$!̬a?mf% kmNvGڥhʊhatOkej/ـ|} %|vY#D׬uc徻.uQv)SN7^A֞E0\մ&u [?%3ோii{Etwܝ1~p:{AΛ}^}Pٷ%ݹ}M`a^ʾ{ЌB)R\t>)w cQ%.kS*J+pw8O{/ɓSH'Ob/rg#4I$tXX$ B^diD&#XrH# be}0z)#"b1h#2&"y/.9{*$TָrR?<51:< 6rM];[6sdK -8;e +`jS\䌇0J:N'|Ζk8&}+JΚu~՞)43\rO|ͳ&驞kL"RIt'aNH`ALn~>qf|ztY&YA& iT#E5L7, 1ŒK\W PF­3E|6+Ku(j 9$iRrxa`@Q!(8+t xMgfCɦepnx*~BF۵md\걇0+6dqyJ']ZGg+Aiy>=J<ƔkA q獕If$wJ'Dx+p$6!`BT+1jvek zmsfΞ2)H?SNg0kkj$G/{6v tK)ED?lL؇6 *_c l:fU.T/SO2$\8P^rM J*N'=B`P`/Cct&tH́A8f8̈́^x,<\ 1ĢPe+ʦG8o^.)׶V1~.Wnc5Q9YJr#&@δ^EuW"H$׶l[{ A~VehdHtҁF i&8.]( {֡e@pP `z 0#Y.ٔc2(tib5Nw^^qoڅ%糷<_vsP\fC / 5N Fh!(j7yIZ5!\OWς rC)D6$9Q1'ól;ňYr!%@#EɆbLuV"NEĐF&s3qLj䡛9MhQ˶;uMn?N- #y|T!)ۡ#V̑x 4c+GiWT T7$(aq6F9aN75I8@5PhER- f 9:ȑEU`f`9$ 9l($pf=zھb10dʠ0*l+2Zܢ)/RbFI[!CqTByR?P^-[l+qc{&Fd2za.`(s2ȍJJ| U&cU ޺ X"5bl!6r41Yc YD44h<RN JLH)jD."@N"BȹM4BdAr]"%s]\sҟzL?Ӵp58-ܧ/M%] & ,~i/zǮM~Nqڹ=: b k+8~H>n$oG J θdb-%a ̟~^tb}m޹kH{GN~?Bg5R+Rq]&0k+'iH0M(7קݽe%~mST]xY`@+ s澝hlۙH?zuV'iU-i,mIΖ5#7#pi36O\AӃ@<9;sr0 vwN.kuY_9X.zcHiH}1W^q|ɏ|f3]PM:a'^89?&uß_|\?;qD#pB)Z&GG `3_WnNijoҴDB xWhװ.[ՌINIշw eB.ЖuY"^|&3? {aVܔ RغpGv'Y?_[ƸSmEHxRw&HmKsQe9*Sd*TcgO6L\^ZϏmL8l9Q.R4D*3)$]J$TZI(*eDDc:sapJ$2qLef;Nv&Ί. m>Bf/c!nwIIqͧBlg8=Y90AjS@#0P$8B92cQVթ!{d9:HBplA`N#-uls/Y//˼l7W%kcCŒ 4+jd-">:ӍTCA$pc3klFިhLCF{>z);FG j! >~i)n@ܧ]gx**]ͭߢw>fp蝻a1Gؾ7ekČ=;/zvSC֧mP)5ʣXY&{ׇ2=j#:͌yt.iKnzӛzF֛9*%&P܇ģH`tL2hG(VATFO-Q_Nâc%$|K'+"7E7|r~N֫on8]jp_s>I, b{5doo?-YXtVQ?|sb1l{y4S*f%tNiV:JY4wm9H@pE z)!XC9ZUǙY1fz-JysE5~A.²&\uPdP/y-sOtW4Lj 3YQh1! f>!g 7,Sٺ{lbj|F&^ jܘŧ,u $2HatAhIpsCBŽ=["=J4>[E 6֕[3#M20Bd0P˙zpFdFH>Xi6X#_X ?Դ!aP,tsdH˔M#鿢os@| 0f3;,8l`,Ej9_%˶Zږ3 "X/(h21Pg`LQ"a#H};!*iPtQ2ऐ8 J(vGeQ* )#&6'xdLN(̫܅ݻ?rΒ#&&N OqO?6S3%4jVbyw0pS7):;Nn55wL0"8@CGj<.[GAOU|6F6:jP(999׊sÉx'7'$Nl=[4D'RDwXD\%-p6g]Y8]Yq2sapx3Dk~wN;aWBZ.T$ \`jH"?WxN\gbTH菣;sZٴmLTO~np>x)m!&C_s龌i2>hז۵0Wo8^3>I#] awMft`^|ϦB׷c/]QWoT6QT3jqG٢ItU,2l:VOuSE6S<϶}< _~|~_g?2so>8XƐãHPI__2Zjho547b,9nWSQbc48fqq՗qf<57Dx S+H6nuuP},bMUB / 5KB!<\l=ф{;ēt$aiEdQIu$"Q $@jiUހ3,8J(8H/nlXA^?-cycjCxoigQ:%T#)k=#)(tqy5JyOܱ3 {Zt8x/57;G@E7Ev^mQ t").8܆!r+y\V_><+!p}%Uu*`B!JYtzy -8?[:!yѓ7ud8LB采 i%fUg?/3Օຠt-&SyQ"SK' V"wU$^DK/_X LrPqJS-9Iy Ɋh@}1I >À5Cq))\@>IPDR!@=Ƙ246C#q3}iNiī9s}+;jԸo.qHo?DIao(?,۩Uo&cOTJ”Brb˧od^Yܜ6GIp? g))'AUB)UYp@ NqPϞ)=WBWy ]A¸D}%a%ԢEH '8!YfN$GJWTs1 @#4RTV%Pp#bNYHSa͵9{$ iqC.!n瞷^s0 b[>~mbH*I% 05æ3ׯ^k¨~A+=IxWJ$䒆EūdM4?8>Hr̐#2CBzD ^P`P c(rI㙝1(Hl,_E(YTz1L 1=罥BD:QJgZ=wB S o(H q9*oTD%N$8ܡ,iR]D*F4eDAe9L*Ɇ(7bN%4*y$51t-R/p9)*;M]OpB׷yrF}]v)z; Eǣrrq%&`C<6YPik"#JQ yA:< $NBO Xp:"UQ*KR9#c9YCPB9`bᚢl<\ԡ}Lfדc_0€:g !(pPPL((nbR&KIdC9'!fe#{YMtP+Y u;blbb"A~yl *DCQ[Fm5 HA qIT&DP Z1! 1(mx3>$vΨ!dHjсGX"3(Fys>-<$ $R+]Q9a4qǡ #q@]/]Hs' qLHwRx* >#:(cYC QZn&)-X8SEE5(&̈́NS#qD=b]ulŸP\4q 8.WYVsn5`:5yѹON0sa18ma<3@:!^ܫ^_ػNp+y?Z4EgF,;YpzŬ %p_^n{o v`O>vLjg>ƁeCz@;![Kih{H4FpD''xlzzg@o y8,YǀOȕ) qSQiYNB L  y O]d?;}qw},R$G97Q1Eӎ2aρQ9 ʱ8(:v}ج;lmhѳg?JYz' E5@l DJx]]Ι8@RP"BVD"luF/!ܦ̂`Rn=];o3=YÝMfWR7JivB촭Vira9[4yW6(s54;+ҏ|621?6I&~/u|'avy \^giXi"u]~l Aήw'(LC .gUIzrn 1ZS&u;\l]mrMz`nq[fckޏnqoj_7U㾎o[ڥ!ܾ6DE'4p ޻p>7ӀRW˅3JơuW~˽uFWhVl hQ{]M=swdSn_yO۱7^_`V}.~봈g ᫏Un_ * F6_#J#P]verEAy4u#wr@]?$|x}pZl7R:"J]Q5Yxk̴.6t_ +|E"ޱٴ~ϻmo+/Ɵ;4خmw .lг@]ugۼ][61maV'f@+EɼSȲg"J5͡vnV'gh<{ܵX;s"))\4 'qsQSZ;ˤ". 3˘#ߞPK9\%Ad1N%łB $ƅil`m@e 鼪j~,1x-+=V{R~hO_'NZ{qҝ.NTmίv9 PQ}69*r4g1S;"3y<1=xbc)r -?zbQ5^eh^*0.@[)@Beq׆NOZA@5=HDAOS#eNpolg<2*LZМ@DA3oGӯ6qpgW{Ǐ)bؓ'wuWlɧCE@9-`x@1p9xRЉ)c[) vD!HPGr ~m7ĥ31A 9q Rn W͍Qm',&YlN*2J)?D''q9<挵=:85l9No|2zP!8!"bL'OH:E9k x+dS*Ŀ"_P[^W4Sc͔r-9>RИvg+87̭^ xWL=|SNJ%"jOR3-iJ,JpŹUdL89$6b;JOjRB7>$ݣC5&ܹ\9 Dx0HCG-A, 40+D=5nێ":3oWhdFq:Zb|.^#Clq?^}n94]O{ċL6w{|r}yV-sٻ޸n$Wt"Y$@0Y`oAkb[~jYa8rKb˙T/aڥB]e뉚Q; I+ 'T>ƢnLO<j^$PoY]26_Mc->AZ|8e*\bcXHV[-E8R?Rx>2O>(jSEH*2Wli?^}jbp?)0'HUfq&ՉS wݠ?)ǛY-Ln^<tvDG+K~ًrh=HrzjӰnɤ\)x7o& /:b4,f)FoHq#wqFaݦ:'ev,V9r++;[\Ro3Ϟѳ6| R`ۿ4DtvYқju u=ȗ6C2+߰vWh'h߬rxG?vZ.}vY8\;xۣuH8FGn.Wy}e琣A$Ϗ68OSzy`cY$q0W Jbx5"oM/`.e|+${$ںŖ ](M6H=Rd5Y)c&kht{B=xŚB$T9⯍!禴C/nxDnº;r1B&gS縧 jf#D8E7+\{9w ~=́/2x /丨ň/)H/iӴ|,˒*$5Ǧ.h 5u;IQgI.³ؚdRNxZٓ4,}#M;^I8Qsνe6qfh-֑zdfZgW:9;!5b}553>fyth3]31ύ9>'~Bwւ|zv2Z",Wl蜿h/< b߬=Ӌε9 'j^x qgЅcw^Uև{%磤b|0O퇣ͰTmN_]18tjhr]wXM+/mB߬l:%'w2 ݺlo;y#V6*GVG6k WZ=g/8 P͍ǾYGcegj{ ̻|f:ED*=~U.uL];{r'+o|龹(jw3ܤJ˺筯8zK;E[{1[5.K۾ޯQhu^]Ν.3׮[OO]l7zC+=7^ 5cjn_{}7No[m?b:d>BmFziDd%̿{ g/u[#bŋ,~y1C.K"l-<$HR}$:>L.Ԗz4bhA̟sjaOl ƃfB%ٜL̑l6S9SۛQ{6~,(CToQʵiw56tmjB1I978l)s,O{z}zB'LjǛ#.K<tv·S>[~-{l+0Ub%e Ze5ev>Ξ[d,rB}чQˋ%b1~'2EWnxzaq@4vۅQO,k" 4c>~?\/l ]O~t-8+u;,n)Ja[Ek xvVZL g.L0ZJ Rllkbm(u6q-:=[D~m"JTU"lΏ^#x]`ۡq|}߅ Ky:⽱lv8CXTx]#XJ61mVeYC} x5  Qn&fcʅRBdHFLn$bWbG#WӒvu;4mhՌIb(y`MRrt-0'2!U+pƞw#tH~t`QKlڨ@`>6*CZ+OʐɡF%H?'w|eٸ+H“wW2]} nN}v{Xq]#[-lB:ugtݵ:Ng?v$8`a\[="0yu5^\_.׳(Bvyx'W+Eϴ揻n䬸[Zawlջ^yzFkW$fy鸏os@3ka숪83dR-Uɕwaç8pӚ =gCnj{&gݎu9ud֧:g9F!Lf Y|p&7.P CR <ҾeM"o7{[|7wKV_HYq^^JzGEG2vXTY{mTs5{J͙*98g6˖\mΎ6ACsGijK2݃*wJ%1sJ;Gbhq}zuzt͞ХޙKօ 2@RܢnE`P-uKDa̘nm h30f/1^Ǒ|Ŵ$dU8=B=MPf/޾sЭN2v1+E{> q:hcRν*c`y="ĐR.  b{ǀw\ ߚP I-#aϨ#NޜnI1i![M{ %C3drRBs?-ߜUZLs!6I2=8.Վȏaߨْ9aU M;JڜԄ&##[}X[ w 蔴j6 )Q;=Q 7k\,QLT.]RrU69h Z'5.YG u+ B]XKyF#eƷ@4Z6)njb &W % AkN0 kiđz1.Ϊ Hԫϭ2PtCu %d.CBa[8 %O C\AY@\ZkhIk.ɷ]KD2i`pDPWga=gl# m 48֎]U (USᵫ-eJXNQRS؎q߂KkPW M%_ ٠&q7!JO2sSXr 2ykR1vBՀR+θx .Đv;Y 1I?%%6;Lp`RP 3,L ˸U7uD&' p @}M4Ù J8 9rIxB7[l ƿj B 8F&];{ϊJ "fZA9=#9X^n)Oy@H^Ct@e"29UDYg@rG0lŦ#b|(`&ZL!d(N"5VQBe7E/-2̉D>A` {`lt~MwPS a]/+Oc"`^ 4=5o(~iY :F+HW4Ԟ:XTPbB#ȗUX3z PA"LĴym>DS5(̈t2-y49jeԹtx cl.aHp#[KM;6?m`^֊NO kx~2Ի߁\"P`Gs,Sx=aL)ڽCX4k p[$z3cZ $ P(.f)amIxVȊcdo9سgp|,3 '2259vEPY ;='iVRnԀ>Oz!dlE'T kᄟ?paidHgY gFMk\Y3 \wVM]R<߬tV 7|+撣%|`f, qmGhЬ!O 1.{g/ǧ,Ou5bL<9}.}ndu <:G qti x3ƕ&ju`0{{)f1xknV-vjqP+!duhfc3bL 7 3Jip` MBp\UrE4* !V]z jrE]Aw2z[$#J3?d=raoި70o +TOuCV F*R;yfZ$r!$oZ^16rci.7ÏC vIgnkٙdsd:dOי`ZESdoPgg(aQKcLG`$~HNU/EruU Kkqq4,]{ np>扷Ѕ@]9It5ԡH'Xי萆äe -B3biG iC}J7]2}pOdj8uy?) x6[wuv)fF2MKW$,Po2mC4"i!P?.b/ tຶ*MI- Cϣn<Bhx;jqv/v96^͝V_4:7Um"ud8/97CCx+ߣxj<  ـm7xx\.,KzcҕS FWPdhTW3U(HW q6~Kr>oׯ^kjMk*Ƴ./M}&ncl6XlϼZ?_^ҿnF;䷛?6/~ܥ`ۦ׻ʃLQ}v'4^4y/x_Wvٵؼ,-5S 0TTαu7T\]O4Sks7ӏ7tݕc}E_!mZ6i]|Rx?*6&4]nbx!o].Sk{VJ~h}ջC|wP;W8 ka=VC[&4zqwM=(`!6゗2fJb3%zbp i1'9ŬfSAzG>_2 jԏ#G&Z2 [͆#tV +HW Q7')"Zk]bʽʅdr+F1bک=2f+DSn*9/LMGI= O(H|iƍV6afT34y/,HWA=;ƕ+m(j 1,wE4 +u(EWLgǔ TW3U)-CuEQЌ( EWL e:yf+i0Q6Ymbʬ 8fwe_}Fߜԗ盼C6un-cEoP1WWUWMVoU{udn!4b4˸YLh-^LW8GfZxB]1+dKS˨]GV}>5 N?07MLq8=q0]ʪVE<btŸ^6uŔ٫f+>B+Nq3Jt]1˪Y*۽e3gO nn0_ނ1(HW 匁7NиY*죞ꊀa0ĝx_JSM%  hqA [2樫1&Dxձ#q+-0ȔIg+t` g9AµF`iQz)j.$+w>N=IsbʽȁJ#ޞLlp28(ڔ&Zn6}YQI#teUWOz 6fI"]1nRtŴJSfP]PW|~ !&#EWLt]1jŒƖV,,9pc!U2Íb4ʹSM1,qѴԸq^7D)bdKSU]PW!T-11bܘi˿gG uAb FW+wŴXg7G][Ab`btE~t]1%+L=+uNތ)u,uFA"`I7Y˴KS栺btz859Nl 70wpƉ]L`@uԪ A"` Y7X)bڈ)K[Ƞz]9`$銁+$EWD+]WLj6[jM%۰!z|!iN(FӌYYSҁjz9ԫd g̸Qv]JQfg7G]8q{vL]39*r+^Ό(13Lt]1e{vUw=-q x37 Q.죚i4(fq,AҨ.H'=MFӴƳݢF^mjW{8KEk01пt$6޹x&y;G}VBjM ^?FqvpY_~Ǔ4긮wkgWnλ??ucme!QB2 F?_|Mo^wWū>H@0JgpuqVݦ[Ww'z[]7w&3O BXYim۳n[i?m[7_z5y:{:ukW,W[uCj6Ax6ԒvI4ߛ7^rW]U{YOME'څxfbB@Ru[L>с`n}Sc @c׻)!P;ttSNÍe{|ײktWw KK躺yܵD}onf ߷̿.n>ſ> w,]\o;>])?9woSjꁯv_ 04M:CA]F ЦE4L'HhmS]cicmmK hL j:Ц09dSwK:YOr3ɄCC}ȽFC/& u6iH6 ./#Kbbźl֓`q$1:^Y)aSe }_u-5%i'>.V~{gCXf&0fl2BtBI56Բ33JW46=΢!WC,#CpBpE &Cnb"WxJ=}py*x9l6֔]:p힎Ac' \j2.#X>.gzlV?Wtں mMA+/ܮf S|7?vVVLB  MpQ6}1/ɝ-[ٵ]#j,{Yvq^Y;ϧ2ej}pr5*w4U5L7GIG&)3[X#o`^\gXtEĔLH-jun{}fC f`.mca6ilCm?cv" qjFnn\Qa8(`熗`!b7񲣥p C9&r?Q@sdD܃B%3hNRt`!t[;N o@/ޅ%.¾lj`7o6-oYo5l-|ۛcéܸ+&ABr4bV1\J]zmˁ[gzՑs~Z\#?<+y*> #ϣ 0&M)*)FD͒2Υ q@$IJFSPDBSFcBbkʹ҄1bg܏q.vSBh1GaDՏlR|je>>nfnjJ<*~O`u5b>!"_q+Tq .Y:u"<e/O#O|-`"Rt 2bFMN#P2JBL-&j!d(MDK L  :-(K$F)RY490l8u/$y4[=` PGbD`рLN{Xӑ]I{<⎝ 9'9~8͛UJ ϓ.c}U,,챞]|8PfOf.[ѣ].=BirLą\BzGQ{-ɟIoŲ4&qF0:kW.~!Jא'\y@C2go!]WCْᒊ],Ge@I9D!Ayy qXCdPj5 lzmuI[wf:&#VCMْ6ƮκW뷝I܃[[p7n9\}dVz: o]ųKZ5AѭSʩjs xC[%[Mr{:]o{ լ;Ė%,zxNw;獖a6otz·v}E&t>RO"-@1>C-ly⵰ާja3N͓nQX e\KTVb>A<)J $وyl]}rY0H%Y4Adc9%:NFpEKej%D+D$`{-DJEFN e0A_8w!H/)#)BF pt!\D=_zN!xU,UL::wM|C񃕂߆P,-9H'œ3o#!ÌN%rУT;otZT>X)cIkR)j65¦HQlB7JJCR>qr/-{A9́U{k˭spR/Bι ]I"S"JFFDZמ# (xbL^U*$IxDu|*a?39#R龈:y6!'7LAC]Ƴ2=]Ѧmkj={t<T򎆎rk}GI_$r(tJ/ 7k #Nz,mPSR$$TIn8G2C rs}O8g9759#P}ᘓOӮ'{`yTZm/D_4mJ;% 04+x>|( pṥ*P(,{,ݦKe;1Q8VgMO&0QAqBp]wEg}r+Hճ+_}}ŋvM5v7=!Cȱʈ/ӿxs1TdjHiZ9۪TH=g$Py|y'( qݢǪG~;CIɩI޸Qk/{3׭GkNNhR<&UhDEVFI} `BJx2Mp\RbH=ؾoW10i-y BTs6‰d,n9@ 2KuP \ylz +փ_Dj>qbp>֩{)Dkjf`~9JK/rAqggV8;SL!zj3 ω <>`v ܘo9-ZMlZk͓o__/f;\q ]4kzm7#x2j^[_!w!-#&`HBm> 3l2kI?LNƣBW1eddI6W*-eQlT%_'"ozFIcue۝*1*5 " Ȏ77_w?~o^̼}޾k{a|cOG-OuÍ \?*{Q۽5,:mzmu/b;hZix6xVvL⬍:JXX19X:Vuh:0 ^*t~5xj SbxbJPrTH<>" ukxQ\8scc I:T&H2߂cSG>1eJ*ky#587bI[^(m_%a[}ֻ2 {rREpP}T}`Oh*?ٯH}pWx[07}[)@XpeRHUx"Q}@azPaE'3.P 0rH=If`>a:0rԑZUMZ RSL!` G牖4$O-#Z]xq"ksv0_c< տ<BcV`}u%fWp8}reȫ'{`-+'{ܵ} qhzdpBiHǒ# CA! jᰓ֤/- w9mDz/=)MaE2:j|0i8ӹОC=I'mJKVu4Z -oʁI MӛYKp$Q*Fi&ƀ :g(@,J@mrpk1ٲ==RvFt;q0~>&65Ċ32(q1Ҥk7Yc ˱"i͟BSs\^P T!"ڷ԰pZ*R*&$58=]5.lM2ʅe. -ϳf|0QUDw~@a'dgG/ 4 ʩ{)n**) :&Ii2$H%-~=o7ڗ}?~|}CP,*tF*hk@+ 3MJxP+Q#ePlu8Ob,ˍ0+CKDÌ$ĒPWPT(.xM@Q1in@x qC 4WP33),yrT@D% ؎pg .f|xs2<%jl\åU܁Ol/&͈ע`K3(V'}4AY eIzZR!E#3܅bFGv;O;3[,@=g`<,C5D0,CJQaX9E*L@9m,*ŁyZa[?rz'v@A>Eȹ^(pU)Kh(`ViY f07L۪ņ9ц␕]mٍnD^lI_jn_P9P8DL5: G6c<`qh3")&y50䠼=ƟtqgO4}Gb_~ QZ̾j&IGWΎ4[|8Y-  ud>.Rs:Z4 |=e0?UA&<^3tt<;wG~F︼cu܊y|AsX]7n565Or\Kk読 Z5NävL`ⴊXa9ؖ:z`V^Jmva'`[8*+Yz, JkDrV ݽyF Vojb冊y&se˜_KhNǪ_ !v]ly~|qZ5UW!ؔQ_794w!ƪٺ"8PҴKC+f)<+>M>?q–g=R6Fvr}0٘]C6POb[r\_Ϸ j v crG9\KW0_/6xn<0~†j-;.D\τ&oSb^ZLt}0{!TXhiwe mojLMa}n?4v-Vxh+IJR$M`E6:d"2zƤR0!,rKmvˎ /ҡY xWcɺ$wZ6i! \$t4:ktZqy>dw ;U=0}P' _4RJڣVXMR W$hzR&bP:Zt9]e2ZANW;;+@_S CphD6Yiu Q+>8"DG g?>q-G&ϱyGߑB1#On7gL>i6ik Mgu3|M&= X2\#)bꁮ!]!Wi'ʀ%2\BW]+Di`G(EzDWX3p UF+h*bwHWsIy U+h_*՝?b@ {+\Ӊ(fL2\hM%vp]eJ2\#BWVΟ ftfJ9y]t(We8_Rt̑A<@WOz :wVo j}+Di2]QtmmCWL0Gt +mU{CW-UF)@W _w7?1~ 'GJћvs3W߳{\-Bӈ:Mg]7hEh6L@Lo*Ju(5U!T ֟= UF+;og]KWޟ zCWǟOW!])OΗIIGavu:ո,pUyp,ͦ(9'ZU:l{9UqWEg yTU~xN5,tV.Ǿ9% jK]/?ч >㯻nɮgX>fUT>uV=,ai,u>ѪXV_U0Up@Tձ,W _= h2-ƱseLUn^c}F><( BCXՈ~觚>o?pqv9]%sX9qߋ@V¿G'{{+.{꿖k/Nrg\'49/hW! ,&r9%Ss\x]h.k~ Eu}bWMwq5~V~<}TjmxTq ?'aJ>kEv]]oKr+vWw ACdEПhus(^[#Q8}q8<]}TOݬ8eI%<./^/":a9SCo^> :ì [n])٬n\\t9*Viz}UHjgH'(/g_ez.6C`{R]=M|D~5uWNrWfkH7=&0יs}91*;m0Gu+ 9L%A9`TlPRMWVo.n+nm>.MoA+Ckcyy71y<Ï x07͌vS?]WN.T{7voggCd9Dٓ!|#٧QgQ~zcͳ,?qs?x̽2|l?"YߺMVgT~OsWz|C gIԽ CN1S p!juߩtߗ烶<ܗr+mxe КDcs|@šFf?mx](i8NW6G ՃвAwyJj@WNCO #8+yZm:] FYzo vgxQQ:Ylyzֵy"ʻ큋/۷۠?{_iw9/Fd_6wM(/}ҙ:3C~[:;;/iYiD{QPi}|x];+:wG_7mwѬ:oxJ>O|>J*0؜Κ/>jJNqd}m'm.a\{i"}B(^ټ{0nɪw݋#PQżw(gGt<|Bx\wϙ)k/3{4t#hRwsyU[yX~Wj r.iNrlIs$c &KUgr93*7QqJR~|Bx5z/n.]~~u4R7?m-7W/j.'{Tjz^嘬8l))klԊ{5sLUc.TQ*6m)K5XWե?Qlڠ ݸR+UE@1בb-ǝmҩL>V ڜ'r t`7'ZZj  9PlbLF3jLzNѤhYI`b4@[?~ޭ-5k|HydMhqjJ1'Zj; Pd0v!1-.oL.蘸pQ-^rJ7@' ̑{}pV'L#MZ)ʕ<@{peP`#tEt(41T`H]4 4f5|/c- ïUh}GQ"^WUGt׶8K* b Zڋ#Q1'Bk} Ԝ$W-wvZxN5bU71HGI+*op_c[M1ɇ%]DDNE[RHHHI?7 ҋJSh#JK}sU.RHVհdDH,`j6z.WP< i,GΚ)Vէ@}Zu)jԍMɀ1!KutAڳkJkCv:pL4&GxHh.2rN3rS0ؽjؠA[w Z 4@smLVjY @rܪCZ]] Š˳±&6VW&:+ ŅĕfPc]=^џ^6-FU @: VY4o:dZ{jliPѺ J(ڑkj o2 8Kcs'[FpՈTb0VjXq21LhHpu 8) d> PAq)JMdLB5Xvl:!L% Wћ+dW40w\͐jPoB+"X7(c)(dk $4( ""*f"u eJukڃ A'b΂Gܬ ݄Bl nK#B`@Y\O 4XS9(y3XGf2~.n+B4\CAV)(J892r$XT5ϳ,+Rl ejA!N9kcՕHޫ0I]D5=()k'1S+եa`rTSn yCH T&%Eaߘl,.A"b C$uWH:мGwU+czh2&M ԙy r9/hw8V:G$'7**WeH;L'"Bmq_À:Xcnب_k*ՂYe]w mCka&a-A7 /Ml#TYǪd &rt%C2 RGU!ˈc*Jrk=/,J茸pQ4I"iyu,C 5tq,{Ѽ<0{':AJ[SF#H܎m }ßϪd!T?V y E*U lZ .k$ S`u0Ov?gg}\ԯ]@,M%C.z*ce1+tmi,!z4%ny1k"!%:]%@_>1gmG]I0SQA5͚%l-ѧ!)I+y rN1^!B ;h ΪmB[v#h543$2e#kvq5؍l:LMLP] C VA;5\CYeXJ cF8P#YfRu9Qb5\,ԚϽ4Kct֞Ewih52+io kNv\zrDM iorAVe"H6 W| F `j\U;R@4Xq0ͯ\.ۯ`lnEQ$;Հ ԭGwH7ۭ6@'eJac`B*"bjЭq5H<RVO ]j313@99KhYeFIJaväDy آsE6G=Tʍڪ+xuR"˕,)TP=`() e 3i PAz;ZT}fQ ƮXBYDqmburӐpMr,RDyŰ*aSW(eQApnlQGb$abRuN` tB)Ѧi.1Q`2j hNm;c fnV<Hk֬U3| ۸f<%4T FٌP-ŸC69yYgy~5k!|ƅ7[ApWlЫ8Ai Y'Xr̦i/zBLz\Rčt4YJFD5SOGPrm⊑d 74JCiކ`ݬ1jŀrri"&b9\Y59nfQp* 1iPb:HfDmO]A#S1 VO^|~l\ aˆpɶe"SBƣ^7[>8}s,n-r\`]fFфfFoo_co>NӇl[f%o_&eyZۇ`@o6z\dW+juqu5iJ?D<\q>yQNZO޺~8wF۶ VOm\jxCUTaO407*ӏ198Ԩ8Vw @n'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qu1= Wh'JzN M{&0qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8HE<"'{ 4z:'@SwePF@ $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@y@Ď d(p='2=y'PZ'sta&N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'zk:U׿,^0V׷ _vf,td\ࠎǸxacur &ƥظУw7t{;XP^,ٓ:lZ0|PĚeox<_mRHm$ ]pB+,i7~t C.#i* M䏅Zg:M^h9=GEWP\vBWNWwBWϐz|Ltre.pOJBWϐ堏3GCWBWmO(t@hI ] ɯ] {7zFt)-]6sm$9zwyޙY؆m4"Ҵ$ʢFPEV* nIɊK@ZUɳwmǷ5&i޶8 & vRʭLh)1ml[egxzXL?ȓ vg姓 mɳ/YB?rru|xy?#zl72c5wxj_W*"!75~5o&4Ƈn\$[IZJ\RR .ymʴ;mR w`KI4([3)-AM^hz5J>L /u&w*7~uG{Uei)CW깡E`Kjt%Jh+Pu5@]d"]10FW] m*w%+ԍ$a6Gy9Πv#0(43M 7Z4-!iUC4wmu%)ѕbҢ+1j8 IU*$2AOJpQMiPk۪*"xt%1ѕ&53Bſ6T]EW)c^__^ҙUduU/r9W|:uyq0f/٣wǟMO˫v{,| ARrtQzT.,xG~{myv)@;4s;wfǣi8HnSym=f_oqv|mZcw[cj*c'97ibC鼑įާjy8+ݷ4&[8?nUe}*W0yj&aB,]97IV7tNM=>^lvuf:ys/qzzM=$@r|JsOSj$=F'&Lg8ߢBsVӲI[pU|BK!6SzS[8F4ͷ2z&07oZ)تꊐLԤ+gzҢ+t] e{$^p??ߚ=y=yp"m*`Dsnsr65K!6 ㈲h=?*ηqMހ-] l!̷ mJ(N`]yYϽWecy=]eҺ؋2) K溺9ULt] sZt%=-ɤj^%J] .Z-ZJ) UW{EyR^>)<0(x& njF'< ĵI߫\QVMQ2$M`.ѕƨEWB⹔Tu5@] tż .] m2*銁SӻJhɕ+,-A~t8`Jѕ&EWLMJ[u5D]aJѨѕ :CJ(m QWC{dӒu%Pg_`DN <`U{n4_]BO(;Cd zH` XȐ ^7] m,^WBbueHtBčjt%x] ei뮪+9{<4{ґAAڵ7/7;(4[πWiACip7Օ=\"-bd]JWu5D] pjt%}oiϣ%,]WLJ^u]EXE`jt%{׵3)9Rdd o,Xb)fWKqNgk-2#IW#'{I^:[xU{1{GaʋMo@*>j ҇BY[9F$RFPU|ژJוP:1D]4]1n4jzWBkMJg^송#C<?Ap9]&.|6mm_C()lЕznR0Q885hJhw}JUWԕC,rKWheK:0`a햮r- AKFӂkQi,mjz/#JQ]13t] %Pu#`HWFWKI:5.٪**1֨ѕ:EWL `JוPڪ!*KN`Ѡ] U3Z_`P(V] PW:M+VJp ֕?J[u5D]jA"]10aKJpAz3uPs$eՕv] CF_Ȑ`0O*6ԻʣLwrUW =mP+N{XRܾΣ-_WB|u N] .ѕt֙GYZΪ0uJ).A w 9:0`v5v hU pӫܨFBtM %٪j; Jpiѕ-UW{UpDM`gDphѕz*]WBu9t]1n0NyUWU&[xҕjz[ uc_xbr%N6̒* %Cc6-&[/΢+<{V^d;kdeʊ2t,Q.1aF;# D~p}}t|5;o}nǀY Mv<;_W}o|W'p?77}y$#Xw~M95Vt\ѕK# #m[BnyxORme3r1[>EY*)}ܖA>%{6&mYGqݦĭAkWSoZbḺ.z ٢=X뢖zG=&uq{L"uA3|tJ_^3I$.gϚimmFyC ܪ|4!bq VFA>qKp}|×wL{uSg&Bw]%NuA>Qg- w9&7kwO֩%Uʗ]Ȣ7;#@2~{%߰Jȵ)SNe40nQ<e,޳g&,J4%`:_Gu}cK~_cyMV.ٹ9F<]ʏ+d7 nɣB+nLƄw'߹ q-\W:d (a !&|\"ɯA-yϮA@%ܛu%U]] CGƤP@* 8H-4ߚGYږXuHW,\4Iu%.U] PW$ߧ"]1(<ܾђ+]WLKRu]9c=6X˿pdid|} Sm--hhi Xs] . J(銁C5܈Zt%T SNaڋxdH  7ѕRALPU@KWBR+Y=|]E w+zlxbqq<`[+.gYbU8aϖLzq}:Z*}rqv¿ج/Cj. x^^mMn'fsl>>j첝l>~jv+}?[Mѵx;|kklc#/|; J/Y!Ԅ97st< }Y?}h+C%)iԺl7 FhՏuXU\ra8V VG峗ES'|EsҗEGˬ_gUoSO.O<#jDjC R@}p(nQmCVhJ k3z~WX6ҠOMwr+!9L@qw͓w!zWrPkft7UV*_n:ߋ9e>z5U~=Lʅ:$rM,ߟmEؔ{W 𿆿kz^ngެKΫtw$ HhZLqeӒ`H3!A**yYUJR|*Q .-o@ ʴ9p*l)QJʂ<49UI pXsU~SĶCd(-L3`7lA3`ߊ6)r6vLq<}[8>3!U Iv W; :#1 g 4>+Jzӳb'23ĊCÝ9eM gBdqNhLVT.:-ʏӃbm&ՓLN/Ќ$FC8,c^*6*'Z3Y †J &#xfޅ3Sntof1{c"Mm&$AsAۃO_I 0nHY[GꙂv,%lWieUHtA.͈NYe~6~2P> wVJ+"/A()Y^ĥ,d%3I+TҜ d'|C^jT9~Bvh…ه*ȟz5[GEJ dYU X*+kL=cؾeinrf3ϘIe%.˄drTdE\6C&†MtSn/Q!2}{?l-`L0HN$fvI$ϹJ+(ŐvLF ڥP}5E@ ӨqS#JU+n' /QC3èuq1@Ia?b3wgpdx=>ެ<Ą#qp5PUC11og̦i%І `,?9,(;]Ն"{P0Ue# ;+vZ7ݙTv;c0OK]Y'E[^Hk}~O|O2e0V' e^d<)2pXuox섋.KfݕJF7XOwzVm`҂ՄC\P3ENsҢV ,-JⲂ,RŠJɩ2 FNKP"|Ivk׿7Z_w;_.+&!O" S%6T)$0a*BJy^E)29շGTscͪ4:v8v=ϨȞ;TgMt Imub2.`l.gŋI؈)\``r6J}3V$2(FP3VAL WpB@ΞgdC\"c0kcy-_^[q9ą+#{^ic%LNJ;v`sao5\Æ]{k F쵩HCi% %8(+cOqO1Y {7Ƭv{ ׭aW q@z# +d"JLQLNp6F#aF#V\$`X0YRRn !D]GjffT=Y&߆Oz j]$ %$R eg}i-;$`AJ0_Yezξ51@u;(g Ԉqxf0Q g9AF!V\Vv1 1irLg'C@Y88;V,pkM^jf1LE珺0zS0 l2e[ɑ ʲLb .)n (q^]rrAG^?n׎}_Sk.fUW#.:|-7zk}z)zz!~B~kLGE}a527Cc3Z8R$Rșng?zB|\d8\Dr U*ɥHߓ}+ 78 b/ pC@k_"[яSp)Q/(l>N2TM. Δ~WM7.YyEˇ? 'C~k3>xĿ\݀o1{\՛uO6>F(g̓ :I 엫i9V~?>/#˂{:F'3}\+}Όy=aTTCtN~vker!q W\9ͤo?|!BH^H®IؤdxNrYCpDŏ+ Ix2_- {] ] [s$jIv⨖f1Q%w|;t,e O Ri7C/BEy>mDF;wOև# ;A-B6{.Tvc%;^FRAb#oCP/O1kp},$.y]z 9-Cɑ=LøO cFk OPRZCt필V{9D`D=]R:\uS ^œ띣YyƹQ]BmM⛯ap /qiR"]9US1= Ơ~_ `F*/Łrԅ^$i뉾QPX̥yDTT+$` dYmŇRB-s=Ƴ,Aq <0vL ?o\{paT+z`tR{ `G>)ڨScd .em ͔1YdEh YGlCP}=/ov6}vE_W $gI!(ZnaF Nx?Ul- ?WQ.;>H&o/XӖ29S=`Cs5doZ9TTKmnaNz=! x):P~Ƕ2d[a9|Z&]htpͨMH%V]Ӵzvtd@ٟN(r?Gym[B_Na Bz@%DmČDߒL3,'%l y!;_܁b !7-EƳwqY `ꯟVxȾH@4G.yM?V!,HBRyNtߡ΢T "lML:+#*<>.]P;ߋAh'ܾE$UQTX;jA6w1c[F"Mc]s G1O]2Dpd(óOt9#p$HV95:)7qz+vL)3mJ$zh ۮ|[`7蓄Wbc8$8w/6JbZ+ׅ C<wAc"*6dqyX 87q܊(2bX/'٪#A3օ}4iSeR(|@Fo]/w۶~rG@5G'(1sA[Sfp[x6AF*q%%e%r])"'X:M%DCdf+f'Շ\\mJ"l5;mfơyCI?Y$z%@EB HHՒO難$lz,!&-TiecSш1qlZ\+N*&-MpBcJ θ.HrcRsK3t%\jv 9YA n@hIOU}$n"uRtHk]08fPQ3Fm2nÕOp&ohdִvf֐2 WʘoZP|AdFyPEƐƜg .Y^yѻA*. Cx.'|āaD3BizZFnR'ufDXB(Xc 2 ؁K&p9k?ޑADTV}ls ۟~nƬV4A^3r.bӟ{f:)wf]lbd/oM1PڔRhkr cMr'2#^Gϋ2WKJ{?͞?]^L)<`{zAKc]qcatcYc/&\v-1g115LGg"_/x( X/Jlz ߼EȜt bzDiP[td=sttg텣D6J$+aY| Nϩ40L 'Abk-e3 >UhZ5ZN*2 ~鯷7Aa.L CHix`wR=}M"u&@h3#2^ĪnZVt|N$=8(W)/mc(ˮcW.4mw?9Zwغz2N_R޴.5gq,R۝<^op'pj5qa{G-Ô^r\.jF ̨/,e/aKܷ ι1\?Vk+9t.0 G(N{ Lco6X/IV<~]'NQU\=:YJ~\4 O_k2Ee~.R]>/m)&H8UNZMj(nD|i^T^2ҹF [蟪8ZlboMTekv {͊YL)/5u(ZD2ݞQ@g ?>1IW }7?3kʌ_՚Ida|:^-b'pVe?޷W &<6k>akL\"p|Vvg?@"oUPnК/$ih"lF{=eg[K[T<.Ss.4Q}l~A]s^`UlgtR8<İgam.[CƼZ[ -D۩G@r=-yٿ>fN_ah(Y=6mc$qyRkMӞ2L3y:VI IMȪ}>oU~oԉ4LWlQFiS_xEsp0TАFF*ӈ̩dO ƼA^lJ2#>&OQR@m:x`k[۾{9㋁}@P!\D@x6h|.1εQ>\10"=A- 돋Ùg=!]oC" oM31AOMc1>o ) C?ޭ-ufl5P(t.3~My,pP߉p6Ŷg, 4`Yd'e,S⃡`DI/5A"ɭE)qeImwWDXEց?hoq(S4JH1|9pT "c\E3 9|0R} r-]ՇLC 18DAցml>f7n`ɇ3t (ЬSgJ*@GM([pj߸[ؠEƍm&X8#+J$G\I4%*.qʄB-2~b׭5X Nr `u|8;Z} 3,WˀLkʟ`,[JU :*.ːa:EI" V).z& s 8!C+ BkY͞FI{G=+h~A,%W5O+~'ϳbkf}j~L$j~Āaồ2)D ֛wg}}hh;{8p/+auA:R^ NSdx~6}q&x#Nh-5<-o]ZjZ (*eD(*>cz/ Aܿ坐BDA޺DMb}eqe7D =tMn;\ M /; Linb>>mۅ{F8ٵ:pVfͱ菂uSI ҞzaVnѐQnׯY -2i1õOZ-T|3† 73 r(@} vi+6O=4:ET!n:PVOnb&x2U%I^͓.*ɀ3ry`֣;']: 䑬aWOyG|47bl+n + `l1CN>Yx"c'v<Z"{u]#PG'Lռnt8mE ӬPJ~^HG׿Q {+ s@a;Q9 ίdܕ1?Fd6/Ձr8U3]!>ܿNqT4ac&h黵Y=-N )\~-->ҳv/U%{-RM4g,?(/}FF73&/h<\WluC*Xj\ΗqB[Q4k<<&/Ubf/;X'PJ`vye08ib-ƄEle2˫}+NS&^ 3!?ϾeUkzawmF_are!PUvRKR&03Di8zX[_iH!8CYb5ݍF<$9hZJ;s ?hYUit'ƃa*8 ,{H0R˄2 sxML뎀HIr<썳hN ]UI K1^*n4ͶCI^V | .WWrOWR0A9w$c8/lJsY,%uJ䭕LiX6‹V\* u8a 0əhZ hLQG5u_-wEy1[x{6$DKёBj R.5cz5Fd"u'eɌX=?$HݙlO233v]̱z{`IedKz nKq;6~;۳73]SbIoRqkaIs! /(X^KNҰ 1/ Rsbs*gWEPX+u[%uY3Ûtz7a5Ɗ/V"sycuF=&D5>Fleb3dw&9ipBvxYK[ yNwvP{Iؽ-nyv(&gB.'iJ mJ&nR=93-pxvL_ n%Srncúzo0f49WEmM f$y_=z2e8oSX b1mg{45ă!k0 -"ӂ5qyjX&LIby{lax9m>3R°no2k,"dޒQF23܂A-ra ө$ol!C:3[Z7n2JUp\6W{a~/qU `5U~gG0#7pǟ.zrvŠ'| IY􎜎Mvky&nP|r3$~ʆ'a`6yGoJ7KkC;{YI,k!fD,sƕ1-c5Qpqus}$9&]?MfH0?NTsRrfRG@FPu\mj3r Հ&oӃ\2}\g 6TԠzVw9IqVo¢AHK<meAx^g|=2O`ce:c!ĹM P˔CC-\u=ZJOb4Hc5Ƅ?qeh;`ku F2sQ +#;zhrT@Ga+;_z.=K9風:&Z0 f ѯ`S>'dF іw&]sG,k_2pniYk9 O fmӅ H ]( Dpq[K_0~l5(m)$B`tQ X 6Oh4.@&Ռ3T(a pxCï|rЭ@ܢ}A>Ӗ lכԿZn)&Iג 6X]A}0u镬weR 9HP)|f+r!*ƷP*:[S͍{;*tô e>ގ_8x޳Z>LhY/gQ6Ńňu> BL[ 4L&Ji˚cRݲ/%ШjhVkM J2!~Jw_%_.o1Y)?:LaѶ#p/Qm7x}}` mAX( )+9JJ18rrBmT^h|?m1},^ ʼ՛ f* >cCk[H.ha~%Fke+3,/[A >IK~Ybm9Jt'O&Z븁ƷrKVG.okX +)L}zDKTӶ 4^{ȀUK?H $9P.ey2%H~(@ -kEmn;@s?)%X {[=wnᏟ^3ܛi d2#nKPkmE<ĺC<ĺ9H !H-ZzI 'EunaJ;b@;͵ DeȜ%80O]r9} }{?m# V#:}lu;q -Heћv=VSՈ^H&ՒʶCthN=zZsqECvrY\|{:SH甩hٶWbA/^/]P89R^ǟQ]7fA7(dv>y^pRGac X/K3o,sS8vijxUh ӋV1d#z`<$=qn49#XO}Jtu"̔'|c;|@zHCy=uR) Ὺ).}̝єp^x0 TBI_*`֩ұH8i!YW&#pi@ ^uxz7%`;0P &Fky7Ϩsh b?tމy}=T mnx]}EEەͽܔͨBwWpj$Z0Qy̐ʌ@A1=]iWkE>/[Ȥ) xIcW+DNwRMKRkQpArWD!Q5lw\T$>4kr?Mx 钷* 45+wq3.̥+*zS0AO3fLc?0y"Q\H-nguNfgz\3.!n34]B,1糞mr2? W'eΕ5zS/T@zM0fs%{A')NR~ Wysww`[>$U-<0VUs]O }r<$*jY&D0y݆ldk 7pƷՄFŔ|=鮋y({}^^5.,'VR{Ʉ['q?`jpFz"^2LCb K{ ĬS$ kRjq WMk~(V-%j}Hs`U9%#o mR,C٬кPB=Iqö+5㠬7`ig͚=XHbԻ&["sQ}`iU[9BDg4BWb]GY F;6(,(n$R/^C&LA9%p= ّ--eHZi?z~䲬okX; O7IE|y\8&4P aKÍx>܆n2AqU/*Ac.ZC4v\ ɭrD P3"^Tk2Օ*IH jLApV]f,3 9 "1F,245G+c?01h7#܍P4^? |DPY_}ycԓӔF|O6NFut𵁗 r +$8oyv͢y!:Хn^EiLzU-;R7*Y݇R|ٽ/HSR6XY yC;A㞙ЁAEa2wu詡*EPbտuڜ$8)% XT5i shm.*ha: k3h(p `3|"<A({t,!E5h`GWkc.bժÛC=;KmdE'Q+r_##NB#G<hHY0îI2Pwmm#Yy] z;Y`A``,j[Ktd0}e٦%*,Jjf'nS x$jiV] jF&Mud#?4bGvs5mPU%{L4 fh.$f:1\<+ӂ7CW:Jt? "67SUl+l0'T< nM?n+pAXc Aف٬4Ųh%tnEэN-j52η#MSo0:ۈXǬ'œ.j!5⤍eiZnP^v Mx4_3^^VF3LL'*z#5b/ ܒ>Sxu $\QiͩFƠ 1f_<[u"kA K*4:-ImN$9}CT_#,*ܛ` ,b~ը0wYêLѽ6\oj4 z9kd 4/?;>J)dלƵ!%xvxBrdJYug[|G;@6 TNF{,P(p"۝(0@ϱʓ4]1:}NE52ΰUG숻bIvgFƠ=(?bF_5(5Z\#:Tl3ja-y^մ2γn7"k@^~C )Q. N}s챬v T$9z#')=KT&n!D_#cdPdqmM<*DHK/S$&Ɔ!&$|[D ~4:}9I5q'9.2H]gB9aP(5axF@8c@_|C"'`hgH$' f3$JA:˥3n)OʻVNkؾ&4J`RAڃ#r|Nt'9PLeݕYR$:G_>w2'֩(;qi[ΐ#Mk}\Ϳ_7`vaeQG ߏ NVjdA+B>G5p7t9?,.COV,9oPdiS:M#Wvv$ƲG\<_':9a F=tMtv& Ob*xjhО,Q#<&$'m6Kn>$ GMCSskX ާYY8,Ѯtcx%Wc\"2 uBXl4AV!#7btƋ~~K!FcdM$}pуDzf(3y8^JUkRCUB)"سqWoJL(H~Ed巧 Xn~xb$.}ޔEt?~#Yʵa2Or- e ٜeNָhs) l16?F Ԗ &ig.Wp?Z}޻>fh啋˹CwQMT. Lb1n2.J&"|/=QK D_M~ ˜r̩={?"dOl!hKCg^l;o_P5Ϯӝm(y0B pGb Vo7clS4}:lEi'ofyuv~.Vk"HK>~WKvhy*$1ڇ2Nc50Q _CѤ?_gyo]-RKYXl7<:Lq\ܓwySeo7$[Rs\r/ŔzY!:t'pqffoq{[.Ply0!<@~ikU%c4$iH+cwN'o<9r!Il,01ےe-X[Ie*Zmru8S-࠘ra[#84b>rt@GHOh.Gq 01]w@ k&zj<`EzaP :ղϰ]d(Y.P;pZr*5"NY/6Y|OXe՟2qN\[]<ԯE%7*CtMRy@5 _] Io3{j*/FaO%dd:ZuO}LIJCel, cBZr/VWȅY] lʛxa{)1r]شTiF `Py9ؗH=Q_vc$nd=؁؋A`YWW >>p٭K* µ0_.\o-8B};|E5PHEEb{ذa.AԲk've[)0B|)@{e._N((MoWFbpا0i+{% c͕2/cZw32Ef{%V}Rc:ϛJ(N8-*@A4O01Ri h&SPIۭkOc:kXy*ky59ۡ͗{0OCD]k7]\O'Swg)RھT3I49+=ǎGƱ/Wm=>VEũ #tƳȝbmpߧD޺ONJFt KH~4 ob Ϸ Yp΄ƞ.5mamoYvs5mPOjd 䎁czBN!W3qD]] ʏkdo1GJ?6YQҫf6qow?.~8^ѓ1_M F:%kd :ΙZFm]&!Ce\_ !d##|0,6+#PyYO6kdTĕ*SZlTtS7m +Pv'Z0!)s݇7$Q, =uSNY@/Uoך!EJuqU-4xh@Ց"ԔtnI#Qkq%ggB[YhYKBZkZ,L#af?l/N"ۤC QWx;΋LNj\q>f$AV MW o~>tQwJQk+€fT)K++}|%1Ƅ.֬} tZ=wK/4KO>omRn_ŪGIil}_|z-gwRV4^w?8 kqhLe|x$4+T8#&(lbtS,'[2W&՚ x2SMFD]7R8ߨgtQQpwtg`,k| rk\J]1Y!YL}_!UW b}`znVQR2kE=B\$:*B-y\5VRV7ۯ?PDZ@:IQnM@~Ԓ .́/6/6Y|Oqˢ?:r {zi~-*#[XS M\ͽ 7RT2gEiyGLq`K l>Mw?ҤyĩF 18 y}LNuNmfTFĹ }m]H6B!>b2Dv2/~-n g6FPw}Lu?xfh+JaskmbHg!0m`R\p3yf7.p ޢ v_z,r6<Ĵ| Ee,1>D[F A^ٕ4%Vd5BU4Z ?|S&v`KQ/7jrr)WDe)I٥izDj. KFUAA'7G*NZc.$i*sld 2 PՉNS~r[ jwr[ o;4b;RI$dKc$p*ͪ#ciY̖XcDsX,nF9JQ- %B`u,8T(12λVw{>R&]2)ĥss'L4S QְeI3*-X8< 5_[8D)_ NI-L\~[AvG[ܕ}&7\G}n\)ʱ~Mָ*T# 5HkCnCzR\."¡|j7Uf;B)5ҋqyN5~H. Pwv5~MT2׏/دfسCy5 #mۙ䃱$ޜkDU#MRPz ZW#uM֞k臘 c$y/&(6҂z 1L`lG2"JەkX9ky9E()B>1bJjU m !)LR)oJSz1%(gbJBQ8J]:O 00 1SzEi1;^}Rn43v6(t(V#c Ś 1QQyܥAFƙ$B("G+>?,;%'ꥁQU7AYWs a#Y((c?Ua\NSjdؽF`ŎRHԷ j 44ةFxj ̷ HQoD!iT5d482;.49&f6B cQO%ppBe̷4v̗T\!ϵS}-N>E491gŠܥL\>et3SJywcm6ݡڿYQ\ɟbh#J>chPo4FƠƕ5Ho/$JA XʸOu?ZXEnc*MV4tO(SU!UlcG E_S%sXM"tb~lQ%,ؾ8KwmtlU0OIn ~Z~J&b=MK"E ۚD%r49]}TMw՜ػO]oK%bF3-3V,vOo_vmQjhQY2_9tLG^eEw5*Q]joQMG{Y%.'3?2ZpXn2[$LbWWqVTKE}ܫ\H.smXcNv!Xę`kJ^g}M:[][SOUR^_$86Q%&Ìjd~sJõgE=k`e^kQ0]>%oO-bpHhq }3N _9qm}(z۽7(:VTDg'1mFϗɚ/-x œG9:>??9{{Qxy`ώ[zs~Pd/ӛ'Ak& V M~x$> ?,q5c-n~Π$FĮ-@<5@Ed,e:֑Ή9g p ;=o<`I96gn>b-Uo{Srap&U9 Z\ʪzoȋ.Du0X1#jr$E4 =ֲ4Dn2rzR4%i3:9}gP$mbC{>+z4Ox$@#9܍dH>zz lQцRrHپlw۲VrDk~{-:n¯^ymeN&$) 6H2|ٰ;Ebƽd^i7~_5.u_c Voi j<V~lF_@:Gh$x牱7Y_>#3vHYm>-,y<Kfge7+mH) Wv1);6vKGYn7okӛke](=GN=9։^.^-WxA0U}q͞obb{9ŏe9Wu̓Ĩ^*bҭDv4P0{zpk@bFёMj1qvQp~ A=]{..^`BгO;=VoϬ^dz~…#m~>moeF~Livn\As:%ߤ9vܸhv͛5M0K/Sa=zq4rAmgij-'lϋ.>i'L:Yڭ=@EbK#֬oN؆T.15AftQ&Xfi7Kn~ H;snh&2&0&8KYm74ҥ1<&iGc΋zDk Z}Y.2zXlh\y-"?0g;!~zwJʻ)[a%NWEk̃K7ڵy\[#_լGALٲȑ!#vNL a\ws @EF3 wVz.$)N$ﱨ$wViM-w!!.ڨvY S|o9z~EJ)0ߕ~4[T1_4H:md*+WH\--"Y/<x)@ˏZ7ħѡy#:ҞGO5V|dkbw:صc;9mؒY,!Db[cC*}ҳۭWmFtqcCBlY$h3wKonc:O0˞5?lZc8 f֜>,;cHƹcopD ݇tVgbYHlÌS-ɱw 0W0n;g b1 lV_%%a=?|zUq9v#BzJqcRWs)y-XFW|ˎu8S<("-ue"^ήKu׋'߂!p &4C{uUv]-꘯F⎬wlq m5l0ͼk&L3ntU]}$*́ 6Y NKJ9WtM~LMp ti=9YOfl=ǔ?uq\'G]9}:Ow3fgH^"ۿp7KJc-p߿[2)GPl4 SoACk_ϐ]}ޝmrHFa-*uw>Rh-hDXtaC:iLpы A*IJܪD'ɧÈ7^^\?2`^Mۍ&hЫ0᭄w8j)Bg&B[ y@-i8alгM[]o(ivF mL5Jwģϡ 8YU$|cW##{t-υ=:tzKJwv1s&b,?.uOJenEquEDZK813#VkkZ  o% GOVS >;& 4H'/\@wo'7G!R(XX{‡i DSXKQD0̲A tv j %0si4Ug.X=3d%'M +oJ(Y>`rBb`OSB ö1lF0}8X*m }"!CECc)xi`N c5(_|PlObr14K,) j7/7H_S9a"DWY'Uٻ8r$W~٧8I0>/wv[>ڒRJRܶmtf2Ȉ>2!˽nona鏨Z!osFɅzݤ!&x}i/J,AR4[[l3)7 "d]&|0SpDz0np}&,rj㤥O1C*L`Nsb Gy}`y%fO hgcdxϠY ~a6VzxjHL?5?7Ibq"ubPyдŽ{Xn D*ب6@v -X1g"`K MV*Ъ`}*%:;xdEi~SF8zȉrIS44پ^{-K[u%5gbxYmd;0ݔ"g .7콇y,T-u\Y;`6{5I#G!Pn9uLXX{ 3jndYYYC`Gh0/Tωso7~2ҶrCդ2nǞ! 21K60C.?^D_]h=@@0גrO`_~6mf`K JK]olvËv%ֶTϟ|{!>\VzLQϟ.S< YOh{m*nvXݣE._:|+l:e^G*Öz1^HBr7L] GLyV(ټыLf'dv< [9oRa#d`Б_,nyS(K)GBvFeOKT Y]Q~u8ٻJO7R4>nO rʑ't~_]\p|WoV[Q7~G$`o3p#~e7Q."/N̡.y[;Qާz@< XM\9:J `.닽ꀼ7>N>O36wJ?_!1!˖f; 9\BXt ѸJg.\ Kד]W:eDzN3dSljNȧ~In\Vsqջ +j% Jl,ޢL$5!èڑ/m/?eyg}u{X;A`gO sіU}VO3/ogf:U;FѨU0I9|(A5 e@*<[>uGc:eÓw(7O[=u>=."7~W7k 2t"y90D{'JxC0t-?[h!DU:uq]欝6(9KاY^eazP*.Kl,6Gҳ†D6| |ݣUE7+"UPx}v ¥{XXKa鶦5/_k/g!Dzl CǺM ޼zqQ^iS\tzg;>Bi쵚M*)|xy3ckMdžn)fVgq8&moץo6S qp[".V6-|;磀XyBj#/[l:/|tw>&ẘL QZJs† R[D\{"DIU&ix!2eb~ViH#QV_| nuy'Ҿe/Ξ?{=Exnl?[Fk8Ha-^,GP*{f_afes.0 |3*]m(zb '3I(`@yq\V|1Î-B-WEעc&fN* 02AU>|?~t{WL܇X9z͝[K>݄?{x0TNFmOʹN0/2w j 8U+Y4H 9P\v9`ufĜG`%?A܇+ĈGLWϬgݙSx]N9ѧkqUޖ1lo.y$EWX PF Vya^0 " Fe\BdnsC QpD%h j !?ɉ/[\ln3a7v)U#g~իv_C68ZK1bI[ob=Z$u'm S ׆\}mHSȱ2_۴ dܹVO>h6̇_k*,LB)qq8r3kTj#ˆ"$IfC)'֜{(!V5Q? No(2߁//y<蟢nK_hH;yjib>fjg>1{l俞ک:\gbp<3j]iIʘ%*S7cKv^'Zd ”8&,2Zk1;\.У98C9% ؜Ak! n#\b s1VOs2)(Db%{nSYϩظvkLή!';b+W y:fz< K=o%ZlGO чt³i[ﭑK[V [j kAnq3v}){9jL uΤ2vok7"X-[͛1]C;Vyfk n-fj;V }*eLr&4b+19qRHR Q+՚ vl􄡺Ofpq8L.3wwD^]hPa6,z#Y> \ r=V6ӌB 0^:Տaz^ܬ .0૛o`ʜjp*. vI86c/g3ث^X1 Dj>D0cf՗\+펟Ouqzkaf|8܊YL:N`D#9jTcKMPzDFvb3!੸}{#DGM\ŊmvbA{,p! Zj6U()e5$ed1)FZt1d N2cc iT&Y2j1w~'9ƣF-@1rwQÃF̀K;8ջ%yʦ+>ֻ{e ova}Kdu{]5)K.lfԗDl7U(` >OG0^sQe(9v5&XCS*3;Jꮪc1&p0-7`Dɧ'ǘuͥA1I:#f'TRI'3jb\aB25/SpPIp̐ k߯.JDB^}z T!n<q񀩋62Ҍ#RJ0t2 $b450䣯#JN[EFf|3̡\4<>6c[q$ךȁsйj,\-$if]ER*^j%32Nb޳2-D:D31)e@0K7%0X+FMf[<d/ԃdn+= W*;Ƨ|=NUS&g;hm@oVђ2]WqK2FCR276k2fI{ԇ?9X3am_ s5C 1Z6lƮ"fRR"\TraMJ;WTRvg0F/CNfF.ѦP{ލyf^dz?t3Y+=78q-3'vf{YٛGb7#b$z ECIW^wjjv߶ve6V,[hoRЕ"iGT$ꤦ 4htCD̓$O0Z=¾Sf!v.^?x^o䂙 xM\_ԢXclϙ+:w4}m*qvGv~Gv=4sR1ժFUAyB.#JxMЎt샧ug9 q0aLO3AV C+#w;$~F3r]oTC{?oF}d?~ =W߫BMBz^v>c\ h};L~ܿ|q>z{si_}`"y)z`NUsĒpo &':cnzIG~?m4}^hGυyaѕTJE홬[9 3;3oqt^Q\:a摵3nFv 9`39}˞mZ~K6tsX|_|{W?./ׇxӕno*ŻR~:ǐc/`=K_bs 24\%w|yvԖLs  Tz*ƺf2&$6M)"1uf]zWWvVؗs|-wyr8'VUxđ j1 g>j$Q/Rg\-6.- 9FWKdۑmQD{}Kr{f,@.!0Ԡ&J=#fW(A- z0ZvZbO[мt.}M\Yƌ؀($~3b(%| ES 58dV Ī?VuB2+Y$Y ViwQk/f`7~#VYs% /+fu{5٩Hp/mL, n֛9%s!ZA߻UC9#51&V 2xjC+K;լPW GURH N`B29r^5t$1r ?~=o5fy\ ;&~?mRࡊZb{{]cPn?%J²S{[ Q0ɥruH}/!7Xoa㹐 " uF\W Ґs#FQH.;/}ܽ L~i|dJ!п6A<׍X!!^B˅A`9:j~/CX|\"JZ /$G[LY,ŨF\U΁Mj A\iEŨF+!^N훑9$Y,&%d%\T# vb@]((Q|k^KT}/1R͹Jy,/'^1-[ "ط)LQ4swc#|}K B}XTƺ8T+LVpS+RM >))T&Tɂjs>pnҲm yIb矗z9$OCKUo:F5hDWY yB29eAT&ؚRI֕ (pUjU&IVmYr>|>8L@7Gwq^gۗU`Am߮wCNmN튍m ɀ u'RrR.)'~ 8P^j~ x½5?Y땵ɝ}{~YՁK7g7|Ե Rßӻ׮j~=srܥq\kOm{a֛AOĽ$o{nbך.=:/辶xY}?*Ϻ-{eȁup)*p6YqaHѻ!@paDv 6f.rsT"Y@cS|h9r*XZSJ fN_Wf\5s{ @97==!&cžm;e><#=웣*xǸ0nϿ_]!,i~rxfSohCҐ YFm8Za_. rd\  B5OW>}֞ÆD0Pr"v龹n:Cn:uӝb /PҙâL .aN9:B!5,B2%LPt?e /H0>?HVPU!K. kM+ajCTY9 |lxg"Af/(OC'X(Y!CAf'',k.},㞘rÛAOܻ)<se;UR[C\0=TV|RBBF<^i1 VѰː ="t^O}[9kLi{H6>TC FfǸp/kejkTAJgby]&0H&!xK4㚫_PĠ6UrlM3覸# KO_<5scփ|Y.;(PI'\4\q) Ee_^ ;3$V19ntۙHll%uEdSF ci4=au8-l(1P5R#OU %V㛄~27LM9WD_)iH:#) n`ۜ1E_5B0"o%⡠2G0P ETS)'lXTȁtLfy4c6'4;#ڀCwCN&iA# w1!{A?S_p8a{2] ˻dƜ R<C&lC]f텫&T>Fs)}ծ+=2:pξA9tJ;9ubJ-Lj\)&!( с!'AX:Ӕr+$NڪiUwŇ%'I)%6?z;d(BpG{;A?S,"[G7=z,xW__o&Vr: jXMB׻[Fc'V܄agF#oua :fJ%DdI+aq`_5QשlnIr2 a#Ո!쾮^dUŊ^j vz:ق^ F𙵮O1{{A<&(s5{e ^պr23a5@v$o4\3QuN__sepuk eN]i P#swn6}Rc|Sc8KmER ᅱ:tzDZ׳եzwQJbN~򔩈aG=x.9\ q (Ȝ5ǣ훺9o⥃n]F~MM:f: >&\teoыY!jzv熖7]j֋?G /!|mqEReEd3&~nzBWFMKFEoI>gښq_Qe^ yHen>L%;S)^mg|[\N~@IRn%[u*g@@[elM.G2/ǑB(].of󌆇æ9l R#8$r,8H8IG㟣AJ ?1]lG_IϏ;u _u&BBM^4,"û /oHW,Ylё4Dq^MO7SkhiL|bOp(./'Bܱ5τkWax K8.u-=M|0K8ᎵxaK>DgDŽkːqD"_xu)tw;1*1 gFY!lDT:SA#']?wacePJ({s_>gOU|#&LUdLV.%U!&4GnT{K}omȷMEv˟]wKEO1*_|u;F->Y"i,* SUe#W E+msg* k[+C)%k6?ڬ%`e5޼]}m-# 2y5}mbn(#~C_dϖ̈́i;?_[0)FG CoѾOa8A\9*B<tp 9.-&b=|Ӈ&&>ˉKRRx bùf\t*MN/ɎqMW6vq|4qAru:uNG55DXڽCј\<;H`yĀ#eo;PaK]kmU;,_tܣ5W'` 87z|6ӗе;1 =d,8yt,>` g!SqV^{Dh6ezƖ6c?͖͛b Ww2w3/;c{ej-yw w܊Qˁkbث’B,$}a>IZ>I#IJbLĮ%/pܝ^`ѷ<}9LU@ ݲR7G*T')$eR; z E5AF8rC>hbWg_͵6/>cs;B+ L*vA~4/\b@vd;Z޻wLwLXx5-%>uF! owJH*eы/_=]r"ʴ.}WҩIj-ɨLd\&a xR#Ðg?XKPimP|IFVTuOiDx3 pU@;Y?BJ}h~5pT6gmLuxߡ !c!2̙Bevt ɈPI;USB)\aK(1Ȩ@YQJn*6 F#b O#lc9Z{2Q"g*bb+ Im,p#r!ZSd wgsD 2G ɕ! L۰x1:n.(,W e>TDVHӆhu r@1L[0&!Z(LהyQnn^!dz(^ּwҼb2liXܳOj+FȦ1 Yk;XY흕x;sA J"B^gژ>=J$}ꄓC)T6\Zcz|Di7fml`&FOs(,R^}iň"|"T.ܞC%O>QAz`*Ct>ϕuXQRd+o<1b;Vy:  <9#"#Y}6hho'A) e /JYj"%eʠ4-ri&RB{( Tgԉe᪑!2-9envprt4p2s ْz7>mW%E%Z$[y$a; #BX|DnR\*k Kސ 10"NU GH^ eD+DDƽLBUNˈ(A8o]i$8hpvCrϟV@"A(K4ʨ%xBPBDU!)bz WknâU^[} u0g"h;ˢ ,C}bLKT!(%b$^%or'/u`6l+K+T<^\~2t`M(,f\ 8WI|xapmS&X NjTOmWgǧ\_0 -TId$++V,"d%%E#!&r$$tvF0ydcW 5@coSh,~#GinUM cZB;yQ`6}?'@{x}}=bZ\|)0*IWFr]!9U,O%kW1q Yig BYFѹ%䠱d OwثDcؖ\9'Ÿ൰A{ I8Ɲ0/eblrhiOV2is+yd*k d.8fJ"0$hH䈂DR&R ŧf"?Ysބ @V8̉O2FV b[TRoZD^︰hCA䄉yA:hB R] ]`F[~J,tB|7=To!/ +aL`  jgf6w.H=cB2$ʈ@!3cDY%ZhO6f2\.%q {zFF+w($T]3VZS!5JYo-9P/cQxڃGSn!gv 04Kx7j0QE 7(oG{]l\O3[H( Gf=3CxIs*"al!ez${6۟f0޹sYi+NlT`j:`Uހ1;q\!CN+`CVm6czKZkPG`|cg4{\P( K?D7: /F]e[]#q˳iY_;,\-eaѲ Ŵâ|7aQ+-!%J5Eg\s|'PiDk'SZ(itaG.1; xA]RwbgXD3IJ8w%˚%xx8W#|U< .+wC+,z\WpsV,z;"/1PREKsƭ(glK %E պm$N .Cp4{(w"Ƀh{ ,Rflj4|mf2{dE)cyv>-,!كLKAiь.9L>dCE+'oJݒ]M0׎%kH蝆΄m0cR a*QqqL- 53Pnm33M* #~yu[dh1; K&t9ѧY҇xY1J]JoA"-{@O ֧wr3(XAe:`Z 9}Ϡ,7I{]I[A#>൓/R6rf>VY)eYڿW{G]$v*ء>j*LP[{~HbWX#;... |AՎhMgEꕚ[)} "T*T:1'+'DIgd. OAҳ>P1ö,(~rIlԵ*-{?-$QpQJ𳇟UR+s;clj0^.< jdJ{8ù&&b\Ɂ["tgUFi^/5F̼Fr\2r7Y3Q;ctB{bٝXH-9_Pj[[kHǍ%8 kFԶ3j>3 ]H{΍hmІKp!8!XGyiA 9{L_腲Z`Y\`C2)k؊r׭U‡\6`wmKEFWuW .`-Qv{CJ! 9$Ek8}TMU5\D^F0+B9SĈW.00c]`tj&*KA7_܃xyZ5W#Y@V P 79No:3"<:2@Z,\btbIERUzqQWؒ&-}UXZa7I__4ը3]Hji#. 9F c }Z&*J=8HUn 1`9;=bIb,MO9͘i9<#} '42!;FL_&o}a{uL:7.ӗq_ץ_ɔM0 EXݓ_X}l|QiCSX brA;8pړB%V+^`ȑBO{ZѴu:&<{NMɵ-a/cYDqʷu]AL2bӄyurݜ.t8rypy1jcepQ)̳3a|Ǒ`_cؾ}4{Ƭȉ5vg L0 }VحzNW.Yzx-Ņ?]sH2";dZ'orov勓SDf]B0tHs998}qRY`.yl&~^D|rXXte7eme|d. Q\vUAű9rFQ$f \.DNiO._iv퉯?QT}S}y,5#CF %BUso u S'w^(rvh+t0&\Gf}+F[= Q*2BW"e$`a4_~z|oפΗşulgnOo.$N09K e.'"~K\ڣ8o&?-ggv I@6%*؜bBJ:r"9hEpntݨO'/nP(z:*͍~<֟ḭF[@؆Z)jq v}rF?ցn0v=gh܋C`B~yș6DJz6։v4f$ i%>,JLKN*n%vE(Zsa4m#&z+="J9+*0fz hH81GgcyNU.AV֧dx'zsOos;xV1&K`0Pu3&0ic=j) *X#1F\OPԄ^B$r4mV*ZUB h )(#vQ6*qDZco6K32}FI L8>,f:Xlb_hu} +`5B6ŵr䧑rFN̤Sd !x̾@r}<:A43'*I[z$ +vUӻ+IF/{XؗL<XL:U1E3T%5(G&y*SY%}1P&pK9&&)!ed,r5ɠP @[lɁ3 vhv!|2yι_!+bKJ+nyt {ɉ\#f2(x Vq` xb;,n.ßۛL*{L+-}LZ EUYikBh2Q3DjX¹;TŸW58mƗ8:ԭ; 9ak S{wp1@D3YbG)+;+f>8l0 btdSXhMB@7E0AqJ ,钵A;2 Φ.,F ٵb~nX̂ToƯjQR?WepmɊ#|'li4W!,O@ȃ2 srʅUVIӐdi<`~EZ E#mg\4j36GZ`Ѐ)%Htpl3d2G-D24(5`sr#H;[֓Vu+螵] B|mF` 'b\k.W|_\\/>p@k!ŰZp@0Y;M}aFXq r&u߯n~q[M{I;U*\#[yJxzIwR9(ջ~Lڌ`f2}}st[LwUdQMm4iO*)e9ʺ+Ar BFg̬P ߑV4AOx]6]'vnZ~:lni7>*&ъrITJXYddrP-y5PΙqqqWި;㮂R5q3ޒa2߉&(RI.W_iX|aWj&Ciɓ.Hd*q&r K#Gu.جl)Ay0Aʭ >SrAׇ~6s*yF3ejLq@:sďc4a'2 3ImwԘ덱}ow&c{|@SLd-ӘgK;qتCb"ﮡx3~:lޓ#/Zx둷V7pdRނq7n2|2,~2V{7ӿ;wD\crG\NQ|NL)+o whE3}9딑g^Z˛̓_ghK\\3,F[]WM-wY'+Z98eoݸ 앣KjjMq4M{]1`4 3kX tgRK}3 0sj P 4aTW]:eGW?A;Ï&l?pf/OTXYDY7'uR7w/0 0\O^}lM&%Ryp,ْ~l4j>.wgkGD=YQukn7^ZL`hj+d Hy8(@Z;6;YǙU:3wfF"[br+(ǔ}{ cHiݑ^Rĸ:9ߥ!:B7Gr[ {Cebvx7Z:#3[1f! a72$SܥΈk{y0{^C&Ԝe<\&c;HY[a /:H A1n\9ꐗkz_qK,Krʺ:g3 [y.}wj%I.^UET ǬB܏:B\+OKz@[ҩFKJA-j0 =j(>ppxSQH+*q㞝 'G8C֙ s@+ODxRJ2%k/^O=CHl4%xeKp䯖Aٝa\Yk8PB1*ء |Vge*4|1Sa1Sr k4qlJ ViD2%Zzf''/YZ*6z)F#gPw ZȚ4A`uSZjQɈ1Ѭ.3&C%@L'}c);ƻ6&^s i=sklEH9;D;JjuV$>?\[S;jޑu->HIX^;CtV=m$m)U۵TbeXQBRhQ,eMRw#mJ}@֬8^`&Dp:wo3B_/'RRJ;hҋ {:&-O@/nY\у]³7xq6Pףe㫕N20fl݉*7.|ybF& J/LOaw?x9}U|]1g{O/FbTWǓ@O<W;rn).'Ĩ:~xhĚ~A xx5oٜ5A) '>r//N_O#3?;0>j(BIOCx v˛ W3).s%&z(K̤%U^h|d3~VXmT!QwcA34X1 x i20>sv!ֽZk `-aaw7t! +aÏ' ƷCarښg%ĥs'z~fz>ɭ :op$op[hJQ7 /LY>GFp=4Y|g=v^HA 9 W"Y˗q#mp}-G opv\CO{q)\o(pB(MeNR@oW1”*MXr,dT/QXR>hDz*:rO8W-{H0BoVɢp|KQ'V ?43vݼ֢Vdc[@.Y? [צuelYǝg5muC[bꆶ嶊mu)E Z݊cܱLx^hd(Cpf_oq*J uL#mܱ;"wqd5Xmu6(> oS,q(a4F$ƧX(QJH-xcl@^,dFQHsͿ〝JDž$I~3Hu&H$$ J;}Ŷk'!lY"7V,d#4kvbDr)DK 0W%J`UK6Xr,a: ldUQ?nYfH&*K^%Oˆeںԉ%hV%#W%s)-(W"' 7nd$w@lkd$$Y)x . nxR$x(+B'd c9bAI4idld.9m~3+DgDV*+w6-hW 6*i0!|kYP:7b2EeZ(eC{B"+Ɗ ɿ}wy+"rF^c?mk\M]s@ブ;g.\8C P'ɍDj!˜زDpy(>k+l2]# Y^FU^D\~3UY٬].KLK`Y¼̒ZH#!F!9/T9KY,;HӼk=y1Lzw=WT\J11kp,Y?xZTݰe^dŪ A*oݰk!Jp;.'\O_[>m`5C5cFjͰխƹ$V|Xz !f3i?ivk/6l ZkkrWbƁwzgvʌ!6"[!ZS6NԽs0{v{Fh'{nE:!tklٜuO ˊX<ޥWƻ7S൚./bqx0{"J蘈>rc Pm՗>y1"?6bqٵ;A#Іw 7 y v#Y(v2&xrJtXU.AI-)Ů,g*&9+>~$UؙȚXRBBOxPQx;#~}g4; 8d4!ݔOAuWqɳ29Pp>XDb4Ѣ,W(Q[3)fbe*Q #.+:UP& _.9M(4ЉnBz~s߅ݓ7!Sx ́~= S9)2a bmb"r!+!pEjIk{ |A[1v8|| 03x'ڵ؂tiY0Qʰ֤NBׄ`\)c;NӺ~ Xvrٙb3PJ4 xӸƛ54fkl'zTQE.WeT_-$ml)9dᙀ 5!(K6_D_4)c,}g)We(YZ=+(bҎVJĘ  zI\-A) ML-n,ڹl{B0lh|UMqNӦ&|=*mƫϫvSiha1zvX# 6XXvKgf391=§XۥgNpH\g}kkFOw3YHM-g^'AD虿v蝎J4Bf=-4qA+v(#23d)kXh~ƁwӆnKY԰X媏 W޻jk V:egh-ؘ $6QWYl|D8< %H?cv֎eOvLwvOLFsnCgQ"(~i0<Ծ]=zϩa7ydH)C1B-QR" jYe+蕍9.2^Կp3lE:*6eS㊉ȥ FedxفAD;mjK{j'NsIGT *,(,B)v% vSbE)/d!|ZPETpr0b79rgn;NvOd4vV_0'Q+q?E5chPİrF^-/ñ ZƂ$Ѧ:URP>ƚ7ێRZ+ f#π)o߾ۥ]=%E EByGϥGnY>: 4f3ʍBݔ ܫCz+ȯ&CKA/~x\g/S߮DݴȪx`U7>߼r/|+_,u ?_~"_7ËOlͻn nPKќ:U|{= j_ͳfzuPv'"Mi~챥JwO@Z窂{V5צjM{o?ʧײw+r@'6ٵ)>~~w#.vQ&#Cf B'-+[ebǫ{׵֤ :( $IǜSxl֞.-\(4ٗ"_{7[^^|q@ۋ_[REӆ$fʚkuiZP_mEv$Ë>.}!_s5.DKB?փ䓕&%ZU]kRh졫ތQcbYW%FvW- `][+VvTukgy+iHe:+ߤ wWBӀxu*&ŵ͝?~ +lF&"s3-2pfF92 {/Zjٷ]0 \7{Vp&pOܵcwߺ>kuO7~ZbBkw{zULƬB/#fxG?z\[ ) ;@/f#;gmw\2_E5q"y\Α <5at;e !A j)z w~ۛnz8< *v}nU;ٍˢdEIA4t[ sigbUr'+Ry#O\ SmCQ3.|c6b΅[$X0|A imǽq)toLo>o΄Q),ɵh\*!p`*KjD"VVyAqw#f X T썩*˔IBBh|-Xe&TwU2uMEi}`L5A摐>?c* P[`Ͷ`DI+r#ԸE΂k 7- rh5.FC;NTbí["ǏE3G/c%? /n'okxl/<'7AvĴxU'I1vޞOoeC= 9fR? qɐbŗy 5Jiw]/9jA-y6ܼ}zUL#Vk^ƴ*(iɉMиԥq QaqyR@2] Gly.71GBb*]Ob(6gqQXUȉAP猃QO ) xڭ~i?{sA.P}F08\ j`T, 񎳃2FIy/ W>85v ݯ.dY6IdCWgP۔C| #̑TjIgM&@l{ߧEy;-薸_,rW\qCt 9B>߼br6^߫ϫv#M^? W]eXٌi]s7WXp{~W&Ud?ܥTc"Xj]䚪ZiCi9RpDzfkf%qyBuH=a(N&íK#ӢYXz ?.i΀Iޘcb/.ˬ$ۖm-M߶ *q&*a i`43V $#7w2xB5۟kDˊ0]6HO3Iͽ-vdKm{dF32S_hܨYT1P ~0rBנ>CnsMB<g5 wd/~bg&vQ@ц= 73k9Ƿ[۸o_|T1b`h~> P,fk~n}@lyd5~Ps3SUUԗw(Rv @K4Vn GO!3^\މWi zP觋hYi`Rggz %"-jtG\b(:EzۀGqޞWJ[~u`#֞Jr{nn֨bߠv{Wk>Uռ%pcZ!lo Uy> 9p$_ko 85[^#l9]~#ɬ<v[7_-b֫ZuY-{PHca7`ͷܤG}Zzc 8*58W)|Q(&i]J[#"p0jNVcj]{V+0 m I؄G&)Be:w|(„ӕÔR7kx1^w>i--d ;ГHK5%2,v+qa 9,5 /=։ s^DuJy3M}q&ԓWqrsJHrFd @pcH4sII?}*B9ۂ(9= eCmjFFbYPt* _b8JA(E+k6r+S1}VX5 A*o!M>MCb0]O`&wI2N._yPg\=muvP %ꂂ#5H Y$·fu(Ϛ*KA^ (z֓)9ZDnb3OX[Q(ZCUώ48|l2\YQ3u7Y:`;5[x!GO֦81̻r jPӲ(Ĵnڬ`t8%s֑B0ݱJXrڧ}lZVXۺ(:&p7.dYH2a$S!T;LRﺰeח,@ ZߡAUcurpM]}xl0Rڃ鹿Fp*܋Ph5]vlr):q謸cձ)>X/V8TgbkޗԚNUjGFt h7"qB,qag3[*)x\zB)(wXmErOԁ``ԂEըwҕ@H0ٜ GG9zkT4o9X>%[cGޥVYmWMSS52z;LshV m?XdxʆFIy. j/6) *LEMB?nJSܒ%DCMKL@Ďix"DF #T(wݘV"VԴP\)c8Dlia EF$ǵa+Q&R#/]Tjc|kFKf0|4 {3`|oϟ~=ۑ& JoAp^{] l4I$P-YmCIoL9k2L%7HHQQ#Լ/ Qn)*#>h8Bdq? m2P$.gC?7Z0).לOW߿럇+Xmm\ޜ KY n9'M(;<γ/k %/Jcc(WEָLq&i$-Tq{)YkXIzLj1TyP0`  e 8XjY @o}zZSa\:xc:t}5͠P4/6l/RBn0Rx-y㳧 ^['?3JYooaF ,g0)^skoG5~ac["#BJt_QQ,%xg \pݧl]xII*y&!ũzpm ..G"$0 f?h5 7ݧ:k#E;bSTg#A@ vqEamۃu ރ9xhOBXJp<YCJQn&B߬q8dxyj3DR0Q+:/BM\ )1P4`|9}W9y0mV؄s;|>y7I<c762cKp?O3L; Nൡùv%F C5eWDˁ;4RG#! ~47HHe00,eZ܈ԅapaSP |b 4f~ l|oY}[(:.Ť26Ͱ ǜ\)xA@!FA )LR#:Pi}#e Ɏpj6RZVoOM'bCeȃnHge( kyPb"q_ˠvԀ^kF'G :&4BWHt$J0b{v8 y!$Ab=" {i3"X \8Z9Q:^̃.E”Ɉ|4[Yy2lCѻt@|p?u4W%&4X*q2CDXIǴ )lf6|L⾃+jUn{gK)Mknnj Ǫ#/{AaH QUpcňo]"[Ԣh%}SVY 8~BH:B!VpAC%XV|T"jY]! ցrk HriaQv5b4G"lIrByY(YlRy4m9fbAs{\( z>b1//Ca\pw.  c4|p_OR&v:#rLs{xw$'@*UuATkGL'IOrIC¯ ̧̟Pui`ΟW x02'khe%/4T/~pPFFԴkhbG}gϗn9'Fc7?0okSwwYh3Zo=uߺ+. U;d0A|{ Bo$64ț,@&PgT8 >'Rr XYC ˔GU+TùN2Al`nrσl{C=_S%kϯ'wSNjj>/Pa``lx9,3k&B\UDMۼKBC|׵dvgnϗEa~>Q>&W!ʞt2 *[!}L!;LPK9_}`_P21>OLme=?0o Ĵ "½%HHFI)A4TbK7J;SB\XO 6CI qQ$G%ħ` 6nE Si-vBH/ONxNx%$99pQ O tG3_Ivq2u!CqݰS&ef2կ ^k]Z9Z{ dt2  {?FQ6 ,9D@hmHUs7^,鲷ft7a5\NhUt<>[~Yaa}`c#2HI;a 0datwF PޓdWF`K]AvLlg G[n=dgWj"RghnHW.vXdÌ/0|:*]Ze[ɫ~onWjIYfOJ+M⭧_3JQ4Awj_m0{;:9\Ҙ"-޹U~xQΠXRfӶ? zFzGds7oz3s--zKTSZQSq1HƯl1vt%oh}|5ĩL@KQmt]-4 峊0.}I ¢H5^L}ˊ4ht-\Fr&7VDB kMTAX)s?.D݉k ?es7W* iINI܎pv/ʡyu8VwÕAC,PBR@CW ]C4a]\Xםy LbX[1:ݎr] k;ʎLujjz@6|r֑UhɒLʒ=hܞ7We>Ǐ7o?fKj?Nf1"M5~ٯ_x:{~?cSSyq?7 ,n +}[vn~МfApܻ&[ŽxE+|;j亣VYjq0Q4([nu,z-iMcW.+?"kJk.>d.(%Oy}Rx'b4]<.4U V dv<*|"d7v"!]b#&@hg$pį[p9{yj:T[pukj"(S:K#|O4!Յڷ*yJk䅒I4fJYodKkӑxj* APT"qW43 ;QpN , 1FR3NDJW|D[(]bRZQ cK€ \iBXtI>A_7nAr^24qjeFBJR0!.0F0#$ʝjP*JHqGTp%Oih@ϻנ7KKγ JJq47`/&6[׮h幉%.\3WWuFȞM:T#F{ZԗD[gݳ{rD\ui1HtըP9rx"1?M4~CsUHl ^ϐwL+jµ.ȯynܠU'S>ǐK:[IbŇ>J~#tΣ+c^G RRdĈKflF ^LS\1g2{@Rhsl؟z"l L}e^mj^2ɯ\IUBGWQ(\ )CA+iʅ4γ8׳FsU>wGςܔ0Cya*->o֣~r燧HE6[+ g  zl:M9qS*aM,\Eͩ9N(DsEF!y}{z> `Wcf}!ԱMĠOG;y4Kn.L(PRoI:ك G* @ޖl|>on`!ڵXcf=KS|jlqGvGRvEhs#|`E21( \)1L{eabhdh1BhxVفj˅hΚFx m>mGE#'%Y%K8Jcj͕ * 7ETs?F6O_N6ku2(($#ԓe^?`XI CR756:[5W mL}ov*@%wdEӸ߄TQGeot:zmI551xs+G2t^R:R+Nl?!6IM]LZz#uBb?:fb_ a!e~4V6YRÀUls`\D,>+@2ٺ-tRR !{vLFfW~V\!'5kgD٣?fuۆPLqgڬO?ʞj~b4>pf ]y8?N𦱝ItI·Pٞ%мrD'ދW&(34lzpٙ1P!\poo)G\R|;<p^X;kT)яd_G0ׯ%ѱx֢%Ah;2S:nQnր)F:^RRV9GRͩ5[\ x_UAxx&[ pއšV@y!/h'RKJhiD(!FO܎ؕ2QcT^pˁ#W, )Fjc=|R [K"9bf59nxDpΘ=bsscԀ8/2+5BAqHk.Rwx._jڨ$nu+h4%H72ƿs 8"GëT>rǫ0NS`os!f5B,+7=ٷ;f t}4'7Ӄ+lt^~8iج} Ѷ7ϻܼFD(2y-w#T$< }VX@|\sxT0I@3КqC '5SЪ܆ЉO_k0 l5 3*E{m 0mK1*ʅ`kp+Hx*  +d)mK}FNj2 Ua *NEF6T3%dhV l-v|e)CO-iqUABh,8`12t2ѻG`D0.Njki^a F1<9`54)JBhb) hb0Y*j  :C0?4׳ d)cqx*8݄s'V.dwv=NحȖƾz0O%8W'hJmq8ԉ/h"GJL&$BNCRsu]6M~F^BN(bPB0 G1$x1n7 .+xpL+uӛhu4e~7/ʣ'(09'+@F"偖RO3^(j~SVSKvʉ k ` 6LP؉Uju%ݺ<-FXH)}u8Ww.?:ksȅu'`1g?iv̓ǽ6[I).9&9/pQžP^le$ |` 1R4>|:9U{JЩ=fh:E21qrjle:V:'ϗ!}x;Yt%5UZܙ*EH1hr7;`g3S˩u`B&U7R=!:84ʃr|TS\T򉇢9&TVnm[EV ;TbZd<J㧩6X2%},](0 C$P'JQP1.IwǼX{^:Uըƀu=5@)@!m{y2jPaԀWZ|xJwG-pb * 4' { PWS&( KOk`@@x?Vě/>->|6v_ʵrme5Kh /K#.崈5*9 :%z{0^$Dn]J,k^[nJ᝜`SPR 0t[QkAyv~:0/N?JJGc;@ŰKG$8&Tve Z譵}lYL_~yvA|y!j"z9`Ř`mm]{ȉɕT^J,4X}گ-r ],ŪpoS33_!c0N*xUb˝] kRp?,#63$OzY<s\v*3>._[3Zx2Ϊ{s;$z0v|or@^c~h0M޿ގSr_m%S`lf\kg-͕խzN[%'`\E? G\L-usT6>Śbw8B9wov?`v3yQ m-,k&,4r>lɒt9-lm<}gޫ,0^^__twZ8>=V\ɠ gƓmyY>]<)=[aN>T![7uFz UB"h`Rdad3 Hs< .s#ڙ }ҙ Y#E;bumܝ@E)>'ct_dލqk'`hF^Hô7MgY=0\O2,NHkFpn'V_fkA,6|y:ԶXv foǦB5Rf )by3O8IX}gD& c$: RE\]MDo{ Ir\! ws߀#J}L*vbEq%a0L@5ڡó>Be>(%H!avH1 U|{#n+Q %8I %<"JHPv,õvY!y |yY)u${iE+tm /nL+wBN.D3%z`z}B&?iŁeD#.@ǧ)W81 <2̻2"(DdOP)(gd$L {N/.X~uC".[# 鴍 Fv`4C[q^=d\Ze~k dngEE5w& `&ƆzȜ؜䮋Q>cohYp薦ˆ R5H_9/e*qbP#B$QQ fVޘ)$\Ǘx/^G,#('H) Q l Q@4`8 O1;IT}PL*>`p@gDM&dH38a1bG*1a(uc3L2 eʙddj9XH-7%XYMm}+Hb$ IaXP ce LKPaq" aj-bipYκ ˠ !*$t0h=h<[K/ǣD7FErj2Aa> F_Ni>ls~yp 59()d%u*"Ӿׄw8V3֧X,V8&8+>C$ycE Dt *iW>'K4XZF (`%<@Đ2!J0EGDnɢȁVIRƦU_"T2Qp]ʴ!~Ū2ƕ$ |2&>ɤOD$ӄ}՞UB0JQSq$RjBgŔh+ !a*uD,C3Vx59  @ڳې-3ފYgR@`hEjwފqY{2p++Z+g6+*%! fnŬ$c_*Di2ͦRz\h~;V6({ݾSY_7<饍 "͏ Dݑ/%j1Áܳ2`-5i'܌&۹[Ϛ4aݭDqGgaoDX r;YegcD؉qr9Vt-7"ΐ'zNw+wo\VECUV7WVBpR}McY(caw;L .fus6>$"ܸadUu|{ &fZNvI)Y=nYRdKJao&`TYX!6Gnw4 UrqΙC<[N3oa-ޗ`%ݹaR)$uk`w;RcN6Z'=f.qRd"3E O,!B3=D +WГE-D`c( 恑W:v4:i/&VMnJѺ"\ߤp'bF",,$`2Kx GU|Ԧ$}vxS^{d  gC|6U]=%)tk154rH(hƑ/aou[mAr6lu,&/E>QJaSy M19 " CtDE;~SC솬{ b_|U}R*u̽f4ad6LX"w,LˣȴH>tȾ1G8"P$4K˳[!y=Y!Nˠlw \D+g@LԪ;&gc@V xYE\ Bvu?4# êm;ANxQXFcFwEb,#t*FP0"i)<߈ H!fJ$$`DK's't }~"zy4/Y2@m**O Ը^2'f,_(t|,hb^3Ey0Ѵ{jUaޞ(07]͢?I$rB|V8 ڐ=}!i:qOf\_TKwijgN#Вcݱ́G9dl:)M- #j$BR "ǜ0!1CBb:uҦךgzɑ_2Ma*b=[azW`L[]d':JLe6`%J2 ASe#y71Y O e!@t>p% r`b\@M QE\eT G2᾵;u0peȅTi~I^y֐}=˟~C'Q\_~~Vh07Cp`fftyA$']^6Lg C~|ESέΖw8~g`ww> pZ`0vot=%/5*)x⟹dg|)J lF{X*Za[}Tb4S=Kv:8`出R@%)FA*X{rٻ- |4rh;_a ܾplgn p\LI˧0we{2BZFWkk} g_w[]Oj%.͜SK}T+ө1~I9~BixP0ASW;J} y4P՟_cХ~ \AP%JA d;/2-3+]3Aؤ@+BRTqB2͌ eEV)yѻLnYjGC!=v=Ygov > g7tH}" SԻҽcmbғ"QʵGz E 96ayS AvzR)BVbkUYV4)+m>+:Jƍ)Fvhi.A޽C[hPbx#N-+Rܡ6n?k$xQ;G"\DGn g<DjѰ熥ޛE%}˚P-cxI])fMe℗ F}*:(ǂ9탭L:T=V?L'w3Ps{;(H=5kz`|,gjQu=od|q~^ٻ1' $"TnU\,Gf7bURR 5䇷Eď ""DB>fkGmnsh,@i|\ d|=dpA ;=G}Bpn45cg>kl\JލLpPBD!] E4*uVHR<Ϊ4dUd$"YfOf,#gh|| ;[}1HCP3%?>PRc֢ؗb2SG8SKlI} gWJ ((wQVM v>9^kf \2:m_R˱Zji=>q\SLp:|=_2iU&E0y}eA}6^Tk~ A1%tPҀ_25lŧ yLPE+u%,s -07:BPL7;N%h/IOl07KEGaU|Ip^HlhͬLznd{x1L_f7)Kx-.?&4idujmB)nhܣ H3&f8`T aw P*:,|q8x: p c(>5=}Q)2s0_BV|~:?}Mpݮ"n|&֚zuJDIJlqmW,{,8]h}<( 8nj.h> Ta.Bn/zBL]U*}]THpb%85CtI #nX0ָyGOU׏_;\ZNݚ[UP.+i`QTHIDS"ӸGI+1pwj׏ lLqN)r4J)C}jS@L5#)a}15> n:/@1DL>mH_.?¿x5gq4R5)`"1TxV,۾g)by}Ǖ&)3"!FRa%^+A$2lF2&1: rخ!1FwH*b9bX1' K4Q$1 d>RL(i$ISGu).q{2NkpI=bRɉԩ֒k/jէ(rV“!`8aM(MR`J';:mwGLR8iqq2pfs,5!b"I RY&ݘ;^Aj`|C$$S,qHjF`RZFcTB# 񄃶ʐr]mƹ^`ҊbRqZ}BH]ٟrt>QEZy4=_ٹlMT=&yIj8 jept$ D;! #ij3Ĵt^fa" bV4.$X Sj(orXUXO8UAS9I&C].MC"m< fItDNEւƸ-NJ8#,Ǫ3c F U owB{Ѭ9jI*u m7Sn]zPTu3zPF vcT΂]ؼ)S~p6Ј,F򺤂Rт_*5S|#q쪈SMTծi!DYa gGJLƣbO݁@$ wgI,O2Dž۹\KL0r/}+gdͩ)nAiw+<yV :"S#JUe`R>*y]ӫK^kT~_`-q;?;υsBx0D\O70z siOϥ8q`*lQIqIԚDg$I9eʧzWLn H$@5:V}c̣IJJ$,EqX.:[IT }AC"0\H|"fx,]J3y,KSe<2SP3͐U[PL7QL63k6J#ZYε,3QN-,20Q_DIseX~_φya+SW8O`#dALps)E0ySpdR U$G~Πe)ŖyHr/$'i"yͩ3TYᬡYJWXIR$,LIє") Tm5:zeWXX4T*H,,j<ƈ7T8ǣhM8?nٯOi#ޮE{_V:v˧^Ub~\_68+ȻZ9Fx~ĞlY?>W_8z]?ܚl~gE0cAsN{F}.8\KЧ>?sg1c@INhRZ۝@b4=7(!L$)]kI或%we>^ @GRB돌'b{G3[XQ`kHN=Ns<%R^ cpG&FajL<GwȚn8 /w# eatI[znuMs?|prcsy[mGͿUb-C1Zd|1sR2z1.9yU6>6G^6Rha Xp#c'߽gǸɞ>,yFqͫp/d22q}lVŊZ|v$Ý=,%u۬*$:K~翗p颼=}אI;Gy;seYظj2_el2=vgc~Kٜ.ɫ܁Y$s[nڈ;F5r5|z7QB6Kbͱ|-fj}`0ɿٻ޶W}零wvj HP'7*ebE{PtDҒFI آ3;/wǿ]xb\&K!g3vFرHDV';fK;\.ذ1 &]b(j!O[UtˍH 4?$wc[>]@`ntrN~Snװ,- 薃wSr7?faM.qy^E W.%25W x_x ut&wvI#d5]6t3b UY%TE7Auк]4Zɞ,mº~kKJZ ~6BaDz:%E7aCt s[,[h~*L uhmRV*K)ZV9ZUvkVDK =sFVRA3w 0fMq\?w~!z͕CtbDFqZ!mQĮQ5wzavE#b%x5wf4YODw/(7ѽQ{t|vk,Fjti Rnd㢾|۝/ u3ͪ5ZY*g~s|e\O e\~ <2Y}j4W(9k"sn`EJٳ߆ ? Jb-G)/ C%n#`Cm`YXa`:PPBVA꾩%1]U輺^= 7G7O3wJʐ^uIŮ:eW2hN1sw){m_vL)s·gN=:emnz%4KH:e΃مN[;uʄsM:e ]R,u-kcN#ZdD2A,S>[_m}:eC,tP7]>[tʜ6(6|ă:4)b@-{.X|ؐl#k-rj$axw@ )O2 VZaKf_n&ȫfdۮ7Pt }72*җr#,$@3vߏ߿;IWv!{7G/ KzӗD_4ͤ>zC2L 砙RL\2QS0a\X:ILA(.&$9,֤"P2fLV(%Lԉ$2i"%YPqv][oV骿YtOHѡ;DN'S\Dk6*$z#/ֲ lfB&{"N7 K Qe0ZXyn@](-`p!~CQy,}yqr<|9O]wZ _mx6dѣxyWnK,wٯ$&Fo8-X\ɇDp3 KFp/QLgƋ?#^.ԿsfM'?/i|5;s[%ι2_{_±OM蓛?>w5E7+ !k`raF4AoUdIQ!7 `gGxK3%6悯ko–vOn/ϷA&>d&39 &S-9VA1ȐPg͓rLaY3F*6Xqdb!q=(&AE&0 ^:=VBROS<@Jg+/{`*D0f T #3d,L>ytEUgI x8Odf=!eRU\oE㬥JfOz*b袬v9Ar R tdSdhrbJ4>K8֘j0CXIXQ^cT]QvI=;֎= RT^1m*/Q;o1o!L|%U/wC7E8RG\5 {։uhhLX'F[e)ing),IPuՀ`FZw8~h2ӏF}ռc~G/M<~*C>L6!A Gv:ozqeª6 b_t.v~ b+:~:-bceD f&`$ Dz.,%Yۅ"fsw"txa+jE;aճ yLӘo$ Hչ1_3c?5 ՚zybc;b akx<Ğ.;x[bcoAOڶӶ8(zzKa"qҔ`#0j]Ln,|17N[UR&{&7&f%oxXڏ";SP:a=<϶Ь x p9>%SamY-6$֞$$'<SO8}k VM{a"x}kyt! j1keZS+A=PcIN+Ŷz߷_UɗlلDZ4,mŘ~91'U>ǜ1Qr@ ͙$É*b\ҥXP,uK UԼ[@rv:%ʓx(qE/GmGUrZoL ixh>^i2?/YIY v Ԓ]ILUhлpj(kqa'a*Hko$6/D.&]$r$!F!@HWQ4=X̕|_k‚zUcVM<䷚B7e33Hѧ>&U}# fX4R۪F*4UvĈ3nxw-VJ=$FHأ#ڠ|"O>QP,c{.%yVुQ6`cdgk}Hsڲ>HClͼoj>0 (@f k@f 7XswWGK7c$nƱ 5 |ޜtĹZ|_W$"hd$5(ŀ/~{^"+h;?X҄ ϵ~@[ΕU  o|+s3m4+ڟ=Y )U_e52V~su -2& j ១@; [$0]׊-DzZiN+f'XH}~V;-_cCUR(JP_C&c<7Uf=('𠆷@Eڽ %$=JQ=+ א =37ɎuZ?n"؃5.D\>Q9]N8tDtVAM6&+iNSuT Z@':/zmkkb`*)*4ɱR1JB$H=}X\EjӛtHV&ʒSS=Wř hH$!˜D"tBb.(Hm.݊>5)&bT!^#)s/l@b+k`2I䵬+;gZ#]GU1hnKQl$EigYxDΠHE j6%cƊ,֌xN2wN(pQ [o\d B*z&Yf֭&:ePpYN)hZ TV1*aC+1+h@æU>D>/LF!V9AhALb7$a4Bfd4,yH\#(!{yz2j #QyM)K^F6,&2+|( 9܈(?O2NXH| i39% 0p43iXRVȲgE#6Ja)b9UG z),La 'f&1!][s#7+*nZ杠33:f*d_գYRtdv+-n-[-ɎNjlI-6K;`/7w9E|[L.rҐX|ʋEw2ow7MVJP};':7X~*"eW SE~_ >׷lp8]}'6hwNݻ:+7F(yՒ9\d(fsZlΎWP*6m7Rpk`OaI,{?ij`FД '+=hx%Q`nhI07;TKt~ra<|@\Mw_9p]4s}mU?5l>|P*$!iUBP@Gc/CywxnљBh,5(?r$%ʥ/ F8E:'#WD4,WQ񻒟"OF=I=y#^go13yN';.7^'|rvzJ6pT?{s??Qn 1@*1-fYZN ;Es/%Y6ޗKɋ@ 5XSL\SH\QIBrBZKW:1ږW~g!rbřc /rОA]պ1]惫xQ=CXyGo8xoqf\9Bкɗ˨|q*r\(6oPZD,ϔD1p#)V%w5 ^ZpM}4Y J<%7s QBϠ%n)v3Ey&Zb*,!: r˱+ByAI3^PSjn;WP)V'6A Pe1ˬq_[7DaͪX+ožU˖>5p8,{cf8W~"Fٞxg 0;'o珗l)[xƐSxMD Eܗ6@([t\h ݻ%hZr0Z~e'_Mm~>ٍaj~+>:x2I+8ɧʍǎÍb-f$1?> CO7rO6<4WoeoIygW =XDnekGgi=O0Sy>|hvnegͣGe 1}3Rh~puU,m<`,{S3VnzLߢ#,t{(Cww;P=i# !3 G : JSd 4W:ln c Hk@|,a3tKSb T )=ީRj+h6:_BT^EHeuR%dmɜmc>ngޝjܰgOM Rao>t|[bri\]nd:ˋ+yYpy|Njaeju)Jsg ,п]tN-ǒX4QNW}:jNaہ 'zX@uSҺ<% *mȜ^`bԯ^ۀȯ r3*rHl)X3ciͳ2aX: CXZBScCJ[ qX[- K,l˜XZAq0Šh7훯zaQB<*!~5'+򡳳8w'7]|1c0})&[ m}=+rL`_gz|amP, *[dQuZ%*agϐΚE0C/sU;- WwyzofJs]TJo!*P h0t'$.;͊RT aC9+xȗ&-rMoJ-,.Vr'{P8RVһm(L~mB+. 3]u^ nfgWȑ6j>mzxJ1N2Xj؉d)ecN2>նԀCD-8xk@'%WHwJqz8: C(!7߉/;0 pbGP=ooiӏ$ Y[z>@t<į|5_淩4 ׿FTp Ac* hgSxB. \K^Z.Oa:j͈fOAeTs:!XWC bKܜ5gg (pB-m[VuraraO7״.aVT2R1rZAGQemer30?:?@ "_gTmz,ZzP0kWj)%f6D TIlsMC5Pڐ̘6kɾ׃pKqܼzSobmC[_};qWp^)P&Z\/o^"볂}Y7̰˿rLCφo_.1!?)p\d?ZpYMt@G+2q(H$,Ҽ 3燒g䏳޻|<Ṟڵ4gi(TFN t|wZq5gpmQ$QKNQAB Ehxc}QiR"BxB ʫ@z`+$B%hDNH)E68/gU)H縍rü DrNE*RRQ'еEhivU*6y0VEAt1]JA|xEZ8ipI1P\Q- JYȳ5j(*VZ*NX+Ng-<^$%9o?7bі>^4#cGi8,Vy{KZ ,2B+:}Fr~ ޼Ε;hCsǯiZP}Btu>VZ݊'Ԁ09:ǐ+)qP`iT&ҮA{XrA{ulp7AkZ &SXCz .*[Th6K1!bc :5j+EXHA}䶊! MX.{xW4CAJ i*RX!RdjD#,wEzJ4pƆ!#1Pg+B4LKL<$D mJNDf;O넥-To]* D%%ZBjquK8ۣL:s#Cқl"ڥPZ.v?Xԡ+F+m8Зܖ}(l7@q~7fY˔@R"CR!y(rFE,󮳼S2ٲA ^!ǔׁa#+kW@k-oD <G Hf%:F5)`S$4`GSKDzG*--(aV9svVRyxwݗvۇ3rI9끿_d9<98oK|" ++rg|> D4;ƥ2j~<#!i?#_^QLgKZ~+! )埁suJ1g3}S4GυXk tr1%?ߏTVYsLDL57Sq_8?c6ؖ@)R6oS˄Rq6KCGX*ݸA^B 2-~2~i^*sC9X4_aemj c"GPL "=,bj0i^0!` &$c#yIb(QB}$CN-LM%e-2A ٺ7R闑K4P +M!Q+XB(4[Aqd؈XØ NpCm7eK熋o&RI/G/$A n]}չWs",=KheuJ&́`y8O} ؖ)b uxmsx5}ۇPf8JT97kwsj/a{O5\hdž!E7&eэs}tnkKj,DL Ab0R=#`S j0#FzaH`)B|!>nIhRA"&rA`S^$OHi@}2ׯg~=ގz.`(!^7(J!(xj<"&k$CNZLJdoAڀ/3 A"#`ˢ(RH>h!C  /iyS' FzQ7~2cMFn~5$HF+'2 hI lL#CXFN5b/ 6CNiTbςSyT :&9.z#3DEc(:Ÿw0m1 2 Cn 61{;T' vX$BZ$9Ռ|yOKT԰3"r!Ji&NDdA +OЩDLukpg&ӄ5pJ>0ynd!GON?k+- GVӃ̥xqpZ>˘bg3/=/^kٜ25U  $v\F DR`CxtJsL3$Q䂘"{iUڵ9I{Gā [ CJ=NxQ%xQOt&*)*9ցu6! +Uu鶮xMtACz6bsrg*b_gZ4` eE[O$]A{IAvXwA~Zf4Pg dЕrlG~ulDǎ_MÏ#/!WKOiB JdC\[V> #ؓ!:jKW`iD)FD@#R|v? *YdtKMxwb\%=&!'ug{l>xn:WUෆg6Y^:=gQBLa'+br(<"{Bu/Z"Cu(!Gn4Xјv!ؤC^y!DSëpòƣ!߯mq/elx?ͦ5Cv!;_R>XA`&on?_Qw?dۻۊjs>n<^|O7|/~eg,fn<”hL'0GkmY'4n.G C5[X&H o kXUsn2%WsZ&ty&aV -om\9~~\ݩF뽆`~06uifGDǧɇMNn{MdM>j5Ț6+) Ǎx>1i-JY/7[=K>z yaS6{&Ym b}`JZ7kj&~jsHciF}q)̄R QiGqoj hCЄ`^wlAn+y{bXã䁡x84=DǶ(Y-rh%R%'5eXlJGerHpUݗWG CWs.V'Jl:J^nY^4Ȉ-cͩ'gs{7N%%Wߌ̖>#IIeգG o;DuB֮J%m6j=,S# V:]Sut-W"WEXJcѾ>ra/*;pNs3Z[ܪYߵ[FvJSDWRgD~iT#$+_V2.N#|{7Ϻcq'FS4O^ǭܭ~:u?fNCVN*len֙[J`.4މýs%!VʾmIJAtåf=rm}[}׷6Nc&ŠJݘ*#Frw[ MwcI{ DYH(P/h,>Fy}w:qaønLi?xz4X[zuzDx~,mm%+CbZ7)I[)(+I{[>rW=ӃK}X@a MڅYBgoȅ8Mh"^ A|hY0 HgFl`XJY핣jꃣQMQuToQXv֙(7Tj "ꁢ>BAe=\vl@ZzQDM^HAf[4 zZ}900”tqk;uĸ쿃E Kv=, K HjϐQu+_Ng0 z%0ÍѥwyvsN6H;`MblL]4CH}{{77gą3F[w !'}q7`a[?KhF=CFRީ3fD^KC6>յ٘Y(Maw7qk۹M>4L\|7f9ӑh6s鼞?zf^-*jywF Ftp2v=z0ԃa?=߹ Ne5 Nշ.Tj*) Nu+ kRNEwŗ<]"wy=S-$pT ߆ٯw+⧼l7L6iNw_^ğ`¸fOn]{PI^;U^IV3L"(8ݹۗV BsT4k Fs㧘ꝗxj,0şYF |hGC,bpGQ( `Uz)ʔ$$k6WmOCŗ F[xiLcIQ,j/п5ܾC '3s!g_.B9 T|z)~ڛCwN:yqġĭ ^D9أ Ӗ9LjEb4?l SB\M& xq4̌ vjvffCI_f ~潁?ټ}(ti gw~3HUdSUO[0Jtdxk5vTGD}=ѣT!}7כy~;iLəwr1̢6) u(2rш%Z\+;uba*?in܋(`U0 Di;#"ב* M#83F+ *1oV" w`RZL_ZQYD,D]ln]}HoY-j/qZ3%m$aVo2fv?^k8𞏯rM#>4ʡwH3x ;q$jT娊sZS*3Q/HyĊ{EGý+"6Şc"[0* (*UpgrG5}x[aB>j5Q+U|j ꗫ/쌳~hXWvZB@raZqz` ~ S^ĒhI'!T8s ɗ5 M 4*,h\G%`3aBha( ȼ ,^Z5ᤀ2$ ~ӭ(F鹞 8Q t|-vSO 50克'q;$뺊x^WGOLA,h,#C[( Ҕ['* أ!$m7Xl/3?{WƑ /cCCl'/ >e)RX俿CR^֐=Hk 1z ?> id14&qzFh*_VTG P!x#T dY^-+ȅlpT`d$MF`*%5ؽfXj QS$L={Tb7UX,Vz Ud.Z/dWguYzdUMI`;cIԆxxL{/J#ʘkᑠ$dZ*I4i k9c8.QpSh0-3b kKA4k嵑Dc=c`+eD@ j^9,BH7>yzui9).S>ɏK7@?fſl$uRHT-Aә]A? $䀧d@ڧh4TZ_\^^Qm7gN k` +GSՃg,L v)u&qhLTN;i'K%jrڭKJA1voyn w!YW0ݸǠriP:cݞRx*[Bj:!RXoֵ([& *A'sۓUT%eZYՀgLiJoMqL"K*cz?v^hUVB9D0uzWMgSY$0Olk%h!4eMEY9[R`]꣦IJE*0kY4*MJ >JYh:%A"O5ɀ7Z%JqjO&[,R"zzwH5ZYUd%Mfɬ* 2qz5%qsFYW$ǧYH&d2)m7z% rz5-Em2k"YK-ZYU4W'Y#d֚Zb'w;̚NY{5F'Y#Hqd֚C+b w8bjszR/ܻg˳ԣ;8yޣ ?]ϸpP#v2j닼+ LS?|5)ii{YìYΡW(clnG)0}3兤YbIVjh"Pbp@K)V2DY:k/) Et@ m7T v$ht%!6'U\:O?-$5i1Vo{WJNZGii8L&wCujW6{E;-4ꎵ5 ͬ"Ut8Z{r= =m_ 糱WE*.t`WT+tRGCO$,D!x+,HI.$D'!{ l/GlQ+\;ɃR1bN`ox 6<=I"ÁNHE@2K\4NKL Nh 5DBP& WBdq8*iAYd<`48F:*@Nw?9l`Pn-IaL4)EK,AmRVH #tou^.%u2YVaHu0 PDcv 8P]}%+Zl C_\úCvן-ekaU;˝BPohWgoOB챩vsP [x1c\ Ō޼V&ܴ{n H5b\:NQ7w0/iBrA{]!,3+[%;3|z& -&%~p:滻ঌYlFي;Ք~iW.iMnϷ ?t7's%w?P[9sq4{Q0wTqΰ8Djh閷[5v }8*jo%*C ַ܉AbHVi .n\|N9߅/|*#`=Ndỻ雛+0x0J'KyO>, ӥi׷0/o߿ ЎFݛm<"6tSed1E˧dKXg !Y}a )݄z@ԹVZL~0\ LbWL)}_ ٿ  ȚRȄI광B< 3(w E ϛO}}jɷytק/SDV7mr)]Uu4tm_hK>\XPk@3h'Hbr -U;.<]G<`_Vnu C4 S-51ǠriP:cݞKRBj:!)^S#A \qR_#'+.^vR7Y݌݇B0ƳHqil ~Z}K1ray!y8q~a0d|yHهA_(ӄM5ݵ(@n|E C/Ķϗ"T-;=dX|)~ٜ}I%,X7QI@eM_;ͳFx3g` /TE^G-"&/f߾n8_ө#vGvZ׀%ò>'ucp]/g F}popM4`ݨ0a #3"ce Sk\S/΀*Bjl 1& ),;dcBaȁ %q]H#&h%':qOaW&yp~i'"=-.D"1P&)0.aO[µ  ϫYؖJ N5%\ϯiJ a/XUkFȴ󼝊ڒ)YDSzfΊ\=lSD%xBZ4:XcF|KDKwem$IzYwFn̠;O "4e1@,RRDYUEJ<⋌k9I"n"(c7IA *nm}GQQ4hƞVzZXD9IJs׎~S }&[XmjHhUոd\UT lmE_<(cK1vo$"U_հn.W/4f۝L qbpxiK/FZS^jchI2Tx<]]`]L.6+"4\i%9CvG"%ހ̱qm3BĪ7\Lw"Q *.0-Z'ﴜCfh18h}$k&y}<]IK enkokmO.oyv - r]$ܬ=:ۗSMD9fk\ƻp"ZC IB Z@WAǝWѷ а|4%jz }ƪ90E?_ɷ(" 9^g($QrQ7PTQ ($8̥lYӥ +c0,+sǘ 1 ɗxUS^K+uTKv(8B%@c[:#fRCY,>z FcT*|zj%;4HiX;G{2LEN=5ʼn=5ExˣWnH7QH\Nm,SNFnPҭY4j͎U|* ri:sn|EpV]=t˯"(ExJv5L Di\:'e)L@K [)NJ{7(b$ 5[Qj5RIz ~{;{0a|GE ͟S_z;F%Dr_q1FKET$|a rѴ<O$kV({sWf=b.⿖>8\-fx.r f_$xv' bq 8BX Ir40Zn-2uN9WY㝓[G#c%@} لud޳G`mR+clTQ uhbL}.ob__תq~m+ؾ7ΗbYX–/_leĔZfsW2N$5zkR{| C}ʏ\ zӝA{'2^jeC +0)%T7d^y4a*(OZ(zv5lXUoO<8H[F- c M)zv1lo)9h E`8cY(IIߗ"'xrͲf{z$oDS`xAp)_ G){xC?Xx:<y< xP cޗXAYp"͔zʀ31́# +x Kؚ;QVܑX˒PEp?S\̶w % pJ 5G\3o)H-&)5lI9Pc(ܞBὅsi{)WL'Y k2>(#Dzw4R}<Jc5;FwbUz@JK%BJ 0%LJT(H! ‰v vNV2^oN jЛ˖f/L:ONo8;x@I3r@tL>!x$t&ZRkT/*ӽ }'!M4 QP5s`F;vb&f!ؙM4C;} IMH~Q-NC}4KMrL)J,`^r%3ڑØVi{9 9 @Z`4 kJi+!ϙ־jG)cHlҁJ/hQy$ߗ8j #$(akWRxݜ&M^7I4y{I$Ct׭{#6.9d=:o_FOM qdR'JgӍW5e{`t0Y1'Ũ`(D} u+>q<9GŢiy]_TP_+0i."NYsk 6Rl-)iti( 8VRNƪ,]aw3y,i'!gSUZ\ݭTNT R׿|?;2C:U ~Ux8<~z)ADAHc?'Iy~XwCH(L}r8%B/>3\PR`|&rvzrﮮ@qSiKLČs@cU*c)eL'r:crzJȸY ` `3\+GUv.U%Jm͠:*@~J#PĿ :C康]b2iG~MCBJ{w!w{߈բwU{p]-oNboO[ͅ%\䡟[> ^;c\qi!H*86c &H%` u%sޝ\C:Zw ϰ=+v?Ή`~3UP~f,\RL(9^#JiZaIbh cJ!Ѵf[TaH2EIpތSA hd\oٹ`mUf#=1Jn"spM\:A1~y{ncKˁ˗o_s8:ᘝF1O~v 4K{AMV-yWDˏe]Fq;] $BV*ԶaRL9|!P5m'c7=peuj4r6Ǔβ끽W" :8Ј=^}NEwoJ2R0U(fuFbsܦc׵\s̒_7"˛o$vڀw1Hb0 -`VlK5zZ[l=G]Ze 钀m#/boͽ1g׋er˳7Zsy(4PpURadQ ц.Z%e$\zRv#Fd]w8#4 Ƭ@mc,QS"(3#(î`"@Px sMs4!G %̚`#>RInߣalc\4v`wp|mr/2EHlF(.EQ1Nj=3JKH]>@J882,tFG*2tɠW]^o gܪ 70R;y򻯞z< %[ټh85K 9iyMF}!FDL^ :3ņʈ̥}El=.xwt&C^c1ZiȖn=TOhƤ9擬Ku*pDϨLO~ARHr6Oo.U&Uv/` :27+rRijQIPBiÁd c[b8+^DI(A#!?V mR)2[P`BYL DPL/K>Oz\̿# oZ-ϊSթ\6>B]6mx_(B£!vzZ /z:Kl Ig/&>"LۻEaf-Ж_8<[]mlbb1T6b)%WK椞2z1;1;K $.!&d1-JR^և R`Et*32pDY%s'Ja<'}`~Bp菽vp0>0?ˮ 0Vk xD >ԥ1=Z_<\VKVw8Q@{N;Ber>%}˟IN>l(OOT+)FXIs1}s1}^E[IXO'"[bS7pB1NLxkeYIs'81NpC:aB08Fr2Aa!j=>\zH(c@׌88883bb2,oP(!VZf J;@}- Lr'ʺr'"=-Zb$:A`1A`:_3!Mf4 BK#%L d@vȩs JA'uDY|;ygVJ*Eύ mJ,{Xx g!p;U6)R,Y$dyEy)R(ʢ$s{˪qRR+#3 FkSnVW4h}?E aK—8-2HF3 !p/!- .R+$mA@V)쀼ja w=~Bj* x_).?'gks=x2*(vÔ8 -]q˘(1~]51+ _?$kG&.V΂)nbŰW6r}19vG8dxo߼ٻ6dW,r ='F./Y}K %Tg~8$bzfH[J=U_UWwUWW![5Mv`:L`J l嬼%Rj4JP1Jx凊slat &SU^tjʤ!9@0uczɏ7ׇL",WpVx]|.Yq)߫<f-;TN?Z t1.B-:xgSS;_`$!LPvO~MaIG&~6hl25o&lC?.#LiqdBJLS}DS–5K}ĕeayYLft3Bv 1상Oaf3Υ74F[{}6l [ă[*JD^*5rUJ. [*QUpNW- sr À`CLn7&̆Yi"%s#/cXp=5<ߵ"g藮bm;Qj,43NsS5RɐTi\:R,퍧_Z E:P`V0¨!c9JSAAV놓6 sip:AI,6VpCJIJZ GE B.fyFiZюN:Ԃ~Fd8YHw)$%cô䲔^!Ji0^sM S3ʅJOlC] %dž2A%5$x*K0* J̃v&h?Ǣ~DJ,ă# c&]PP)KbEH<$vT!.g}^`N (Ɗ}ż2:=-NL 4'h49HNN}UۃbtP| cawNلK9Z05 cX24֜}J0 6. {H"G&V!^ӤH”TlDk*eqNoH@0q4C"fRK{4)BifZ+B 70k?mS壵Vˡ7*T`% &'AJcn1mO5DՂKLMq<\`Xp7O" yK56̻ ;}haڑ pab t~wR~j-k5𧳏U :z9yQͮG4Ѫ¨ pnDW0:li0wOfFjO7`{19l%呡7qp? WᷫdBQz3Q 9-|^+֝MUkU._~|=aQғHC^T=wTm+ǧ`r1HQwn tDZɪ[6֭ y*Skd-rD떋A侣u; 2ݲMn]h W-҉:'RrIvbˬyeDT`ߎfwc(d(&~5wg+{zY &d޽GƂ[I,sn9wTזC sN`66윋@#%jǶh 7GTQP*ⅻ`QFąRn\ZZ+D4 ˉ T+n\[iGaHJRۓ8 JIQW+]VdsR+sU2KKocVWUvAJ9&^2Ž-DOTIA!t!4/z0|68w` +H:O Ptt B(m^B9ꜱx]C 4rp0KX>EH9 !b $jɝS%h$;UsXbÓx7y$ݚyepT^E>.vb_(8Cx82Y}6ÔJ~~IV2DmNޖh+;h; 涓6èwAnҌ@ivSOlA?֢_j{qҐ™>) ~ i–7PǞ:ԹK2uLԻ$=41R j"h?kEs[(uVt3]ω)%^:ԮkgK\8 szEp(W8d>G}A3jzGlp}a8JJoH(GpM >U5)R\%VDf 40Ž* ARy)A"-Qjb,Jmڪ5B!Kb9~"T_L$PIU(/}aY+fUڮXbbi!ghEup`@y`x 1HYiPV(^ecl<Gd3l2pw6k Y 䃔 8EPO{ RSʳpoftbJMIzx_jJB#sQ J.z=N.Z>=(D'6+Uu 'gJw.t[.$~apzʾhzb0 `bӮkMCu^zNwtu:1ߊxdscv o|nBb"L/Ï}j5,Zs븗J-SgO3~;CUI+I"h$(ٟ%zC` DPKiBt7_kr,+$ ǀMH Dzb2^cڈ]l22d#\ d`c4 DOytKM;A!c#JyhʶlO&9ĿOanп)8GBTIsRȭS:EHS4\ >3Lup]`JX- WeAcA8KρvޔZ.aD烽ZPAJ}DsOw@X ZuN'+u2y=R!ғ(_8=N1:ntj4; WfR&{n%)B^xu+V;D)da J'{ dHb}!Vj{#됳&d&IWsxn!AHߑ @ +rQd42_XQB+o9gmaLyq7 b$/vs=]Iw0*v1$I&_%ⲷݯ/C]=:;D(B'1QW)=)%EM_dm1r.]}NqB4':AHO1 ̓ƵZ1B{¥QwG]n}#^?{FL;E&)ب {Ӱ~|_ĀE _ĀE=`] C^t?wن!N~SkX4&%=bCe1{ ҲtTy AVOZ0-~ u{Ϧoxծ ūڊee#Ż00R%xB%a-z=y1OSjl/1l5/ #HhAfxhhrB$:V-NӉ\KF1f+𿻿"W1+BB'[1#o\"0$NIA#xЀQCW(=j fB(5iCdئCUڮ*Tz@ UMJTzmØ1([HcU$ƪ#_ĩE#U싊j.EWf@1[Ğ]w ބM~]{pOnWb56O>f+D׻0.fa1i|Xܾ9;DN(e0el@tp'*7-0&fcn3<dUҾwYCCOӸ9?PD֯XV0oQhPQ2ntk#.溜e+#xlʂ%NaF&cng>5عQu"\y/eim-(cZg,}@m}cpGеu| VR X5!KgHjEwLhERNj\?{Ʊ=[uw&Apl< j %c'[=$pHѡb5n]]%n߾A`E @,LS='\ l.܀m,6'OɈYmoؑT@`>?nzAk=kCSӍA,*>ըm95 >~W xE~00c>\:~ciСttfA"G*xYr1aArhiEvՒJQ1ԩ=!Gɿ;r–w ??^Y b^Bh6ɠn(43Þ (X-Ov'G֢)~Ex0Əc_oۜx95Xx*A|H^yбO5meo1ypTF= 90fbrgnوClA"\^RH%;B{ڹ6ɑaYH:V4j})Jނf)^`qgC%VCa=[]ya|V1lBv*‘p5rBYː{[+-&ռa;95М1$gdi'xJ6iqx~-Z+LҾvHqC+x '6p>{~Go6}vx|閇G̺xoމPmyIgf,˪9:, i, [q*A k~ӽnWٟ us)W `d49j{G|bԍz0UW55A?m`5.6̕.:_l޲9i1Hr qv/ -E'awk1靭æ->RPMzbKUgjw}H-Qy[TZ=xEsaQh.=[6͢؇/j>^<Ju%}sVPK-\G$q6(_f~?mV;}_1򝗳5 $;y$OK &, p;db1HufKKru6W%^Wa5 YiD< bFI "˥'K8hNVeJˠ,k}/^p"R>F䎂3M**N4lȉb5YLE69G[_͂A.,rs%-$`a0&)lea5Ox݂>IX;=}oJioG_ڇ]hU¨a3:a)Џ`k2+1SXzx:cHuJJ#,AhR0jlC9"y$̷7Mr2 ̚ѿ!0ÛObӖw> ټ=o@s}҉3n~ﻲYB_ <U{Gzp 'dp9 #Zf)]fǾ}K>j RC߿}]L<@Jh F fpsQ(EGR¤0Zu"}~.׬ Ϻz)m/tDU%}iA`;ƢIe+h*ƕlfV]Z WMy>VZZΙFqy̆5{o*`1_k2ٖp%b!$pv,CF܈рc k)?_( PT.C1E;,2 ʴgI0&Ț#kRIjgRe3A&KBDӎl4>"p`̑p&;YUgR}QL/y1Bj21W,3bL =Qfk ɣ.{C;yiR4,0eg,N(!Qݗl` V> Af݉}P![%q2YExAZ"VF؏3˭## 'k6K&Ayb^RoVLecz .HOΏQ^Fd$T&"h)I~#:%b6SYa_Z@zO|& ڞ ǎ䲏\fa e!#/8m/qTєǜ )xk03/Me@Eg^BY l0S֐^ыl (lv|RZuO)b.و!5\D FUAQcyP2L =1o3A@&glQa JaQDV9FvY|)'ڿ=1f1%%I/fYky ^dHI(H.36IcvadaV 01)X>{vrXs,I,wQ%RUP2A%19^NFHƐr/I(%C`1TV$仩!JJ1?X`wM[EM- cЌq,|gl7%WEǶe):PArynH#Et#yYhªcd $rr)9ӆ4TPtQ&ƃ ɉG4DMpYBrEЕv-aPS ͢K V$Ӷ~$Y0u0Y}Gg}#}oa/#B{7S031/z m/ez=W&PEQ<ڍGdtk;p̛xcTJ<| .ҺMHݺ[wӒC KDJ՘k_"d.'iȱ !@zS9Ɋ8B{/I⭹j]Tlz۴4cy&O Wˊj&z]_\BsI""|s8}WsףWWWf6b\w7DՒY'e d+;D9+ɻ2ǪP2#\##7k4[^V.:7g%74`%@!:X3l7aw)6Uuϛl-suaiM5)֤XZվ 2!1$ej'K|LIHbIWC!k$ꋭ}_bkߗ\0feײu=+u\` Dznu0D \ioG@JX1&M@ym2d /??߾?4 = 7-hS.!<}DBi:{ާdzo㋿ޮұŏ7\>?u@OQ-e*!;+1/h6]MKiI`6,@\xShZ)`/Q\v _1N0V!Мg.O a޷Q;V ǯƄɈAxT (d`JtMF<%N^z3 `$c9H8-PDl i;TwiDP\fz=dbmxj 5J2ZqZ蔥y%([$*ejF =-Mc*6\kf_@F Ѿ,dSI!gay_9iit]ߒmIzīd*QtnG4|Yl~▌NYzr1W&yZBi =Ӟ94]}xHWWW\1+Kj嫋xn.M( Ӕ?ݰ<(<"F(嘽RǵӍ#ω)=|!N9sy SzfX&3a.F77ɕ#iE$]hs}}y^\ʶ ^t{ X0̝{Ҏߞr_F7ęonv_(Ҟ?~}?}5w sp.j_-W^v/> G1vc̽t4ckٯ`?dD'=0<gIO/S. ]TҜ;ouQIC]TzTX[!NfeX02֥M%уSw^/|K_נۥ.UZNTJRc q3G@*5?6tG;?sΙhӞ&wQģVX{=խ sqpZIx"ۀZ""Nx_ņ84yV:n}O|t9X(Z5uN{j@|k{CJ颵 Vuj'j$q Q,h8'w0z:A1ɥ{.>o6zN7nµBjQ꬞kMza$]XUU8݀sС?m:ǭ9kyP]n;]<џ;rבM]qmx'"^J{8\ͯV F*wQODh̞޹guY9`Hs*m(>uTLq=32}e~v4QdWM&zW6}NC˲l  *+:K)lA)jrvn05X9́ZK Ҙq_=5qHV]ՋG(cV) hr!,zB0Ss1w0&uN$=#iGxJ1Prd,0P-3Jτ)jA8Uu+O= ۳}3ecLǎgwG'nwȪf4}=6wƛ4(ūKm*VCvA𨓤?x̥mw f 2k-!NZ٪;N1, a_:Ԧ,1Q>jJY)Vz;yfgsB<~t#ꛉ& U甂n3NQKO~J~GsK8K:cy5DeqOk!v=: 3AJGxo`u0<I\=x&U݈˯9LXjL7 ڄV,%c坰z$9M* @Q6(熑7:%֥ {!IC}BQ՟XM%(+5S31${#q:,"kɥ2hNJ h'4[)AQR#^گ?4)\f&\37󛬋C%62i.>CE7y0=Fhi=٩!|TIo/.AvZրhiTp?$7؂<@;@#W"\NnգhC< Hh%Q-Apg$5wҍ߫uM0f~E,%_SrvnsR v ;W'V]-8 f=¼Xu^??mZqmW.nVĎٳSp S !v/ _q/!qnz~yS^z6zv=3Ubn;>e-X8O S@bAp4TkeӧA㴼J"AuiwI.fS*~rCwCQL,=d&5(2!>nJo#7Jy˛{G"Q78&ѳ0DE+uywy"b>>1q%DCENؐD 'm8eTQr:BU)<}mX na>L)C -"[{)Պͅ3#LLwUg!53f;c)ͽ5X -u5SXNe[_"b Q7(s y^oY-5TP։ԏh6$ntܺz[W &\PYTDKDpTlSHhg}\i 5i2,.R?R0Bm!&"笹, (U 4㲚1PE!ޗ?F *5tK?C Jk'Qq<Կ޴cR{5_W6Gsi=@'' 1=~=!NS{ 4AxpssOAk89o;oOX퓧EOMa!RUO!!@R¹ iNۅ!I˒tug#O߳f\ L͂[$'+s NBþu.{">{j7jjF" j"VilXOzci JW8)U*C NyqS9sOoZ mm㿲z `*Oi:,/l&&)-/<]p*i7aQPGo ́5)⍱0pX>(zAr$@ϣhbqA[heV 8pLK3|"+t j=:́5n̘)[8v~qGRIn?ZG1+P&U 롂 O]p! ꂤ^ZZ ҠY({vB_2(SbyB HigʹgDv2s9#E=^/mh!NJcCIeh(!˞ 5Avs BzxZ+EdV1f#Ss")N@ K'Jd^ ҏYc soNKX5u#WJ(<.QCQFI#^,j0! A5]!l'aGolݨGSV%)St-S֦s=Y`m wMm:k%$TשĥWбdIYI2]V`XhW;;=33Okϴ Tj8Y_Yhx;z!BثɕF/t3 ,ҭ',a-ka-Ж h%"u+6W$RR y>`Zp-tNmВTBݨ 5/1XB ZH*Pց"׬tQ]nG*{( ^:] 鮕trMm:>JmXRK2d|$R3uu͔R=SJ_ hfo%W V^|)"֗oi~o;S TR)$x ૊9)x!+.M&T/\&L58Q`j_!S1J q4.kjfrn_g@9P Uq!je3| \Ks3K-ifwUֈ+xU>Un](^p ݊Q>h]b8AɔE\)\Ƶl[. "cLk2gv=ˬ]" mq1d5# BҺI·Z6H=.\flU28$ L)H Q 6`o"FAƁX 0 ǁjK*$D <fqMa6m=VN+tѵnNlb(Ǔ308 yi,/Oe&'P /y:ԼLjTpMjۅ^QI-baJM$墖''NJ-@ oXpv,gUּaqo&wCdKq薐f!kwôWhHڈHq`^Sn{m@OM`FxÐS̈́"p夰0Nt @Q|qjAƁs!JiQ?3q[V9=>\Y,YR,)KVT͍p˹X~6T V:\nZ069 $f 0F (3xKXE-Vz|rֻkzۡg6G1&g6s??;<[zA5ɵp6 WQ$]|.<|bSj>׍^ȏ wχ*>\Qj_ @8$CHݝ[!A_xiv{P}輞.<=o7zoEG^aeSg`WG`*?´+/ ;ӛZoGa~9 ;g-n?DZe8wb,H2S5lcS{kꟹR'ۭ!|hB?\),u1&Թ t9(e:ǀoluC8xJk#oKN㯍8w-kuk]㇍ecQЊ*>]&%y<+5o/HǀtAv$Lуn%\"\r+4A,Qˤ*PTQՂ*yЫ^ c(u˘SIi Wlv20LU2GF+SN¾ |JTr49ldC%*PɆJ,o$s%%%qIYaoLUtr((#jt, èfX sXX-oBHGYA2F&2:'6Z]4a= RNJGbA8X˛U>ɼkӘy2#ƚKUZEA H&FY*=:r-)"ܰt醂5)Qjf P5)X@ g g g gE jJ7GJ]үel'ײLDJ`jIX&kYz($ji{fWQжE5s6xH"(. ox>~||e~ f f f fE6~$0! )f(|8:(m =%Hv^X ) VZd`1CNl2xiE.D—' U D^$"IV\ +WvZU$u(1˚4MTQ-{ ЈBx΁9S;a1J1[BJGSV9ea9lɑ \ am2p^J~~;_^m>mpKʳ !n^̏s>9]l&Fho97L>:)pGǿ]oW}9\ֱHY4p wȇSؖ$_#cz$YݕWnWb[?7"c9*!B՞b-E8w2^}*A~{A._zMYiyr  /YO'G9[b?@oę{ׇu=ܙxEW:P8 אQRzFlLN?=2PPOu>7=r `Mj܁)C5tNL޲x?7$vL z&AormO+}n6 a`S@8$3`badjr@&S#Ѫ`k^!!{u)F$1:&[B!,]du[cWmN®ѷqf}7, +S4%ZeTJuҿm9TJ3C'uĩ7f'ks 50?1}M1}M׬?SMO̠YVik/xVPM~D=E׽!*T%k#5xeM{zОuXm_RxnXs̘uSaCip臖XliV4΍/1FR>U.*(By۰LRB%_oQ9!j Y"1g9ٛ:ٛ}g<)hXhl=i eBDPrL^ycNm R\vl$%uQt>TG)afQqt:3UфB$LFS6}l%TR&;J;ȹ>;6DyΒ:vT ؒ߳^g=*6f#v#р)͟^mca`5zTOd(A=^X+FE )q#lT%21n9 Eyd:>GˑH# n7I=41e !d6ڦ>-Np,RqhNȿJuaV3KIz˱h00F\ zٽ^ٻv% o)zyMh<]wxRD7$DbkChWw78PVG-6I %Ba؉T5f`14 eàhy,%rHɖ؀arz !.lQf5EF| ؜WJܼw)ѐg8@mh&.* *bYG=X5S!?bBt_͊.:xf*NA:,sҮ԰MUT7regITk>jɮ 4$İp?ieH'S]3Z]vxIC,W&DI>P*{yU#C٪Xr}r8p5R`^p&c} x[R+皟.Z!XHhñ[Bl)E}]Rse̷+>{$:U(~j6h̻>}u]f4loOn9:O p GzɵPe6ϽֲASzCҙcWԕab J0tdu%v\eP'[[Wו`e>?=AƦ)+z{R#ز :d/{m# KǂlJEm }ri-Z򿝶-XL |DiGZ69 /%Q9l]36{J6}uT{lEweR41Q9T =3-gĞeZikKTjn%+DTJᡭkLT7$?1f .LgM*adS=9H|rÞ|vUW h-)ΤJZW}PګĹPh杀ؓ7q޳ My+qF-emkZOջPwԿߜD'1W|xyeqJl?wz{V~^|kW iC dDrZ+th*r6Q.: [VH!Y:r)S ^8iƿ|9^?^?^?^?]My8&PnY9x$L d&_Hl ^L1q zTD &rO'΃p>S<Yfj_?-oVIШ ޿NSvaHsWX~WoO7 7.-oInPw:@0cᙪ=# `o8»X|}%2ƪ ^ºǔPG) XW's%oE0|)bD;u^\_^,#'T6|eUO(0! )EoEFƈ-vlE+ 4Hz^/KHάVzaP*tw,p6nVN8<h$]6e:F[bmNDg6%)9E(pt &qkWӴKriFɥ"oe=A Q!r)dR ;(E!:ˮ\vfl?.K*Rara0CRf4j~ɶEadIaIe`:W)kճ"6߲sIޮʪ\$U1S,ՙF@8mwK^Xx淗m(9d!bYO:'JXiRIF`\IbaA'9(mrPtغvPTRPGl3XT0ǓC%R4JÞIgqRtI|q{B,p[{njSNG=j?+ f[ߖ Zq򭶳ͷU?=_O./s\'CLzJmhE4I%핤 zo daӊvI;I$`'Jeug!2H!5Ď<:A)hbzm8I %k UdmEOT0x% }e'' pR_rQxa1%mmZhhC|F$k~E I2ygE3S (6QڢCc^EfGnEEj>#Ƅxu6KH\эݶNLF*ΈGbjg" nFYg"h PHy0mW[$'1C)=aE(~d Nj* hG>(]xѺE R 2u^kؑ4FSSY+ ڮSp8?,SNC j%챣h1d fu@+#n@I֡k`)6tBGPNUA۱uӝ۱(k8[ِwqRj=7,yQN ,rjo,Odő&v1s24Fhc`D9o0*TlPlV$S $:X/eh9٦:=aYvuw`?#8+2žw`3 ׹NRn~n^^ZA-~ua<2tz5ױ7WW? KbJ3JTɸ\䎬#"@Z_fOԣlgjG$ܰ| Tw>l3XIޥ Y x$AS=ɦRMNMďes%gGªȒB b:]PzP/e2 9ib戌$5 .*#M{{c=~Q %}dezg+jjLwJԓJ %%Zb KƤ`2;lsP2bjuQ(Ӏ rX,lT-?4 ܠJ4F!%>,HlGoETg$ņ+'aqSlB:TmuBsSv*Z suR*/`VyTDUYa=u.jPZ;K*..#֙Nc~ AIY7Dn49PaT{$"z J8"7h|̊Ef& hQƸ)*Ҁc=~Qe;; wc@TX=!sRˋ bx@khh;B0; ɼђ ^0#˒+JkEI@uqKa0`#gK~BHu AaA(f:P!mkp\-pOޮ8n2Z~6+K3U3'IGX[?gZ cFK ^L VlT|h;(gꧨdf^D2"J [=ImRҶܽ {涍d Km k)ֱ]VRŘyHvT<$?k 5}t3b]rV )c !{NWZt?grÃ;),Cs[D2lZ:Cț)*rc;}vT3l')Y%p !vX3NpϺ!*M+H^;(X3) !vPfPfOGMガl}KewqL8jFlQx8Z d;)c feMY,#ݓl[a;,㦩v2}쟫)aqk( Iq+S(L`bʌFi#0hEQ/1!:T #xij-ٖѪPVv|5Cfi6Xi-xsVO7AGDX>_L|\:HbӑN x&N[1IF :pHq$ ; i0CκH)B%ppGl"nAlh-}Sˮx5{D`Q:NePUL8qjK'#1uwIS0\ .#iP`Jzb[Q V:&]7cHK'=ֺpDE^Ѐg-/|"b^cF,ʘِ&Zj,( a贕cg_Ih$% 6Įb` ]څv| 駵 LEXz6 \3BĄEqĂ;EQs*gD0>F"2 E . .&i 1U_lwPUad2ȧ5Al͢r3@Vh(`s Q y Z3j7 67BpĀD4N< Kd* "eI]%cD-&dM#B !޲s 3ԣ$h*hC 0i9@ 6WWKX)[z{!Rpè"m<Ԕ 1( ε'ą` kAG`3;Ȍ"lCݖ!܆cuR^Z-?R aޤahq[*~ pLp&ZpxvԨM˜%%p֤Xmt#`6-#R&C ^F`bn`JYA,Aól밡e`9$F nd7fA"jtqɨ &!RӍ i n L,vi"E0Ғ$oޠR.!RAH6JH"Q6I%-U H\p-s8nPВRiΕJE|ej*QY\ $ys9+H%2SVjgHn6`G(لXZmhą8 A fsK)p^B,)֋< Bm}rIAqybbi i#@|z[K T칝=`8l窸ps \lpuNق6ȏؐUX| =6xgw[`= }BkB=_D'I#Byb#,I!>t% 'qhL8oCxhn.yB\#jJk%!v\5 n@ qS wYY9joxAg6EVKm,* U==\͘g0Md2(ԻC0^~އ=n#ծh{X 0yIN cz>y,qwŢ0%,=L`,p?:XU;M_:9RRv@uZggpHNZF'7YØ+2NB&UG3'pQjc!r&X#':dJkSU5 u4[i\}p8R<`UIVe-FOܹ?㏓ΟA^{ZAp,H'UuqՓ!7~ G.ap $;+-G8\ӣO>r7?S3VϜ<~wzt|x3gޞWG'<}oL3c{?}s$M֟Ϥt?ϺjN._to/:KEH+Z[ #`R<+ +aZ~;_n wMkb)˾kncKi Ƹ  A%鯯Y'u8і}eAcݚ̢Ń(Yze"}l<[ uIAl:mX9=WD g49, _h5_skƝ%.Cg~i/h̕\V9:L;oet_S J72M_`ޑ_i;$n|q4Q^){uyUo~z=W飏sEw0j̑9 'Ìz9 s/~ 4hϵa7Xx|snұQz;K'xzbuCnp *8ݷ7SمI?mqp)i^D+_sf wQC]i%uG#ZɪT:V"~/]2}`\S%)8xѓb˽7 2{`?^% 2jIw1~)|Cb,.ӖA&ǽN`pX=$G.08~|WnSMK%Y\H[+RÖSgZ: OIS%J8*v.ӦL7d rz'늶NYh~k>M1F1^g|;R%pTmٮTJ i"FfwF)XTBpR*8.o)Ք7L_qɂx_Y"}Ex0釓(,uv^0rYwBW]f]-s[8ڐݲl}{Um7VWQwf{iӋOgD] dF K"29ԥ5()Epع5J Q ꨛW\iosW:dM; T<`!snp΍toK W4JFLX@CUHs Ew,ToT"KEe2kgK]Gظ'5F* BL47d:É"Cɉidhp,ю9{gGu1J\*Ę d|QMu}?VDJlTӱG۾=W(!uz4䛩:ho G]'~({,;>{nyGnT=M֒]aJⅢVe|i#Vi4}83 MT]q0)W٨w744֙=Fe< evy`NXR%@ۓ^e' NͿp8Oq?Nn!$4 0Z֙2Pbm0NHBPCDb[y*x| NWߪzHh:cm勳뿦;[3yi<ڍ+|k)LMm[zBbM_OE]OH,I<C5MuzE"ԋB,i0EoMDkG'J3&Mo OBL%rz 曒eSfE;DZ-HS*RJD_E&1`tɦ4(,6~Q"đI(0֨@;|-}qRv&1ȊL_^}?C1^\ EX9]+\)N@5[XΉ$L*h,3 5ط]W>򌋦2(bqP/N{Oi?İ_C{\iux֒ڷQ1nH"8 AX v.Cb "$aFd|e`"[k&Ćy!lxw(M]+ę>5%P^P|JpE=CZQrΏ OD#I8DG=_Aw%9<]tݦ̊l~jjQ+ r^rwQ5-dv\^fKEF`f޵m,bE[G=mmai%U8F%%zYK&$sw&g#"Ъi]_Tu\ *1ҼRbD-sI, Scd ˕_re엫ςxO4 %6iB!1N!U$$uDю eAˇ&UѵA c00^^+~zeOauiHTSpIq(ljfb㈇R!_TAHYE;aJUtD1G {i0^@yaMƍJ[ozښr.!fwJ+W "}>tp0ei718/{hK4o0JƇ?>@k)V-7 {Ya pCo y +Y ;oUiB:h0=rX5$F %Κv@<Č);NnA"s)Y']kISk֕lAL̫YL]%k-7 ~INM"Wڢ0A`Umϰ)QѳU<(Arg{cUzz",.8@tZ9ܶ8;6ڜ0s?xUu;|胞/՞r@"so.]ܝٜ97X3: ys+)eYv ?W2SL楣͊H`8=~ xҵ7eSLh޽/! c!.#SEHe|d9ՌIfqeǍлwsGAZ&+.nx$ER9 E9 )e5CDo)SI\d7]%ԕu>+̯ѭn-QhW(wkyE~$OF~Xao@J{ %eİDQc BO~1 o_Wri֔eK nDƍo,5ņpӚgXf3TD(+V<o+4mBqSf0`YDq+~M^踯ePTIk:pCaY떹B 8Si;w'உ]M<L)U ʽnBQp5 -vL)Y5x! \ժ@klZ4@y )됱U!꓏&!)9-q0$ O]|*[Df(jJ>1èΐ}qʕChz5I%j۪I-Ҋb2NE]R>9%W A\)4Zao¬b&%'% X@0v x .SǮLؕI2cWԱ$ZjbMQHEB Myc21Wg*b;dю Y/QU/L!$xWF^~e44p̨O CƱ CD>4g`Boю ߢom9~”:!h(;1q$]ϼ;(.@X9 q 0.ݢ^(M?##OJ:V,sZg #'Y:)kEtd\>hqUbR 'R^fD5L{,cO1$ tDzM/mrM;M2E/҄X+;{hCXP<<3Y,VnY|Е)Į .NS5[x ~Ih1QL Z'ןݯv>*=Vӥ:]\|X8U(#7Bj(SviDONy};]gDNywmz-F"g2:NUE:_u饉;vtz3G*viH$v$ M;pƜ-d: Ay`lAq<dفe ԠL@{B9}-B"k#{Ë޼Ym_/cΌpD B;c_S2N 6ׁhP&p 3YC }1Ї"wdSV]bwU첚Nx!5rZJ*>qS$+xwrR]~]t s༘o4( ZrI-@q8%JQgcbQPĝLHѣTJ!\/C@uO[$Qqc( eJ"݋Sd~FRe6؉0)f.o.7${ꃰ\ks$µ$b) Λ@o%4/\dlVEXc2V 3i$ )Īmjp"R6rBzJcM*ʙT 5̟h^kxUZ(3?IG8bc(=4c)x{X2(Ħ<)eǰP">vҶk?ca?;XAsۑBP=xh+JK|ٺ߹WO@`R֬)mWRWH:9"݋Ԁs~ogYw5F*'scץ{;w%N< Z䠶G!QZtv *꒘6&syB!.pv+ ΐhYs֯0wVno-v&nE| +TugVTXpM"/$Ni?7@e@ Jjc?Ok vgdXk%9]2;_ U.[C%s^[ `p8ApΧp#ڈ+줹>| s!X?02C)E儖_H#NɇQtD͸08DIc/18z{XJ]IJM;r]q8:.T2% !tӔ$Y\ֆ)*؟ `KCBVv$s @@lV8Y!b]̝=j*f. ϸvUl=͎ܸ(^Ӯ/"4s#w3ONfׅcj%#ZUnCC?=x81J"HD $SVm޽X\>0BUV`vT2V6 W33BD"4ӞB^AiG^HY3KVZÌT=g 0rLK8M_?:. p覀ONYϿl5#飱Qt 4SEoeOrh K m ?vljcC+R#mvr*6?:Ćho8V .RSUcÊ ,33׳Z[sjHʂu.= Ww݈s"97:ʊtnu+;׻ւAӛH˲H|䉤 i~rfMuEQݪan(*FD8]s; ;43ům5rJVv N-k)n kq-j J6҄37HwRdiI\m;p i8r w1 XiOq~{ڇ>tz= \/$JB%ǥEϢ`OQ1t1ňs87-@]]~

PvYaޢE\pEf*v!bSVXйIcdk =eE4RT&JZ Jx(ӒFۂ^S3ecYa۸`~sָ)VC}ޠc ww}7`X0D*yʢnXVEc82wrr',Gcx]UwS֞qVh0/~>%~yŖG/5qS;WB8 j|{4Nzq2rw3#a^Fю]9͎37Cvp<Whc?=~ j(3^ތ\N<~Nt[݌1#1wZUx>Ol)}S?;S#¸/tnp*emQ8g sYN`6[nY# |t#湭)8+j2g{2#WJ]g{2"Vb Ȝ2CF QOyUP,>ZM"nB3 J1Fuޜ#? &c\+W#I$(6f볙Mgbr̀.0IAuWj 5)hV(AZ$^bɞ>P}Ϭ8/kрK#4ѰD`VL{V<|V?Ậ?GX74 z=O(SwyMbZRg zt+t `^Y權)v@hfQx3qiSkBvjTU; xMubb` $H}B3y#0uQS:ќsrJen 41OD͓+/ V)p;p B'vo,N_NBs0q<{[fy8FgcOs^dͿ|~h @~9Lx oΧ㑗 oSx[Zz@,惛7dF}߈쭁``4a<=!y n_|z|F.a YWY7/~O} OqjP 'w|.Y?-~Lҳw;7%`l?1\^O/<~*Ӈq _&?z?<"dO_ϷA\CHg*ƅ**!ս^{ 5`R O'YnGf{bʮ84{ߦ()ҳ<i *K/S6yv8?g+#M/e|5ތW;{m{#/IŦ9ZHv;e13wי?AGlyBNr篧b"ՠϥ?dSez`j_oك=!y‚0HBo?j:DZ1Y)dyS1UaM;+C)o:FbPl$Տ~1#L S"q68xl`2r_,Ztq?pq0o{ SL Nn«,Ukd+#@!Nʴ@ f UTdCЩhء'Fφ3fdftkmȲ &-V {Y؈@^+Rq }ou%)uIYLh$DVWWu9~$/uzKà+>J_L0ݷmP5B-}$;?|HbCH-~l"eABq_Z_lŤ2)H$ l!g%]9'mzrsxA$-ՁH$?Cs2<3dE3חWHCZEAjrZGb/ņHQdY "* u?0g?$q``[gJ!x_Fjc(1E^ZAA O D'C <m ]K/\?{4p<츯v5 Xzk/zҎ-~,Ռa4$]$hGQ͝ I7D4)U1s$=/{!|)呕,S*n|:ݻmW=U|I_^YX_, ?t—KK]0 iêI%;?|kWUy{>}&Cɛ]j<(5aRBy[i%cYJ3xM*@# 55novH$I5?HcG*yRt)ե6J呥hxr/P5#pr=yI Pi_eT?εM\smX8]^+;0^>ʉ~sM= ~qiB}÷7]2 DjƳo5:/Ys'f<ԣ3TڣZN#o1ycB_i>QoTl]7Z{| νFɝc 'ߛ٪'l|㫟_ Z̯7SWo&]lOle3e&L˛8׿_UMh+s]a\3v `~8*S ?RT^%rF(ҡN x79I!y&0g"xmdg9($oϯKK a*pPkRj5 |Rg}w" &)߾7:)VٶBqwB-Nx1Yqᦕ9oX"d;}h$Wꅗg}yKC Du&g%e,zShIpCN)x X$YJADP8z t KHbQ/{͍WC9gg/!}B!dյObyۿ-x$dAy(C8V4r&xGRh= d48,9,N-'mĆYF(;/ K{^77a. bx D,4=j/Rj\kjW|5 VIBu9AU=Y":ʫn{zH4PgB-Ox(\JI-E,$Hr fA X;-K>MZ9inj.sҺ\LZ3}ӼzCq*hid2̓e ,d`8gPӚ_,@.2XΖ{(":"21GOhX$ k)TJ &eK&6~rыFJyL# Hym$9V r^ICqhfUVr"s}H(g^%;;C^#do!{dyzN4Ԇە&PѸoI}-b ^Jrԩ63Ύv&s:hAP\!YhRd&*JL!9!-yAc͓Gw㚢` k6 E|rD AJQVCP+:ӦD,$Z>#b ?# QzCI΄Pj!v ^" \?⚂}Aq úbLk#b@ӂz OVKѣ9(3@>2BbY- 'Rk00)1R9TL@E׆KfgJXjOCӸ:x͛1ަ[%q 'D餣<`%0(H4Ev#ޑYl\:E GB _Qi@F"GM@(sZ}Ct>/3?o)H!u٤tE{z?^pʐp͘R b^N XNa 5c|J " -gSl◟v4 N4\M I8z9<;L鼊v4h d>ΏNc}Pº5קLfcȄVGb8Š^^}x+ELR /3-,g(iP Awʓ-O4KFk7`[k7@AA&)VfsBjİG[s_dz`ֿ &j`[м* 7w!>u _ؾ?s6IijMJThEj03U:"(Q{5" t EFƐՊ#PCv@amJ81H%E"3GR! ht'%"rkpBD=$=5ea aqV:JqD_NݐL-ryAWFxN !Zg +X(e^!'s7o/&O+#b$%_En4!v%#T.ԊW#_vдWjt ޮSz `#N^6ݦJ %RvG vGk^AJ7{gs ٭UΘčX #eJ^\f+U>^A }g:ƕ2#$< 6b+CKRogyTL#P "ڃ`-QKbgsQshrF&&悱Tk$2܈  ]4p\z24+,*R_֞ꡙkJ*uGH.cst*Wds9 F 5ؿfzDkLC3] ^ &&i5).eV\7Y/~u4˘(`Zu)MJ@v)_4x^q8MO㠡j@Ax=5A6U_уK էmy6\^!M9T}ɉќ(;l `1I'!ǖ9u;AZd C稻#Yu\CWTi1lr R65<#~wU/(߼FgZ8B7R!.Cͬ^KQsW=g(p$%A5]*{!w.o@bOeDlo v!F,9t_'1G>rh4Xn)F{OyϹ H(\޶BP'@3 z9hg]f]9h{p0"wZV$nSUrJ>@\ zS 8@f׮>nƐƃvWH<[W6K?n-*dWutJRaoUtsinQ(XizoqC)7HEUA((!y 8#08H DT6>TH*c=mYр;DF#ILb u}6ݻ|'\CފFtESzKbtIP&oݕ 6E?}'ӣCN_ /oObؗPP̘ftYdtin=xc_v?])jОOgTE]qJL"RO .4Jh5Ew ޟbm{I` X1Yr3KnK%27|&,f1g)xjs77AɊ )g {F_~ahD;X\̿]ť5k8=vJ oGr{zIk6z@?Ǒ|zY`ըS~42OWځ-L -.蓸k2G%388N 'u0MĽR>&8:szoOe݌Kq߈;zoAӊ kܷc۳C ~qba2-M<*N#F[OD62L=*r1 `e>J3 []Ĺbqyxnf(J7CŋUpY2`_7YY3JQY5S$z,мͿPV4 ^ir?sMC?Y}$+ ]\3ՉlpuKzolz=Dtp:ִr CHA:DtOB%PL[c1ӿeK ČXPP4B)eR\sa@0=JW?F^0'It%!#G\al>h^) *& Qw.hF\]g_^prCJIAKPH& {ƊȱŨ($AsbMU&Z=U*cR_ Țe`.'Dy+ \0ea&,5O&Bs6=f `GQ}ؖj_=Π蟧|;}l.]r/&w/{zYYcN>W'&*G]tU 5=$Yg''ٝ9&7؄5kP@ YG/Vt9-ômkcRh+6;)&Qe$$%ĜVR"-F+y3u}N`+|\?'"w NX*y=-[yVοL̸0eܢ}q o"ͦ.`iGN]x{ D 0f1ahF4VyW(kldF EB5pXrd1P/5 TQ*b@x9f3 kΊ.* #-k<J].˫s؞ԣ=q• vGSo (oe~/`gU YGc F>po֞i D[]S&w޵7(&r=c\%ڃ*N՜nܟթ^*7/qV}y#{0/wS2y?镉OlR&sJ%Z DS-sN]xa@i@'ZD׍w* C8f!ZbxLrBp*9Պ Ehl>뀿k;,\=hA8+@KJ0m%('}LrC3$AӈID^}.\YۍNMGID&$Bhr>X"&0X- BX>.̥_3Hg %t Y -fӫB2Ñg\!/N1'/QU Jr55Pbz~6qytׅ^O>fӁ u OBlkN_jMw^|{]uk! ʕ݋e?K,YdϪ*AŲnfZyǐ1p%(Bn`"Z8Q-]$]Ze2f^L-p `7xvC=41m]o72\(֬q2<:}\km#YFTL`NEgZKE+14y3{w,=Tv;[Y-ϼQ;y2rudL~><|L*kx_s~ ܞ ke&@JeףlSLўpy~0irnIp"7]ҸؽG^ʗ,w+~6YnP;3?3o)<@\7EVgdrw Ԯ磥 xRO]ӥR/cp} Xמӑt*<)c 2͈ (hAmĎknm Ĝ"ߧ GoPܴMC-X лiUz:G'>$j{6WN#`H`>"gp!ag Hl8L~5ڌcrfeLAȗՇ(JY{oE&$`%=hZ)&kr'(rgBHj0 Sd[-vEQ'$tu+;7#K `!C $Y 7[o}'M`5#G @K4Z=qu]wBOM.>GxSE-je5|޷+hiOۓ)q)N-T_Jxs餦z̪cGK>I=]2Wͫٵ[ԕ$ևɛ_NiÍ\^-kN 1U㻅KA5zM#poiM$ꑉ̧ͳr`=~}Ut; _üEoOELUni7^hN5h2^Mu~\ևp SRKR\u1+SJx-uhߔX${Lr8]ͮ_?m}m_޼:jfS{\4yV^t< ?rI-Ұ ?LFj"'kaUD1LL:6X悈Ԝ !u-=(-hUZbw k?ZhAGB٦Pvg]hTWBB Ó8o>vp|4}tu2YdZQU?V999YދS !|*:uYn2T2`-Bx& a8\ZpEHKm !#]| T)hMazC" c'6Y84Y5hbh:P yj^eFU:0E5DFuJ2ZT[UCB^v}5t + qV9YU-C;m iEr^*޴Bum*z^Bby =ùd!xhq-blԪx\G &qDFZo78h @X .Zz>E*1L{-qZ g-"p=' Eq/.㤃+e|Y /2e',@PCQШ$xЁ(\SMxtݻMGoW֦J .ٚP^}^==EUdU2e}yx=PN\^ٔ*Huxo$p\ޅ–X*y6[>J1KVޞlq *~NGiR |eI4E~Mŝ&odW€墰Nk"L%kָb|)gƉ#Bc)W3{6<!j!H<-"J9,}?Ƹ)gԟJGqlTly6I-U"ً|֬q; "}ɮύ TI=v0R3R' b'ݚnGj-gMbTįt]b<84I n(8EYYACn.? pGp7yQD!/b4wg,Gښ ZUpw֦aqф̉Y> c  Ŕ/02ꂦ6TQ/֬qµJ²uY6X$ 9H]wƒc Aٻ&m$W*en܇"fbbwVk=bG媖, :HK]VXG5 L|Hd"Fs,F 1Q+̃R^$LDᚊUR*{mtkV9IƑdKhPvl ވ /?ݲp20 L .f!Bp%y6鎪o=Γ_~Pk)>F \p҉ګVk$B $ qe u9uq#+. "pQHd5ǴZX<5!ǹ$ C%bL(J"Z)D>>g`R~#J3*1,`Wۃ2‡*rT ,7:t1|10Klaq)CMn)0FX@U "PጓD a$=y>vDKD+Zfv/ dbmΪˬw0=.+ms345ƽu(?W'aN3UrW~(ErzRܠr…[(|Uy-uWh)Pq)) Jh%Hk(gJCsiEsXF o@is 2gS p ZJ |T++th4gi?y{AQG@<ٸ"<!H-c<$p R,8LX?DзF spkQ6ca%p ER 1e&6rwk@#K.Q밼|."1Aí^] :Œ$"R0$QeaRiנ\j$h_KSX0Db̔`<?Ŏ[dms&r@3 aYO|&lj[f'#kd7>4nœv' r?|;h>558qFIB+:u$6jF IRFdøxqÖ |(\j9U2a$nGƻ'P `u|+hnXTpV8JrT]-:Ug51D>h"{9`z8G"b=]9\ g?G}-0kOm>8e}lM $ZB 1|%; _cP%!OBܯ:{n`J'H%Wz-HgjFze&OS\蠤 e?Mnf=\cp&r/3Ϭ-xƄ! _aR2 >?-…)2<ĥb$GDg&QOv$T7A&N heJ 7vBmbn2JCJpTzqr?}( O6s͸Ӵ&8JV)+$J1뢔u܎(I&rRMԢ̧RTn,%a#KZhv2z"{XҊ43~ tY0Z(IO2I҃ CgyQ8P\9=/ N2L. 4g m&%BG)G;0J ge #TLwQ0|}GuMZXm_YixkQ,/!bTv2Ȓ$v{bdwb";_K8:ݵcqtIYp۪A5P'1Tٶ*[sPs #%r8zOwM?}oX߬ ֘1*q7Jdy(Σ 6=֎BuX{rjD[1@98 1BP1YZ u?_Y*Y̘5bYd?UX9;6lIB^&TT!OeⰇ{_^:ѓA`@V~"XˮA>}ws=sy9v^ whSWqK}G|D)7ރ*3>mV˱X1'#*]zA*&_Sv࿂0VT4SK8ע>sp늛BHByKԙpDLNb:Esk3+MYQ ٤Qd r&z^m?ly2dN$oXu|ٖTkΒhpms:QQ lSTnu@%Ǔ+  "CDgF )LBܹl6\~`;/O§ ]x6g}e6 ЪעwoKYWes_&l CQΫlwM`>#<=v*U!>!Ek\iN-ӽB%ԋP0|̳&)SWnhpDr$X_Fo˖h›Gwvھxt^_BǴZ9%~ 59e)`6Ax$ٌJm8bVBOy1O|e&p~'w㧀|Hm_:_ D %0q#ˈ#t7rW?O.'['+uqX3ѾmTmR~>SCPkZu\]­a."JI9BL5lzmvT%dG.wR~8sK[Iba% ~-R9{\n2~|&*i;EkusV)"4bEXƴxgTsIy0sяObrf=2"!ma5Is">U`ŌSc1q)Oa) yE >xvlmcfS{}Ś78gdh)/e9J8KU9xv+É*;D1@΁B#NxV>#YgoY9kgm W ܚd]%qZfbJ @cF?t+@=4͍$8 nE]AChofNL!R5f+Z+t!R5״ǖk7ofļepm}M ڪތ ƒlkcƼe¨BY% 6 o{KH:#Brp`$u_Jtӽd G)פ?Gsttvm7(ãhΞ=^M18VG)ҀHTe/'7&c>!H +a(W0ma@EqY-("0Ќ{vW:=S4hjq. bЮyǀ!WXd;tyupzݤZ.}C,A}瘇/14W-Z]?__f|vQp8}d$%r˴EMR'Tb?{w)MF[>䨸DHK[W? Um#pܬ< Nj,$\%=د[LB/77K ;#?O =Iȴ_u_ }}qg&cׄI؋J`+cn" {m}HȄ%*vvm.\fKcT(a:TvH6Y*+cё͢j;6^ԛҫ:P kBsKc[1`*zStnA&'Ƞn]7!6uS,@:Da{EXBriyᄴ+­8ύ&F$YS(YPFEbL(pT)%J! *>g`R~#J3*1"'⸣֠ߚͦЅKUt5)YTbؐ}2) "F"`W J LΕ(d^J.$yvIC/`:6mJ` .g(" :$A HFm52fОbXeQ_2MIo.Qg6{KgĖsmscDN)ׅ3uH=y5D's$)ܭNN#([\)C\?M۳;;ODGj .y5-Bo7_TC?{ƭ K/H_T}:aVReS*\-ȒLv.!)xLpFht7~H~s䣏QO= L`Bgnt㢖^gXt{ '7ɖ1EdCEEIt?苙_qK-EćV: i,u Cd"#[KzqsƩJN#OU>ar#wח|r,JA)`.*+`GB.G_Vck8G-'wU链qYmϞze)^iYp'i/ †H,K4hz4FGT2Ӓy&>jq_sgi>A<fy -gTɁ8ݬD!R70b]P3m<8mQ.Ѩu$f?+{#5 lWɲ]%v,UӲ-]Eti IA[㹍L :Xm-#Cԟ_j/E4uׁȍ+m6t9>:1ٳOtB[nV9]F_Ƴwnk8GZF/'41\*M W͉ъy+4qi*"ⱋX9"F)FJ2RDh+5kC4k+!35Zc8IjBN8ET+cFz#ҀpCC`ġ)o Y! SC2 {-@vN@r0Hx["KaD/A[x KmVJ{28ᑈ:8L(ל>C[n2ݸ8umHn Zhڳcɣ 7.1edVO-Gм7 D\IV\bbHYεEGRp<+zӷ4w?''=6Eh&)0*6piB L}Ȝ:S|}8̉r/yz%+>E/Niҽ]fvp.]~\1!/*< кqк}#`.P;!]6_ݖ][xm%R(l5)k-(z)T=^.~=}齓Y.+bT5 lFOn'F)cs7^zݼW"zț[o^7LC0RH.DTFNQ8Qs .Z+8Bq",%2zl(2P 5ש JHp+s%qLU0ETPKi(0I& ӀX"YbHl|N1٫A; EBR;KGY/Kw/1P|2(ǃjwu$sU":LR L)JDZs'Q U}wb8/ >FnEۅ%`Qm:EJQaxeͅD1ё#=ޅ~4㉉>.0Ya;Lѓr}.-]|_]_nsxS9Hvxf3jF,hM$g_XTpNV*6KJ+@nL&V޾^FXKwc?beK/Y0B}yPO5/S*˞ҌNi; |]ZmY8+qq9乾>N_s{Ζ;$% ɒS!{Iyt08v܂>LxwJldL !SIbav cEKL1/XaXI1jt-C^r\RLJ cɞ,R1.jVX" @{>@"UH:Aow,s@ [Wi0Y;B1KmopSEOb*ϨQ{^'<2p$=`^d) =ǒJ9n3$@F7gi6 i}Gj C5_`"Sˎ(= Cdt]疡r6DUc[Ts9{B] mʼtBY @T$Bd$*Dkly u"vN? VB+ɕUW{= 6t?ޅ7T['vEbTI|cn7`qzChAS1G-FF4bRm4'0!i\mzZMŅ VNS{+XKSH4EJI!Ukoj DEq_ˠ ֬"iid; ~m X1ԟ9).P DjJ>InzE噢s̥l xVW[w"Wq穟']N[[PW(cU/(hfAìb% k^@ Aj8nvϮgtz*pszFuc 2XAqĐs;rAhϟ^Z/Y<ўїGQ r7AE0CD԰/`Hs^tLdS4ђ@tJS(6ˇvZS4Ų=VEΟ,ůneuYBcTIOWϮ>wV?yZJ|d-[Y[~t->O x ۃ 2EY@O*<)/^؃gP8Xw2({Q=#v_ T%v/AS!Gl7xluq/M(ȑB{H:eਥѪR "`:.J`sĔR`Vżk~ʁE?sr#GwX<8I=h𛳋frq3/RϳFŢ.Rh:04mu{7]X:)D9UhXi8ڋ?FTf3yb}$s.q" ;.+1ڮr<h5,j-3!H&kI!QZ5[R-a!_Z:]Gu6q+weerop(Aǝ8`%>kչ-Fzl}HbCׇ99V Q"jCu/R2M.2SF^+^zaGԜM}&:=%STT SdE+ _s2m6n<arQUOj{MMHIAT@FLKN}nRr,G9B{٤b4-:('ƈ980B(gFNL5b̟N$oiȔ_ BD c`VLtQ%)㡢9?c+&C#HJ2ɭ 'c^Hk8B01z5ptzx&NUAʅ\-;,vſukX3ER_#fak;C礠J"Zn#kV8lV8œ 9h+­¸&^QT'Q)yܻSR^mH#-ĕVM\jrq6 גGCT%`)a` ٻ8ndW3fV,yX$ާ nn6;FIv6{}%=3ٷ&lKi+oU@S1B0tL2Qz::jxPH?czHϘ޾-9㑤Vnnž?1ÍcF5=r/]^&cI\='9*c2T쮮~&Κ[h̀VTi6<۸_n#ʞ;Xms2| jK`QnyU-FOq}H1دJ3VX~cp\̵} tn-U xp'MUAǫBa?'މ x1.EEehtB+NLM?SE @9·v"]}Ј6ͫp{ 6%L7oӓ. s%–3{Pq)$\|:Ezjzj)$ XNNZ&05 i:!)/Lԅf,[_m*+gq@JboT|Lz7rdrY6RTŌ~5Ԝ0b+lBOK4 IC4 ICfPP I=kW*KbN8fP[T0Sz IԌm("=dXf-F0S-lDR* 6I`ai/46]E:d~`G6Ɍ{%A;{|UTd`կj3Z\NA9t^ ZQ%~ף*,A!Z ؂F]v.UjceMRDQEA5Xm*Vye ܈*-9*D0\/>I;$]2-JeK9>d, aV)5);~2b"KeK=PHl\,;7aVV*ߍʕQ3~Q;R!* QБ4${j/.+<,`EqdJY4hkZK(uN( 4S .3l@1;ٌEM=p> I?[0Gڤ9?UûoZ0`= |'wo o^P? 3= ϗ~B 4}(e֤=NzuqY!}!O8*Pb.>=(?w/^T\O!7~CKR?fGB!$"6mgy5o7'[-_~5?j]G>mA1E|hAOzxaAJk6uNmv=~ -֜_ux+AvËhߥTՂ(SkP҂m]\PWm DzTaK9'Iy6P`s5lݢԂbQ;3Άա8`wnlmm{ߕ2om!~A2q5`("qp=]]|*Æwԯ^7v)K,9M݌־~IصL^>?YQF׀;&Fn[~ ~ѷo{k5njG`~rZ/n>Կ4twſo'n9UW+Q:T+[+P,UG1[иԏi/:ӭ{8^7Q@"} 6<<@ub|NwtLh iVb%*WŠJJhcFZojt8ATW02Ӻȧhw=5C.r. Pz.B+KUinӞx(p &00}sGus$cF9Mtt@9ob104:XRؚa8oO2m.?Eb"z`=׽R|ݟg a?b1@烾X4;jFy_{ ^OhEK12.瓨DTQ*[TpZqdV ~Zl&t;H_B pΟDuAd~no&Mp{k/ Xo N G\-ȕLZJ6 d~ e6g2ҔZI%O}?vsc'Q+3NR}JY5QTY+qJVpǽ{;HNQ!*6$1ouDsM@*R*v8L({11uMhP #eU+CnEYsF+7BoTT ORd7wkdk2Ä ㍚ʧ:%gE킶 X֏cmMYJ.G P)n345ʌqTmD|+$y2%b{`x) _$i*H+xi@zjXήYy"-{WZ~ ee=}/Weyc}:6 ۭn!ֈYDMl߽ IRh~j"d)4^}Z(#ά wسW8j=K },zn@zӏsT|A%V#rK,QzYdFu^[EHEt%ST{w&P7x/pBHދE c8(z#M;8GJ RJ&wha&AёLNY#OQq>EɻWtNmD$ 2d:ĉ q?*zr #5C fe%,Ij9^/dQ|z~ԣNhsv-?%iO2Ԩ>M&$oM&$o7Sd5ՊLB$H@o7Pf hZ 5cQl܆"Lsmq)z͘֋Ƒi $?zTetNtwcyb"eB3=B5 j3OHMI0 R{JĔɢ$fȼKKϾQyzkhϖ@ECrTްqRMx ΃00:8`)HwROOfCH"jdX o>&l+j, KL:Fsr ȦC1ރ, d&6 i$e-o ;33?ZĚJ.N}f|l-\;ں*KDQj%eVb% EYV6ju6vY1Ĺ k)$i"N$ ޵N,͂âJ+zhy MZMӀ5-1e|uϏt<1#M%z3rԴj(.H◈-q)K/s=1{R X^]y, Gt=QLc^꜡(%b#1P;(dpn’Mq^)%"$N Ӯ8x4\dBb'Ɇhs9ZK`qldNQQ1\*Fͪ=3˪.VSi+&/EԖBkÝ+4+BZj%֚\%g@@n.^(``hb@yqu> P|>ṵ9ůaK29g^Fk#E޼Uް!CCtE'7]ůU'?vHٚOIڜZWê5XYI Z1^"YQTuU3eseš?NDaD*@!*NةR!TPqB 1~lŁ RaJ ?tQ+ %m $ `ㄑ~aV7T+9) ONa4rl5~T*ʐ4 \KBE 0?Q8/ f*,@{cSⱸg8[ʍ˪.f%2ˡw@Jp/l,~7ꋴe-۝]SvǝL<6zxɘ5,4br7Si 3$9O%>+ϔTEӓGUY`FыZ\?S&~" dbM '|dK?=9b R, QPcyhsHP'Nz:rp^ƄsԢT]uN~hTٽR,&ODB`jL%ʊDuxseRVZU)r= Yh21:~*'ʭ*9s=\;;ig]e=Sb߫J^o̐pTj !͞P 1A)d*OP||_\-4Qb-8M mFb_JeVRʯqUV" t~+h\ǴԚU730 Iu_~?ª0nt3{HJڭ2bET⵮UU=R.NL+I-7T過2y_yX0hkP\~7}U1;zc^9z$U::!cN{J} Xp Q} CW:O7 Au[TT?8˅k*Wzގ8Y뗹 h\R뗻X,(_{dR(+^r=`@|I_̈́0B/K(8gf:-@ +0F3e0(E JܔQ(Q.~ y ϛ77w4܆\Ϸ0.ݶ5jn{hw/>TٍXfl~xلMلMji.uDYhժkVAP8t vZVEm eGn}6MB~qwDc.q$B"7rwu}rpgo?~dG$WMQH3T65OUW?/U2U.Disvxa¾. %DYbrJP<S];-9GR9֒)imjQYe 0` QiT Fn0% ^Xd"i/T2;BRI2$V$`^iAD Ѐa܌k᎜ M"y%a%S! -ςc> $.3U":APOp(3äaBNPjū7rYh4irDHg 6ikS6Ƅ!vC c R-S'J-C1PH%%*3g4Rtg.b/[ ^IfAODD)M og^P3ך r"F;%`x=r%[(Q•G2ۿ'5j;;}n}VO͐vtz Xl~f/]wl,;G S3d,{ 2{Y39i]=nRr[=͢tfSgmZl]W;ғ0[wpGo5|J m~wg Q_ϳAz˴tҙR-4*ɃK[^jM{"PU:ŝT!ݠeE:)3Ev,+=RO$]f`kK t MMFbmj}u]vZ:̨"h#0 ܺ&[W1TM5̨-V|iFoz0ZkT\^djB90Jύ F -9"?3D*PqcYR^gCB!d& %a9S{!"<2hC~x6jIh]9h,Š)Lal`dP''ɒTB [mW@D1T0'`C`^PsL 9I1#eFٻ̇\Jb˜Iqy1Ie4Q)-44 37˥ 5L?>n2Q` rzHV+e8C_/ }|Rc oY)/0ԛOgUl~MJ_ޗg[?Kן 9OOi>_b~|{S! rE{ y}q!kcfF^ˣ~9L)0P KL˫$_+#)_=.*/bXDpImۼCwepR KN}2,~N2#n鏧~y>Z5Z 6 (i))춊NAVS BJ=Px5t*]iAvm-;ޞ@{Wu%9lVx/D]kPW w _OReͻ|Y;Ek+9ư.X5R"cv 0%|j{$`  U}<~ͮ|er P+f#QtT#aW;ܦ)DZuR=,sX'2Ԯlq2I= O?oPC(c8=83=%">^=lU Ea+qw8V#ƪtdkz #HmA-Up~4־=P8fԛ B}[m(jm\2r^2DmcD S+dNWFJ^M c♄ë^364Dn]f뼽_j>P9YS)ΓwAM~> 5ʽ$ŏ~43.<|J޵56~ |e 2o>[1OQZ=Ԛ9? nsWcAT)CBPʖ7L{dW })Ӧ͓z8kH-[/~ѬCZz8B"|aaDVP6ԑc<Mc0*ݐ7{ {gشIRk`^KQ9B@S9 Ɏ;no\!jߦ ج(#^ Ø;j];?% /kthFKGG`V`T#S4F %ʁ?9*0}R6b_̾\hθRcIK6bLĒڽc_h!oﳜ* uP*+>GS_XSW^VkRq+4<L)b HfoS׌+%X2YFGhA  ObЀzTS+;N~Cb%numxa ܨKFK}Oh,ij`Fh{IH \Y]7,jk p㹭8OQF !hbN@^qC&2b%swzVƚ4R C/-5|ɦƋ5 2Z] n8Ypj8H_v9.U" ;b Gz#E{97bkY⎋Y= jػŬc@G]Guij`Yj\ۏR=ڿFis𻻤 ₱ xp?+:`(c~ÇI)~Sz& tZ>]tʧd~9ys}MF oG/'n~3 +i~C^8NҢ72/?Z_&w0h-{~7 yPdF(lH>/`]*Ig;H^DEfweJ{dg:YNvVNVwe\A!JIg@1qo}:Jd2%H]YߎzolZX$Z$vvKg KV}u\A?|pϯ>:(@lB**W&h3 Gšvsb/ڴb[~yhMX  IV!΋_M _0aI^mO?_^.֒?h|MeeHOnm%/iN$irr1yk9r>{u2ӧF I͎h WD网k`qko n}i]/P iq^jm$6y4셠.k΁Zd9 TTXUJ%Gz4SLkK{8P3h%f DF3$HJb!3hkiC \odO.Gl.%ʶYٖ<+ےgm:Mr&ʬ c(#)>[;?nyq;K@ae؄F,!FiVDOОHZ #JzgʂA9iҒ*mw2X-QIl* I l`qksα %Z:S񔬳1aX``ICh\?4]16"( 0KBb=ZV[Ld@ZYXaAҡ_ Le`4dCb Hq+&GXDҖ_-`9[UtV_?׮*^OYX>=76@ݏo;HS޼|Z{`hUi>2b'͗:> c/׿sf.X|\=_\![ i|έ>ON|zu8h̲ͅGF#~bRlv$f]e6CLF̦@jdjh؜ ]lx rl,f8%qz`d>}3c$M]j|A&kiG05!Ih?ķc?Pt]j$Mk.DgiU 7An~/s5Y9dIw|;$8.W)]aA ْq\Ѩr.%AKK]YG+Dl>4Ŭ1hyYCSMlF.UʬdQMðd*̈8LGYg1=oO=Lp^pO5 nűU; ay.R6Ukk[0Cq/wRZFHIsS w.t{:JwฏnO$m]q\Ӊf .G plr}ͿBd e0ŊDcY&gaKbS麅~scY|E6Tx΋mƻ^Vt&b8-Jb q ؎ CTu}RsrNw'ޓ%Ҵ%Ja)V'3\fYIKZp78Q0Un;=L\v`Ԉr |4*sgWthLv)xғ 0[U:"y'Q$){οJ"qHN1c2훪A$i&Z鵵Z[O-@49F靳(nW[bkb8&>czWnvW>@%G{+G)ٗi?䟟|eӁ78[ޠQ{迦}(~z?_O'}ύl}NuTvkEhW1C1h<=d.f2⻭ QIa30 l)=-Rdfg$B)A+%L I@m U,1*. ' &yvli{鳴5;l)sF+KEYBS+W \!R̨4Ljg¹⪠D>R9ϊ\`m3ś;/i .+~}ye jv+4KtBO4+ P蘭BH`ՍcZE M/V؃[*% Ef:0C95=68؁&F:Z^Ĝ|}@Φ8 i[do] 9^V f Rw9_VA-o Go7w>19@e|L|cNY2ՒCZ^K^oZܫNz߯S)gJO~If/)gD Tg(Ie^s vUyI,XXrr͊uFZ!XD׮LEDa_L&O!&M>ghj21]h{?b2`-섷8R Lxϭgŵp|9Q4q]~_TWp+ܬ^P*P|7W{ h0'yg~(-cP?{ 'UyT&*: Rn3)xnKe,%Uss 3 3E#H%1 އan;ͼ#lg /3hXS%D&97[UZ"N踐"f.TC5l\d 20d/dd*Mz]2zʅF* dQÅ4ӻt_o m}g vN φ,|kR+Wy~b~z6E\$}&$#dQdTɣӣ7c5X)Ca+(g { .\rAP6$xy;ihL90CBnRoL7h}9Z 3ZU%ͫ^8TG *I\g#?-94G޾LeW&>&JRuR(!aQ81EJ) ڋP\|hUsII'$׬/+(^\2HJl'ɡV2i;_qL; &G3f?|yS}bon%oL=fa(_xN ϱ )$#rdKT>χ,:7.^S\bދt qQ Eßa}\ \XgIZrFpɞğ!Yo ( gxCRJv8Z'1a*-)~4߬~ox<8*&Yt~};ɮ76V%כoweKF樖Uпw!XfUU7Eb _j 6ď|$+˲5]M]~{g;՗ |2K(YN*\J IiJU%+FyITI$fmy|>[Mm,vGc~jn!8.?͍ahySjt)(]KوmMbw1Yƈ3,L Bc܌+ĥiN+J䌩T)-$)* ͉KUadڧA@:'ڛ bwaUGA&8`e3ܒT>FAIv8kһMtR}vWcA. Y2'&J"4.c$UY 2CX Ris2d^s=8 vC3+~}ye7*啗bO3F`><MwT .ZoC;8Y;B493}t G-j cafR$[:M/U$&#Bv~;r~m KlV$>{n=IgR) Vм >fs&-lQ=~О rZ"m4q↉V1i! ݷ! 7휝M?jm,4dAIpр7 ."$Xa2,vSha;H˔AϣX Qt/=>1C-(HKaΥn8!*I%OQY$RjL rcTI3Qd@")A@ 4*v;wp{6'!<-h25?1؋zsfz@=cAx?|ޛ;4!s|T1YE/vy#9B 9^ ػOًkX7+^biE5ՍA[lX‡͍&ާ}Z?9^+%=V 0|riZFg2SJjYu &\[!̍,g<~2F|v7e0O)~㙹ɫɜ;/1Q]w|J‹F)q4jMDЇ6Yb\HTscAb aW$= -b(K TIEV$LfY$/8)e.@!1#ǐ.FZ^||A7_R bYSNObl 푉\q\gyѝdj9bN݇weH1Or ">bX Z^ۙR؁ў{~~JnG@Tqİb,lB4E4SDgvb#"[0S2Q ) 2ViL맛ʹ̴iPX$ M;o/m^E=5/$89BB7(i7[>Y ".[=HA"&2L2aQEzήP1P B`;`kzr#%_0 ٶB=Qqu(eiS\&yM>!&gwBjN/ ^죸3ܚ4X J|0K&h\pZ%&ªބѪӴjUvr{琜G4dHޖ}9 ^mҘ8|~ݏ߼9Td2,+79XRxDΘI$P` > HI N%F @ibJn 4j<' 7ek-P] k;s֗GJ$,ֱ{蓘=1~_f}5tWmy*oyRhw8$.[$,[:'vAߙۈ1+ 6v?nA""lD{=mNljwG$ RpQ+8x3grG?‚к~c=|T4sh/"]gKMhs/3m=y{ǟKM@f +oCv*PjF1gqT &s}z\Iu*D ?b0yN`<|Ѓ݃꒶׶[3;;uU.!}!S>3x_Nr^;Ѫم8hsi~^!;kT9̴ǰ %ͭ`鐊۶ #@20*f>÷[ +G:Nv~/R*+-ZW9a)uꚚ)z%9ct4!u/43,%Ps9_)6?}d_l NTqOC&$lKѥ{(Ƥ:VPTV* ShKꙔT*XTR(=psQnЯKy!ldlp↨-Ntط jDc>~-^2&9%D=89`buM2Cq$f{'IM0QEjed?6U+FA)! j3|ٱۑxtb'!8{Va@ornf7|={o':Z<2o5E.5W7Y'"Bq$E(h2% X!Bғab<ؠ&6L ЎOwXgˣ4l e9HU|B +!Q1҃ >MwbN1ŭ|'BWܗE8qMU]s2{8v?_vAD }slN9wqqIL YlS}9JPJps¶5gM ɉvՑPAiN|'e59ǢR2 u0r)QufS$nzb n:DLTώiG=|:HPC7hپ24ŌJ39c\Uʶv?bf:\ˮRLdӊ7s/)뮿b%,yu0+dݧYIxsyju2 NNdtzU~-??y~wG$ = .sj[ysX.Rsdٗ`5Odz^Щg k"T_Ⱍ\k6w`⁇aS"/Ig+`0JOpԃ=8͹5Y Y:"0u\PJp_eCm2TCUU1v)ULեQr5wwEƕ8E/iš_'+!@ɋJbАpkV`1[T`+PM@l=D#2A6{=&b :.B@ފ.p JSwcK8 (nkf|j#dZ*V Q𯼿]s o(nm1\ieb) 1()Ne:d_.}L-oo.Tzڶ"]Yb,!(mٳw-Dƛ0 (@[i'(C$p F6{= ȟCV! x';r "59~BيД*8a!,:o[\ocS:kObz$R/J0vDyEVĿ+SHDzwH[mؿJû ۾{ <킬&Y& trj~X#D.<| d$ u>d:dTmzfbcIed* J peZ[%@f'2V4$Ӷ|>E jqx2PEh 7Z5>sV/vMɡ?~f3C~F 5~_tW<1ތ~}L_s81Vee[&_h9V/"`ʠh]J^H?ĭ>UfO`- Y+Og;o]S/?#h˵IPL;efݷ5q$=XWicB(0h@b9Nt!4BI+lE:4 t.P-BW{ L5eH\#%Q(b"&%vDlq1c"VH(5Xt2RI*=cs8۳Z(R&6q$A"+b :0[?į*˭AÌ'S \?O 0ӓqsn]͂,> 9\xYT֌&e19WK%31qoLbXZGL3H`R8jpd1jO 2}m0 A;s/ ܙ<$9x_nLLL!աkQ w( '$^ѭ[<$|`.W,ɷWwsS^@ \gWx>/04[Σ$M'jǙ+"E>˜x>c<'ݞhoU/vr.A櫙Eɞu9 u-RЋJ݃V|/ ʃ;[JhLrL* &&<'mK<6#„FbIșV'0g# !a'G[Ml4DZ&T[l4H O-w rVw}PAetP$9Q $W HDL2 1 "[S =0\{'(G/<½+=.UmظY ". vHA"@&6*+Q+) )Q8"Qfuu_*!ZRl[%ΗpMN~`CfKXsSZ-35"3ðӊHO0{S 0V^.0f f;+e/"Z"*E^Z`;3Yr$(dB2g ,c+gA8I/T^ aP;hv K[|.̔Re7j`D0A QF8eMpל׿ICev/[NdҕDE uc^:Ux9x6Sm9EA . %H ՋE*J0tAK6=:kbd~NcCyS!>7yiL`H]7ȃ8=V~AAS U_L 9s|sb3!ϹG˛Byf-rJ;Yn1 :n,p[T`în彄[(YbsE!:T'#CNkw~;$_t/ hu} y6|~љ8JMl.~hޞG.蒌pTCl 6xn :LRH3&l7{<6n@t>׺\yFPc~N<$!}ԐთڵќkDA^'ɷ&8sN1'\wK@ӓ>pOC8 ͇qsbڙw V"[ DSyiGׂZ:>]A2n'e`t~猴ZvZ5'Se"hr[=*VDGS'>A҄jQix~~"΂>mM@ҧENByJri*.^oD< gm0T)Y^2E&b(κ%:xè8f <hP{+^N of%AOr1X") ٢E=,}I( 9$#$=~#ӢARˌt zA X;>Eft-åE- -3,4ˋ{'k+::#wYbh4A:{?w~M5?N&1 UIH)uɣ*$%2I#dPg 3 ?\&JA WQEuGQBRhj':G_R E]~ZvVX-v• j=>-Μb^Ɔ~NݻTSSӰfj[gڽMz,y 0*47Fז'TK)i =JMg"=Hҩn.yML[8IUkK3[n ĤeU.=J)#D˥ W^*Us] d%NwBh}M gq־t֞oBJAhn-gԭhۣk\ѱk=B_;hGb5Q U{lZ? YEPaZء }ص]m(;iwwv`ٱ4}K B˜IUo7nU

[ 7K1 ‡:?cl6M}юK+5f-qc4i;͸R9}Wz6 %U5e21.` Ǟ3]/H]nl"W>q( qD ާrwBbB?.tr3tСxE._4J%8BNBl}]IԆi-(4N}FTzt(i6좫O{a{Al d|O'cT:c1?,Ԋ\58= )$vq\b3;˭OJIǼY=% RBA#gERӳƲYrJ~R߳GK9Y| xL:}Q?mNmw?[ɗ$N rseuDf'{TY4 [ ~ J3= S^S0jmau^)%ӑ<= 3rnr=~R=f}̠z+3 c-+ykf z6s@rc5;-a wk͍ኢOd˾V/oL_ݮ:ɠ?x2l=1 ^gnCZu"K/x*jvg+CZ^w4 .]ybr SAzc|[ToctIsksҌI!Ze3|(o7w6z+=-Y JJ&M)c-6T 9^)]܏*E} 5}I{o!ւ#A-$OAגBI)7Z!zT򓆵_vՓ$' ]} 6 NU_"}-N0Iy oF.0*Ti+LPl&B3+w\abd^0,x dTH#Ғ`S Rk*Q$DoRއ%-}_-$X]l+d2مOkٲt)kـ|-Y4j7HUi)RH ^7"Ȕm"MS1_X]B3 b<-K+Fʭw({Ns>à ,B5KQSo5gHA,uXm@dvQMکYe0ҁ `u=lCVT@HnC`z(3Ppt~h. g-xY 0//S?#;˻%hܨ!AC|#p 6SaqHAd(is~;胉WyR&RQx 8m$Yd&:<> oFun6]qfת uۛ3$ŠV yBv͡6L͝,~JeT~i wܼ:scOʔi,'=_s=?҆-aRaYvp;˨ ~kC4lQ\.Qp'SnA%cmm<;kCڳ :JE/dgWH_n *V&>x*fEOT/҆ngoŢtJ%8 y|7*az0J^U1}٤h 7BZnTARi7nr 1A)o["hVio HjWgao/'!QiвyFKsz [jǵ9⏛:v>]ZY3s9FִQgF(W5QÀdU7;gHQtN|6iFbImBZ9ٚQbi`+Nxz'0-u1vp\ d61ӮB%V$<vGX׷F9tb 9`Ǘc֤? Ð g x/]՛[rΎ@hFqN81xF2Rned T vY ע->M$]/|\=AGcJ#6߬TT{M쇝{u֥┴ R z5-֛ü`T5Qu)5R{U IQ[U 3ȈE^)d n.Y=PYTtEޖ!ZQ1E#P܈6eG":ѲrDF34Tp0\/ֳD{'Y'ȝޕdtD,^0aVrrC:$vl8@.Xυ-78BMm#?ݹAzwU>0$0 䦦ՒZc{,;PExZ Rt7EHH 7*叇C{fpB=RC4N,&o>jVn{Osω@"bC|p 8tw8Z9~` }_w㧟ݮ=O{ΰwO3{'/p>NĈ~~vvxËxYk4o D^cᵣwo^sş7g\=0fq WԚzow1edz]g4v։,uu;x$r4{gy7K!Z ԟ/޻{q;@Ǝ|pG/fYB%sBݍofRU Y`xsYlM_H ہ2&M.|mGnT(Q0#`sMQW*@y`⭵_ه;3db=|-< 646K)F} nZ;ݙ?u w+3f`4fϩHԾnhh V#=^u@J_=?֭'bR{k'g%؍{dԥMnCw̐) H$*z9@IͰ?ė@??fM| Q@$|PK?{#DĿ\_]9'6Ჿ.c}vX&B,evfsC݆n85f[ӿp/?q^=E$~6cSm{vڜ2C!aڡhsxK $|k^+"fp5 O᧢<Z;;+&>(koO6{ӣf]t×>^=y?eW,.~z{s`=1dH,:Ⱦw:peQ͹lo~\Y;% M!㇄aD1 RS E2UW'9弔52cį0؟N? qrJ^__a\7??iA|{"PD4h*r'< JB Z CW Kِ kgP&bF7 GƫWUgZ8i1#߹~t'Spk(j|pga~/0,q Q)&kz\ԋ6uۄIzgP=2UB?ю>SH !POGq7p("$Uw>o*WbzJJK9~r{Ka#p3š@xFdXQtw`KánB*+]܁/".};[㼭6?66ge ޖaلBSmpڊ-h w/D}v~y~9~o 5e\o^ƚ}J_uYHIK&g- JyF#EbߡSG*t&2ޗ 5ô "M+fh$ %}qGjʍzE[҆En0zp3s*C}jL!C0'tâ,6/F.2J#1U>Dp@\J Ϟ-RȾY&z$q\½4-,88q+g>}@3lyc&L\,_WR<'AV4xӃl4]dldđXi MŘVnm&]GŘbV苟];di؊Q,eMC&lI;nlGwCc׭&pY.+L2K4qՋKD>5Pxqz[s9@ .Ozfϔ8߸o;{7 `5HumdbOSdj\?fvZMWΤiik|bgX*,I\$/M."@ZR&Mm@\9O8Nm*t绐y.Qy23 㤾[ _2'dN-)M ~3\ c( @\vgTFIu*q% <[*fN+X=,~jBQT}'M`1GCdL cVKu!x%"*S]DuWxc#1f$̌f/S,5&-/G9[㳩)X1Ry#P $EM*i24u$(/tBYklI=ۦzxءoȒ&D墮*F}2RQW"7H8.Ygffl=SPfYG][Hͭ"rnuptteix3%qt׬4dd|h-4duکd \xڤF4xΐj<1$Іm~- KZW>V':n 3G֕߬cE_*/[˭dr%0>ǜ4ƪqg֕t5;q,*[L05V:WPc M\i&Jʉew4q%S\LYR9EN]!,ՀV^k԰A̓zVbۢ,\նU@Tڂ˝YQdiKKR/Iq~"TЗh"sA ,jVJ/Z$G>_Z^xF5PzLULozW*Md~J%$΋ =+rrWF3J,*M){56TU5 _$.||BvX^*Mx~_Uу2JIfZѴY0T=pEwABʧ"#1/i¶;3knD:N񟟾给OŔbƧ=ˬf~"!v<IR҈]QJklۢl-ڱmvbƄԫaM媬9ȶ\nbR*ي[U[b&+]*cҢWd3U*9x`Ǜrɑ?ySJ(b{J-6gUJ%)c*̍2 /VK6&Ii`w(FXˇ.NM=grFZqdbDŽ( ~\Q4`LrէoLþ!6o75O쟫iF̕`d ~".AD#[G؃ǍVk߄hu/=%T!&_"O $ҧaTih.CAœWF,@TP 5s ֒!hINmFU'ո8jPdGh $[gM`NEXR&0+N?P?=x(Az~!S$$ŠPו >f78-žǨVUqqI*!DC?Ն>F =Q.T"\"A['&1L{7B`]>!HhQ* y=/:dqg4d3Dg咏oRR¬4+<0sD-a 2_g'o~iMzQ)3{"4tomyzle[״ZN:?uc;;jzM$JK)( Xzk_6h,BN$yS%QkiF-SABuЈt/_ \{~ITC\Ѝdh9x/9Oy>9rW i`Kp* x:PL~.4ON6p}w)ejKs*x(e޼IY-81P=`.txмTcP@{" 3@(jqgݭąЉ℠,0Imt^ Ǖ "+h&`iK w=ui:ًvJu3łKMmYXo]%";RWH 9ҖC9virQe1RR#-;]@r1]LkFrՆgXŰ8'Œs!wn9^8hOsή|)λE#HYў稬t` 1+0.MX=T㜁]˝ոB+k^҉H~Ȗ㼝pҌ#V:cFzv69X؄4)GuX@S[Yp>28 9$,b-&"{j1RN&/ՆjBAוsV φsz(x ea&Q1Di2>^)iZ#(BDsm)1jIPe,X:hHAsUI%>ZIN {Mc 4xA58n lԢQ{ʋ{0:}Ni$9&![ p ;M'Sxve Y(ü;D,1|9Q5J$4F*,)O8HDcod3bir8J x"#s[;Xx `i☼8ӸOe}姦iCvA='.@ Π߾&*4cH./xG޽}||hq}!3"bTyEE/~~E/Ɠ /S\~.L H45lZÎ۽`$}R WoB;70Omԫ(A9{'`nǔ`tyj%*eЊFc d`aTQt?G󡢦6fc|ۦ[]]^6;_ٲ)j@KnkKBDUp۽]r\ Y)!0Q*z!B":9_v)@ϕo 'crA2a=5DE1BBa9[ q.fA&DOYe6tT7CNq}KX ٞ5HSFy%55Pum=Lg$!W{w(UqG2 nX 4]|QX+bJ4UkC bqɥhV 阉8p8`GZ +Ww1)q0]¦OCx4a4gΞٛޅU?t?|o9Ravw;?&Dx}0i=#Gǻߓ\\sSG3ţ/\r׬z;k,b*,`28gx ܡ+'6\|mϹCnj-yC.>bK͆AN6$v:SۓU"Pe6=18uJ%^ B$Z,R'&N-IZׁ5K8sJ; "xogX(,L`eLcp2p008l.q+(x=.3Xr_ت(V8$ p+ u) J!W# @lV7H "JU, ΃xņo.z__,d" xOû`wL->dAq ]kzvgzY]dL@ qvD(LsMƀ1r"5A/Ψ5^1;M@ .s)+GaisX/c40.64R8PI#9p gP&_IMxLz׎ ^Ӌ3Y9hYx4]߽8˯-(Mr}q?.377H1L@/]O ~].^\]^bKcήcқɝ4OӻчCAץyq1Mco&Wap_C|N["q 8L{2 VaMd!(Q.2LDNh51ɚ^ >^N<̦ 00y(NV>U* &I(\$`wń@r^*S>Dj2m5P (dv{E; ;wHveقѕΰ膲g:;sƋT_ { 18&^‚s n3 ȏϪa$?@b|G~eVi-XmYXuSxY0EX D8Fe=%I;GF9D9вmz FOv1)' 噇A'SDEgi.^;1=Ua=< E&Ѣ/1.~72)X oF,'^Z/¿7#QZ-.5.ߞ'"!FJҦ5u(\˓Cꐔo3(ÿ-R)BࣞIO7~.Wr5U:rGZu@e3!Z*&Aqj[h,xnEԼь-5$h4&^AfԠ+<|hwG[oR˿5Tb{͂w[OcDZw^'/ 8f5'AgHIѵ)կn{JPn7LQ (*q qk[vn}SN! A Oߎ> w?19k*gmfbl{=>O'U / .k-Vp#몯>8%*rXi- |,{~VFN^wpph ):*έEj~NY)¥뼅w/|O>Tz^u`NIi:`:eŝ o|@ &(ZG/Gϯ+8 kf>1\q[\x\RZM)L273+ۋCodOZzqyK Bsd6MmܛF4F{^/qBt/~ގrC1fјa-LtEP] [1ޫ4BD%ocufR]q㕧 ~Nxt_v::DE$M(`1g`12hk FNzsR9hK8#!9S7tdIE "#y0^fϴe1f3NEñ~/+V Nj^wdvC}\Jl,hQTu;@Y Tg֤eOиװAn&-*~yjLMٙ &gG_:DKjYJΑ_G\3Kyvmuo1HRƛO7T'X3X]m1z8Y} ͨ慵 V(cA15VX0s930 2!G,`99Crix8ڵwK&4o770^&&WP3F/2`VB ?x%x#4V $jFA1%I9pLX.2l`MPh"jap?z%9^\c8a QhB̀𬡜VP"#1fc2BT( (iػFW} ES~ †d!scIj6*Zn`s osqoSL+8@}q1cf?>BDL9(   qE8h$ƌ($s)DM8_W\y^6_{_u+"u>3}*讀xIVbvaÞlRqĊJ ±7Z1Vz3úŠ ʨ Qn@X1 Tj RXT8Lg!7ᰃ)as"&Eu:C@J4muYЄ MmYZb,8=e8ؗ;r0 g~o&-^) Tp6 ' vS!itTzp>Y ] l}[TxB9ʰ4hYL8CQGSTiDĄi=fQ @X~24C TQ9(5n oq1fdhEJ{?X׫a7-%wuØu4,TIՂa FAŰr 4{۽ xuc8,k(QF6=Td|}Oq$5 &bFim9f4,a-9 W}V!oE`f;#*,G P*p:@ Wa1Dzh!Ԕ !FZA7$;hb: 5/S(i|\c }ͮA!]Iֱ?t|Z#_h?n7W7QLSv,tjIQ|22:zK Y{쾚5tАy] fwWpA/wk^).z`cz֬K^NN8/Z /mD˚dT0`Pg`ZxP8V'# <Ӟ5j%$g%R!FaKH Rf **\sKI!FZn`%f5|?KnStI-F"!<LdPZp;L V+CRx K>%(EmJ(FY!)MΊ MQB0L rT:Zvx!EGv{JY RwH,)&# Ou-Wb;aGy[rp$S`錋?&Bpe5)}#~$=_WG[0;hW"žza (SpAeA`LV;]Խ9RKWq/CЁ.plh/T}ּ.%~1pdzdC%yILT$ϑG^)&N$B^ǻ% 8=CI4||IHIwWlGc!YDz"!TٺC On>Hs=`c}1rq/e쉈/) z;]GoAOoc~Ԥ ԗ>Zr 5>|6I9B2DۋuQ\}%'`1"~"CHRt,LEy$oeJ R plc4|"D=AlAJj^ 9c9S7LRX Cf(AR)&~KAoA猝:eqRmTZ%b**Rag-YWحki붦⥨g C}Z^X $f$L3&=] ͛?V 711{tO۲ۧѵ?-2q);ݷ?~d]?.|am/ rU麟᧰Xs7ʿsV'M-N?>f(`lk`A#/}pH je>|-5@rIכjd]n0z)W]8)iiӦܮS@a4E8TB LKKiV@) +6*#q|R Î# ^ve!1fB ;eZi8a2r>_V$u(߽5|{k&KL>SKIщφkJ&lyX36\ϿtKcc:I`!'84{v6e'5Mx.X]cz^bE'ɬ9rŨe'%#= u^mA^/'Jsh͊B]tdp0lOK[K} j~(Q(ePARM0*[7q%vR>QN/>x@(s'ݗ|趜K7Μ&x)h }d†5`l Q*o<#sxvS*€+@^}rQrQ+&9<;J1nT_$BC}2}JODX1L +\fa.0/BXgrqDi%8\}E\67pHQa_vFr _s`A^.uSZ.^TNc 4q*<^|>JpSGqq˿lyq u޼֋$\ ;})0KاrrL-BzϞɒ9DNgȐ`Nɤ#]o۶W bI i74K[,S gQe\FJ&u}ZjCMY\9oVے!+.~r*~(*?Cs̟-dE?Mx r#g4Ɔ ,2J  CBJOVnPo`7bi6`̛HncG0trsQvdoҨmF o{ҳN#i1\%97{'?@cn Pz8+dqMf`c,B|Œ,ߨoVg˃3J6peϦ/knNm_F/ܘS!Wkn4νUf:5W&1p*7r^*iarF;& 1Xk""Q2aĘqH0c0F(fѻfܧ֘'uvؾj QŘ҈+"!K#CcG e+bڈr:ރڣi&6x*HVr~ouAvfH8ou0<`E@muhe\Ӛ G-Xҍ}X2$b ~UYSV*CNiKE5 ̤`Z ,}TTf8JgtwR &+qgڧ+Ē;̟8Rj{LOdQ_u?q`Nw,Zpx+UbΒ68}ǽ+ʧ8w3Tx5VHv`보5b.Fm,P;sy$dKq>^ת%!S\nQ&& ^}O 7]AIε VzՊJ\m څA+\A>>zW[q *edy^`Vvi&7$ΚR; ܗ4Tů~oeL7B`@Ιxw~MH1Zy:ptUZ{_zmkwGIVua#(Y[JcܚK5-unM~niv헾c(#,wF9ceKoɸTQ4Y1\f+79c!oo,וs%5H:fhkpǠ+coWd!C{Ìo=jPUon6* mٍPF1YÒoqD>E? ERkI|qjnrcI0bdGF8TchS*A7W$/]Q)"GQ 19xU*DRG1iѡ 24XQM$KKNuV ؛x9EƎPPE4 t ⚁ßGP0*qV>T, _xF+vQmk}'L#-m#_PY:#:Z]*63 .rs-'M^yY)Ucy@Y=G٫3Mrކ># PnDbb5 雷7Ӝ(BGyT~G>4z9oDDpP>͡l8}w߄8o /5 N ɨ׃9|Vo7.)R{)JƢYd{[TD t .;cG`}Cx\tQ"+<`#=:t)gi bKW&Z'FTrw틔ȉf9DMti]'Vk%R L*eXkhQ!!['ok%Bd!Dߏ&x93qA0(fp 3'gfDk9dJ`in:h跷l,X_KMR';S2&+]uއ;3)/6-;tӄEߟ:XdYr޻SD&H-0YRN O\-” ygwvHok,X6hL(nPŦ,ՠ 5p_41`bUGXfǚāA;舋(r!؊TE@%Tna8+蕏 ύ,m>WB{ "`vjU~I6E~aՄ!E'^_~|ugxYp© :lpRpƂLa/ xY++J1Rl4Y u"Y jw-Q“2:V/ 'rٟ=Еzi<9 ]M&hW|*[1%_3x'WV/KYdlÙ/^c1ӸwI{R < g]g`k~:{>hՊN3o?MB^m{γg{5d'T]eC`_ ,ܪTT=%2'uQtk&"s,)ThCe>;IL̞z\QD/SMr'Y/yjk@3='UH Z26 X bjB&qsMo_Oi R.sPHU"V}Ko=2TWrC$0TTrwQEV, Elk>yr粷w$rP ;u \K=8xbɬa^Dž=[ùlFT=E $^p*LRqs"˽9SN{xH=Q\4eLg PHrJ,rܤv*ЯI/ʰFAtm:\Y<'PscEJh&s4~BhO%45蚰it§U!yǍoskb0~,wjW8#)t\iw)=y0ϡ:/=t1V{|s%X`S^M" or^ *r+@״Omz=Is4h7L/_F;π !Zg9?&_688z( FEJ`TVw6|Cvyj5ᲾnĈf ܄>|tdҘ`:rH{|H2:(Kk i OC9 ~;1/bVjۀ9Bu"mJqϥrapWMzO{(v1i=U6[QA EO/,|ȅplX`>U(A6ʴ 1Z񜁱3X|`:69 Դ&: qgw6s{PΉt>'r ~|qV  kpPa:`zW&\ϹR)IODJѧڌ0v4?\5Vd{ԙ'n_n˼l."rS%ѸuI`f6w);+/9B]h"[/5%X+AJ2nV3GQV2iEqnQ) $  C7P<_QmDˎ)qE'Kh'/g@V}HQp1cJ# x#E*2+T0~.DgZA;E(vYcs9V-hTZ݀yf\-(q5K=\dEJt)P׊\`PT ;W :I8ۯ]Rduf ;-ף44 )ɵ}}IFRD@q%RC;f+Bar3J !K0J欢\>94.(rs2KiM NQkhH~ءL} => }?. 0O6i >Ԩ Jػ~? NbdÑx>PMӛ3FB,nΥDoޞ d ˯wPB".\D9a(g<\/ Gqxhi GՐɐJʍ:.0Xaa oP-ӡx&$pgq%e jlȂBp3CS rù#om8LݥH 4 Ol< B*hef&Dj4 *s !yHn2[.NE[ @Iqno\:n{&}^<6(kcT grX?rx GӐNBk?/8>vɏ㑴C"!DҌ焠jL5Z9BY\}BtYޠ0׎:SR*4iisam#)MmR*A 5,|YJ C&qvT!!XSh\;Z8YRf 6ҡDK|DP*0QqzW; xejM%58,CfÃ>g@ RFq:oܩ{O\Q Cc)+E̛3q1n_S׵5}Ĵxug? or5_HIQ9H^˧#%qc-j<&W{v2M/? ^U7ōn |g".V  N 5UQObkwGѥPQPhP~wóSn$be]yot<*UJr6R;BU>p~=}U?tIPM&2{zT*^z?ةUp(Wy/EEOh_c˿M8 2v@Ed$ rH2/ΚT(StsY \Jf,e^[klns9P`kIjʄ =)O!'94?Ca8dvkxo}LzY |>I oٻՙQQP_eiz3!7Ey+y6uP+ ,~#?>of~0 ( VS-2.$GNY0<FѴRCl~HW//EU0#C+X'*x,f`jQ 7M'~{H .j,1[pOiS^aies+92d Y38E{ k^xko&Wh׼ VpdJg f ,x16\$c߾Ԛ~%􄟢5fH6!-k̒Vgdfk' eĮLN) T6_ &W՚#J&ڲh:Δ*߇m!=>S!7&ߏv{y:wFW'a20Gr$J ayjrEM?oBbYIJ65-Z"±/iw|"E7(Fv% Fc.Ӿa(R%$x'v(.uBvv`K7YԉbyKo1 Fi#iEj6u5Ӈ65 ;ԛjuҋOI'#'1}u2԰n#*:޸%8=viXV-[;;)vݢ[<O*Oy0h>/EKL*7'b N}=5^tO|WeX f>Zn @{:. 1~!֤ վ7OR']'t*4l'"Qc\6ɈSR3Ӵs%!ej"kkG|97 :T]ukRN>ܕ:Q $22xjÀ)(e& wly ;kgB7NAv@'k>\Ǘ}z;qS[:0$]tMʨLuSԣNB#i >_jvxNM'nt|Ⱦ,P Of QN.xS~wEu3<8[7Ų "~*RB2.ϞZ0\ `=K|}+l[f+SlN=I{}̬IkvNFZ)սJ3;3_:,Qgi^SDҼ*%F%J_A9,hi/ի]Rs ujK iMuGRxT c~t}{W6^VI ,OR"S$g 8ΝSkǹl;&ќgR$`֙,48ύ#)R FYҭ,-_]lBarq(9!/[X2$4ɟ7 σnƎa.#v_?qAP 9ex$*CLfEoF7sswqS Doޞ d`_GwqTon.~2S>\/hru<1"x-9.P0XaaCD n$džJ8(?z'x7 CEilp۩_ar-l/{K ꙼N&3䟏,t!UJ4E3)?y^9>ܣ}=_v5ޝyX!R%΁Sʧ+~߳هẐ[z{=B< c/xBTc:~'8 ec F #ŬoΫ$H8Pij<)ҖN94`cR{AJ1 D;.cS(vN-qȂ+5 QdXQ{ ;z<qs]Ym%w9:R^$?L`1=b{d^/(f7dYS3J@ L&{ǗE~4 grj*Z1"-;]u@{ >~ nəBh]_znQ_qgsWaPNJ}X?LiMNo<|\`᣿Y:AӾ%݋]+.4%J/AY^Xe5عF%, L8Nj4XaHI&Hx>Fa7˜[ۮ\z00 17uz Vv Jta0!DRYv%m*462WyGY1ܙeéRp{5|\"$UA ]Ǵ➩7| υ(@~cXL6-N]p&GŔv )8CZ/K΀oe~P*̎bcY*O:)scmA:;;S~L#R崗 9aX݁ygnRO6JϼR<ڛkIh=aMmW6R?2J0vH'Hoq0JrobbX)5Hpꊕ-evͰ+w=8\طuePm&%u45 V +nwр!IkuEo(;9&8D8j0j(pi~zÏNn}~ BEljSHRp^L4\YhZSFs/e2sn_y™| wCJ={HclZ{(;Uw?{ht|N87:xjq҈A@b{ x̣P2UZEyrQk:yQ\\^-u}EGӨ$R2K$%}zmp @_Ս#i[[K@93``}INHד.AԹP8#h/Sz/\w-O[ؚM1=wPBs9x,C^ωLuW=v}cҔbtHes&tRDi]<\K]9[|ܸ0*jLӾ.]gFE RȔ,޿yZ( 8KpHMg/,^ $2⠬\JXMҾ tN.|J0n9qZ0U H$wšBPH!LzMrZ(YC\H;lBYHk $ݪ;\zwf',5g4ƶnbށR-+膤#@AX)^ *"`5;c2K̅"љ&5S t(wY†+z_,&hac#s5SP2*^P  bk0@N7 Hj$( DN͛χ('. Q[. p9U_hO(!od>`[_'*ͧ 9OKGޠIXNM_=\X"DO|J|*{;ByF@Icc*Q` Sfi#*|# ֦-c7żL>& /? Es~fÏcɘ@gݣ\aY[W>h0\{j60s"Rƽ>´MǓwΰ5BO ה6Ь@?Nz*7+- [ИmTz^CAZZ@Z3º8ͧ{7e†OIa]m IeXA+C}Z5kC_ 쌂mAxVZX >8Nx<h/ F/#!LYn`/6lcX46K`hgAbUGPr<xNcO9ytĥ͞3 #+8_ߪ&Y:T?YX?C`ҵLNӯہv7 2(Bxu+M(lǑ"w~ͼ3Qd A;43: Q$g &b< 82QX1*˲6RjRS*L 37CÕ 2"`(9` ??KcjӱԜo];ź)Sƭqa+6?|,>硻}Q|huGBy)Q"-iH7mm qEkQbYonwUp?;7d]1CΨBKn*եy#FZސ~ocT:4P;y,pjruoQMG}_˲◃3x>9{o=y`xp`([Z05jŷ幚}F[Ǻ|'3fjyYqaHtBBpɠ WZ=~pf==xhQxи^PÞld;4 BeiҸd;XK`An9KռwcX(ŕ(a{d,[e`kAn53W:SrMhm,K3RT%eS4FL+ fFd#{g!#=\괙=_ZE.=?Lj#n1FƭDp<4݈=\֡A̓\4Fr "76vIk! d ,dtUЀ){bP[ )]1NP:S3&Iz8NnB,u'}޶`gCy]=.]!t (H.isD+bsp[?g ~**oKo LuN4rF16NDmyL6 8I>j%7۰;TYPz9D(5) knӆj]zOu*FNբ>?M `H3n`97  #*2p[5 ThpD1`m+mgX81&m;w^\8߶C^J͖_'KDΨGB$lH+mVlC2JѪg$y -Q9';~ew\\Tlw}~&^q_?F2qan_ `<}ѹym26>x)~ `Di_:1#[4mmxV40ԭءIKv F!!Jg$NŒVR 3~[ ^-w"׼xY#7a1)ե,E)`uXX/>Ϊ/BIM%\] 싫 D//Qi%S%8>_d%Lˋ/%̎ss:b=l̔725Ra*I%NwŁ .a9c>t׵fg?:oFBZqFBƤGHU3m^fpå1aLX˯ON F'|fQT8qb .wd+0ƸlЗxg:?Ɯ9w&q!N;UB2mKNcOĦIȘ hA$` ,(Xmk(iRTT;XUK7`b ڻl~ys\ GmFWf~5W롯H*mO>оɧ<\RF'E-C4-SE nd8s+md :gK sIӖ"1r؃~'/4Kh 5q2\)>X"8l \rLfo InJ^Z_&N?5BnH,O`Zqwn*0mX!w/,?Ҩs7Yx4eņUKrtqc<5̬7x5zS5RM*Wz3=1e_ bO=¦~^<*L4뽛ѧ9 s]@\f2056Ɓ̳1LE)w%FW_]]= 322@`(e @i6ڊ CSr5"'/JV*@N)欕!+py24J(Q(v_CcT«] >$>++9E΀e4\(3yF_sQ2Þc-NAĨdO\h$Bg*W*Eh5Ϡtc͡uvTW4.rlL5״k:84bOLF\ [º[+zfpb뎸Ն]ɹ{b98JeΥ&h%ecR>і9Dz,WXNDao҉Ҙ;f+8w0  $ "!b1| )h`\m<[N )^9282EM^kR)w(V$~#ILF0ńln1jk"jIse2sV'IDatDkʇ: h9;JOu%ppPj"r_"';j>^5CR{1JWri5_<ذXhZͻ5'`fElAcL P?&${ɕؗu4ԧ v>f oa~S!> ^$rsa#MZ SNs\?Q.ap+- ݰaeXv24mߟA$+4R]AO}uyJMދ̯^=ik'ԽȜ"9 7-)/oix 3xg7rJAA8x,nf*X9dT: ,#Igcb@$-0V:3݌_8Jc݇G^m&!s!ƝW2W񔩤!S83 c J8)Cp7`0yYO`J$V::@9l8%>71%VX⌶֫Y3 ~q ^0`Asӵ Ђo3h|"8~HC5^igRdzԠHP &`:Xl#5D͸+)P-3YL"Sģa l#=FLMŴL` #7dsP,h$V0ɍiX+!d52jL<3WY!20uE/"UE&Sӵ6@ӄ8"HQq .XB[S{G7A\屔iɏ|Dq9Fh㗊FrjEK'RՒNX5Z2~ߟFX]DKu]!ߎ'+v%X45=ĚDb1qE?ۚb ـ=gu\ppPd\s(".:@񖒇;ˏt*'zHdaϱwf{Dq2'`WzJ2+kgjPs1^fT "QN)ymW 6M w{LA0wwYJON[,q_J :>7&5ĕ6/ś?ˑY,?MotywYA AšΖٵ[Yc pjӐ9'P գOŮfOz >9 n/_g`Ȇ/aE@9^7QŖ=*t E9@2)#J0kO>mش`Pd8skeE4jKB'39/ vix҂e!A7I"nB$uFobaM<_(nF3:./C&TjxxxJ%AUGGk)fG&4H}1u@ď—r8ѼPg/&X [D7M0nqͲ[C\hҞOMP6QhtPKXV,WBD]mc#(ԫG 8<#ɐz$vJ(}-jňֽ&a淟W?ųpk5ϯ/jrmC_EETj=vkZ|ٳ}OyBYVJ+Tu;#/u_ی1{9mD|0D9\QͩǔշQ鼾qCv&9K^W8՜sw?|xFKO\DLAWi0ݎ$8so~ԙ>Xx6y#vw΃٧AzeJ3yVnւrޡ!b$B{IXy{1 y5s6¬Pgcqb2 #[BbhMLF\p2A9-'ԂRϯXcHN\2a=z\s?~*ˤ^+%6l=2J:Z/td8p44>`W5].ҝ]CALֿ"=^t|h,B;ƩӢ.GnԤ|F.`kK30.v>6G:2v!bRK/dXriXjl95T#p4sd fRJ<$Iqwޗ_ }Pc*YXri,,w*%ȰM kNm[KUPbw{l}4:nlR]araLz|abKL4jHۘ~ǬO}|it'tꔴ^vw(P ;fPj핇Ȋ m~D֜Vz/QkIegv<~tk)0fnHNm8G \4#%(<13b|VBsYєzZlvwB!=4 `ej]ߪ"xO3:dK%I7mqƛBK^zW&)/^+Y@#W]!Y!@gR]ZȣֻY = I(k:c>YԋW%c'yAGR~, z/$Jg!ڔ*=ħ1V>Yu~Q,9c.ܪIB+TLO/Zhr;0HwMH˵VEWM<[-CjuBQW1;Lwxaegj{7_eq{w 0}=b`bokϖI%٦I50ˮ @>#{#:#qHUk24JHp윗-CH'Fg*5ȃt4Lk:x+VB96hagH&WzMclɵeCn]~e4/sUh] 2*.JP `jiH%X#jA Э Zi} GM%ӽ'SKP㜩|]I`ZD9l$7ynx/yy]xh>sNdӥ'WNR2A(Ra էTa75TC4-H:;BN1| |N11ܲuө=Ight!C~ OQڑZǪ"~|z*z,KE茿9'TÔ'@ي$7UN!I 19Th:-$m{`Y \h`J*wKj.U Q".[*q)U*#*)GZSJ(D)UE,.pdJ 0ˊ?٤HDiUx!(e1JKJ܁hdb0JYT, KE aorOH-I|O'Ep Y?;XYXk2ٕEP-x SHAbh=> B93 >\Վ+2^=yr8 LhM4Guظxdz*P'ۧQ{jL^G7/գL4}_?> /BC؇>\>e|wa ;xѯpt6_{O>OinSέZ9:߹_On-i}8*yw u$Zj}M}q۠ (R7=RB[ ԆM;UPjCyS;IaC2Qɟp:Dze'3]5ZVlǝqgTϞK)U&*'DsudƧ[|q^)4C~U)R9R kkw4)WwjSШBPTI%1ĽӛUk3FCF]:UjYdo4%%ǺԵ@Cq~nԨQŎ{ J)\f$.h) <=hļwN?39( Uv.c'ւ4n׀$o'8^Ԗ'ꬔGgA+5!ް\\ ݡxdaT(Va>c zuN vI;2PN&ỹ%̺"`XA)o-u8ϭ3ԄT3ho ZUm"ԏ3-<&~ 3@Ő]ASif k}FTwӸ.>̿,wܣ%t2ǃM(z̜ jfκ8Pf)FK2u4a"q:ҪFr:EvFRKZ~0-XΡ}Ou \z511mwt삐ɃNtdK:q׮t[Y AS8l?H2X-x"qNS<8@(i0ΣF.XmKo&'* SH=-U|%Ɠ^g-M1&Z Z>&`ߡ49#8! MNk-)QdW}"44g'^hoGi~"Srq qֵ&;; .,,[hxβƠ5sY?^yF/*5l2nHO õ+DkS>҈0'7}=Rsnu3Eĩ-*,,[2yZ|LtɟB5Lweo' WNI߲.G2 'Vq9:RwU Ӳb0t%A{TK[՚ޕϫ>tFh D>ʫ;x;n TuA$jSI%|=/?e#Tgy5C:|sԴgx gu]499e/c1m'zI-B5Y^8n I3&rE+97تԈ:Uo%SC+@;%SG1Eݍa ׎{ |UE$t9ɬm]r66,Ӵ5apvI ӷs=]~?)3՘M-wPT.5gm\Gi0v'%?%X߉m[*)iLj6 ŪeW0`NpIֆ77JJҷA%a}Dܚv {^C !L_R r},Eu?]P Q JA0<(e(*hC\Z%XÈT;cL^3 z)w+)a%{DǚJJw+ULmybj߇W)haLI.I;I~-ld ?Umj?VD EpNZUU2gRʕ"=?ч߬ߢW 7z7yA0$ib戋-|q&Vnu<C=S#@gc Q9+bj[e<8CeӢ;)' haׅ"F{^ xY g!ਬ b4@a`6:pFhIϸ>+~ɸv]|!HZ#%_ʼn/}ZGxA-?QNd)xrb#i\^BYܿ{I Y=TGx"_&e9x4BuϨV NVUpN fGh0ɠ8!x&̌ |\"܋5SNWWO%r'E5VjJ &\`od/ !H "?yI} CVr60yhE HwÙTHAʨE̛h{C|Øiϯ3 ;l7M{6\4s%xx<L 6|K$DIeF8!*DG~&^ES op9LKr>~\V-WW2wئ-~|{?G-c?> 91 0[:_m_1TAhr}G_~BS1Wqa}_ ۣQ8MI@]nl&2mj Iz-5(qaݯ@dc>QÀlc]qe0}cWݛ=o>u1'ƭձ[;:rcpk,2SXMS;lCTc61vlݝ3p;f9JpF/NVշqgpW$JK aVb7XmyGx O=4Bx;lIk#rVQť@&>%1ms÷-J_=7W>}EAw,B`Lhws2* T-M)٤M)٤_N6)w"((!*%:P1Z\^P) !yl1VJ~TVJB#Ta'Q3M7w6뤃2Z"N\p;aqPy+*A:<$ XI8`Rݧpw^`n*TsQ 6ͪtU쪧l(LNsSc|DL5":hVZY˰N5HǮhEKt8YRYJ`.RjvJ+ A>}w?t"LJӇu(+UsO0za3haqdŋlw[ SR|p7;KD}bwO&/iøvzm]ouK:daopYm.Y/y:-LLZUv4AKQl LND.?LZ jPgn CL)kpkUHՊv-a-mNF)]$Tՠ2u<,]Z1jC{TQ[o}EWUp `cIehJ9z5HsBϠP话:]Y݃ɓݧO}> t5mHUű zuks$O)o7mҔ,dW-ھsXRYJjqT͸.%,قuĺ~b]?_f],A0Y:z+TNvPdG"n@q4q)@duufIvrZjh'rtI ,8S-EJz` R‘hB l MWHpSa8DN_K~\Cf'S5)л?NO}~c/,z`o^Rx%RK8|Te#Of?^M}z97U &&<8La.Z.[+"wަGOk}D9Z#T%dQXܙ/ah)K??_7&qX4t9 j&u-EGmwa*#\"rҨ)8Tl<}Zt)a=b0G2 LfN_ [$WK|_7hXc]ddry}e}TMdJo/V~h;^7I*ȡ];Q'ZR?SsaYp`JKe2bKG=7H0,h$tKoyR - .8CD{8dCNcV].UeןoJ8QQ𝝄 ) ~˧!`\@>_\*0>.{p`ĻWQ \[ ~wՋq.4+Bo_6;s>=Fp"p m Tc$UoS.9x&\=CZV ؛8/Q"+@.%@R'[O_4 L{%Vx#B(>QQc&QwZiu_)X3C+Hqr_]7̈́"RWZx&3<(D-2P lV,ZyĠy-$+F87ui}rĹ%Rdgv<(`(rT>b+I:>eg7WٙӞ遶kId:i9x$=EoWښm)WVyc@8($x̢w yjPz dJĎ^R<Ӈ}CdL25+NlE"}7/Kj@I1vwv WNya!jM%i茠jg>TLW_b yy)b F`(B~3R"[@(uxeJUF‰lRE`9kUowgsHQu'qzV )m8:%.Ig1!6)- ,GL@`Xa|[8B^,|41~ YEg3}mg@|:}{x*Uœᓕ@ Vi4픆]ӾOr~det7zMo|~pc-)\@mƒ L ,uhNMt8٩!nLrdb4nR0WoԢRoy *0G~ڸ-˔H>=v1`-2-s \DG t,XTІ"@)b{D[;5[4?^5;?9?%*7;8[)ɺ2{@y(LʂYN,V%ٙeO)͟i'^A|X~uYéqZ-jV. $绻q- Ӽ+W` W`dow똯#wTH ɢ/։D4chǭJI? [?6$Y. \zVhlYa89>eRTΐ[L0N R놋i,,嶇-gڛfTי v@g~;f 25;zf5ܖbgTj¦Zq +rߖom+L:Bjdw{{rLmFHӋ.0BxNQYU,m C )̐`qƆa*78M$On)}s |x:`9-pG%kH"O&һxaty: ~k%ŨW"^Fw.}]IZfBԧe>]C'a*gze#7y3iLE:nLл?NO3yͳ f>MRYՊncט :<z8^F.D0%G2XH{m8#I_&3 >%:+$~X9u5lf5qըFK)]eVUYbBMRB]'4.%̠ [B29T j0pr 7 $A2 .dŒ7 .y$)a}LPPE"!&^i-y *q;fHsʔk>IS2ua&)?]Mr Bӟ/Uj:n"^M븉x]DGD`8A?P0Xa ̠6 PKIc(?}Q RM ]vd{aZ`Ec VX!K ıCq"2]%4Aq̛):S>C̀<|f t}Vуý`/&Nn8@LJy.>cXD6R-pM໻wW_}>x5~-~⳿ Ol02[JTҌZVd/vGQi i{6^^n뜋hjEDs(f*,-ˁ=|Ǜ:xapٻLp*fй DRΚٽ i !2הrUp Kc[;yVT,[Rx$a"}$C0qIZh x?]$ugGTfIjfďR4"-f2lo`wH#JDyQR`6sET:N>GY7|b1S!p ` /uЬzDҸPa,ǘA}f0(JqK+QSJImt_53ʅ8J0É/L:<}&a<5%{l3]?| To- B;ssKd=!(XM@7`?V D9yVrpW T0B5Eb/7gxe<WǑ1,U<ēpEz-Φnwp̿B؃|rmecr Oc1He|Y#pfX&O q۵kSZO@NV< R|h!=cFMWXِE4KX-vӸ-}Gv: xZm4ٶv&4U5!!r͒){qnk7Y[.):Qf;MhvkBB"zL ɥh=!FzӇ+;KO:AFfd@(n/=dntHK͖JHJ^K ޗ.5%ah4_Br%&`Mv.uprRbt{M%usTqv!ZwRI6 I@Z>sZ;ݟ@gOk\s\7lL;(4֝qmg?JcgX(q,*ޡ~LZ=XA i3!ve@WE5wݫ5Vr$ ֜ z}}ֶZ"$7 XMppI8Rm}`[&t,<%´K& 5 y L}$o.FwwOU-;'st1ss4NQxZqn2G;WGy"G9Lr@຦;XJU|N[G4N{m ѻB|vHP6ie-(*Ttit0lbs6uYR!v%lyo(S I^IlpE((8惂.( $^i-yB1W6ZlrB*6}gҸh9 K0xVZTi+:2$< Q;EF6/ߏ@#lPgY*E{a~0:{ޝ//.0K%xx>#3`}DY.7#4Xůډx|X UIުr`ۨ׮VkX$QaClȈ ˸v;0{3g]&n'^+jIKI<{[POE-ao͝YP?*ݙ u^ so%Di7p ܁(SC$%aBpcE0`!FqKگ*wZ^) ~цEm&3Zo-op JiTSaozT,TW Ƀ} (Ql]aĺBLe 9FRp'cC<ua7BVI%vYkk4܁GC5ZD&j3R?Te0Re.AsO4ifdJ8p&.*7~zF7we%Xee#bfbkt.kND69l>8e45SIJB,I#j;XTzNJvUAJd%,9]'%h]r@sj7-|(=p ;]Ei)ݭ5XtR٥P6*s¹D`eM(bW,b/TDL\Jz QsK _Z*y6mr_EoD 뉕X4ȱ,Uot-R#Om۷AZLVB:&^fU;>7/MlO3|YngWSB;Y\~0up?Y0O%Gw[9LGx1#_WTAеYƅR8cЀW>|'!/gcW/Jה喺\pւCg84ˤV"tǫKuU %ˣ5F%i;>#aǜŧs^Ud9B  V;%C (Pcڷ `Ă/qZ^_˃m-|+5ބO *}SH,t4C(ƕLff7.t:BD>|7ˋ"Ou0%MaB(4EJŹ0FQA$~nL_pVqR +RAͿlNJ9E&,oWt@NaU<-]\^Ef\A3˕x?6Zq_Dž~uu/d6RJy( ۂ7L9̑Y5sI j8.DzޔD\^Nʉ_ ׀fk(/yh_!:Q9eB2J}¼RBEpS!"Kx܅WG![۰T*N[Ťnvĉ[X#TLZ iuCWƄb8&%7b,./R24>߃3qp`1V.cCa-[$ rL!H*" z kIc83 y080I #2\m R-!PSGײ,x 250dBeLp- 1a*|?sE}z|vQIY* [Xq03"=V8P^S!(ggͿ]4,%z`<+yLb A,+wC%HToZĀqZj݌O~C\'aCC`ڂoA5ZʡV!+,n;!Ǖ2|}jYdYYT 'M|pD3pT\`iʠ:zP֤kC׵| \ ᛤ[D@E5h6F=?:@F~8qɁZVExFC,?xHwwf}wu\q^e{]]Ո6x"`!b" Ksjrc)-AJ1l2$8˻r@b =;qC {,kY:[gG+J YeA5X[I Zٰm.Dߍ@$\SmDGOrd-lp{?8E +]|r6|gMhn zIn Sb<, =5䬽FM+>C G'!{/=]|pLÏvܨV~xqR(u3wEyGN:b{=XK ==R|o_-nQu@U3} ӈEէYU 戀d/61wǦ. BLuذ&bt IA/l,H ),Ei MDIv /y$٭sa ȒO,V벞ڻUDgͻo #HQlN[62pͣsp c$ 7s6$8+SxD 3F 'j V3f o|Ɇzc[bBP(1;5պfbF*YzyfhcJXUf ;Q]|׆SuR`?C:ajDKuxс, v&di:LDd Y6:k\NI>a]mwfxOmbv Ma"Py'e Qǂ#m,h,OzaT6C8 czUv&gW3 sS0g7WX!9=#uM#XVyn/?y$FɹⓀVMIsztsql,'݊Rt}6T3>җ)Y]Lm=-$A[ ɴHs/S:y##Hd03%v:LXO"(a8hnCrk,FV'jmY{wZs!ňŪwRI%b"XtيFDru!,nmu~dX|kk]k=#Җk"Wrrfih-_k _x\ľ Ĝs5E͈|29gk-ʵ#Z %X۹﯃ͅ>ߞ}ZWߝb\po~Mx7OsZ Qj ԕ?C*kmq:^B-$adtCa@Y9$q%L%O׼wTOܞ;VC1h`9(*ҡ&osp2NғbosRo608QTr,l ܳ $YM UjQVH# rxզ-W 3r[mF“~9ѥ*}$fp}_ᮼe +Cc|*y乯Z!H葲[o/z?/|?/zxS{I@o0U$,d\\OF_ZpYm^*!ȡ*!a$yU9Ff&ۙ!=K*F>-?7:Ö saӕP`w, cZȃlAq3ZTOU{_;\aF]ٕ~]ٕ~t7`7hMյuAH\yBqt B!]'Ԙ;8l<{c9YRp(%JvJZveJ6hI2߯K|? WYMѲFhQiӤ*iXڋj"Jλ`ǜ-wZ ;/w;:ކDbF`I4$ MV7zT2[T6$;JQ{u%K;NG mUTJ$xM#AMDd50''P<eN2 m8l.0!G! נҠ 1$ sF 3Fc\ Z-iZ{{Jfc-x,$ͨN,IPPE JZ@$VВh JNL*`̤#Dt.#K!ꠥ=77iVSF\ny<~t?!~,u%eYXxz#Hh\v/J!rnEyd.)dCpi!$FD8Yiɒ+]Fmc 2V(6nڲ[zZ?Pw};G[ p P.IR,i.JIcov5 4#fԪd]^/f8N+8ͥgוD9b f[l_!xMmn,K0|!OmH"e9ҡƆ 2z kQF k!KDz>Lќ8cǦ8wǦV36o3}y7Hxe r,'5aZ&}RjHp gti5{{wWym&y# "pPC0%" ֮q ZqWP7cJSSg0|ey[u&џ2Uu}.C {_֚8j-+1\Qn:.1-{=vxdoij֧ۭ1̈䯲B4n`Ekwye/t+o&|5jz߬kpӒ;rҝ] d{4+yfDˆ~w)gAc8Dr0ba6̈́c~wuo@so麷bM!,7-u*ԕe3AVz>h=I֍{{&1\l+bkע{hέ0 &bj.B. ƓǑ 6j 5n{/1lz!W= Kc+lZ5 gQ('Zj2x T$a]1."9J"ƀ9q{H{DTi x?Ʈ$`!H`OqS୶N5Y&W`j+U-lKv0Hy׌Ia5˿nPXe۷Z J)6dR&OOjfPdrp,cG=-=n+]fFIdnc9{|X"Rp1 J|{Q}wh.ľ6 /Iܘa>Wp̵)DKPU*UUܙdRbհ$JFjI#/zW^s6qZ p["lu>Oރ۸B֔yCAc`UC+l@"3l R{QCG-e0_0ֆZ'^4aIɶZVX+/AlY^(9Ey>h|\B^",NA+>Ąʧ-t?&t(6yh.2G1iRW%\ÕNjj9?¦p\q.a # _uezI ce-81}]5On<8} t܌rdHӈ~8*sm5ʡ:sc3N2ϤY}]gX1T2o#l6CѬHhK}VWExG?ܕ!8!-찗bx[Y.8a30O`dPHF R؋Hi}r>ZѶפwwrY |# -Qк^6"jbG'P\Uِ5EJd'U6NLo:B7)wfhMn)DD ҡJV,mq60/t3 ys)2_lBd=!&+*gB0ch\JSv([:CaxFeXp %?c%0ǞHI#4IkgO?.J d_SsZ+-H(MuY~k村FЍa|'a69!f<2N2BO_cmďM em*j^ /EV9[-+۽92b qbU?p?n4vG6f]4D.IzHkgw0 ü y<8<TF>f!ey~%ooό"mc%a6P/3̃)LG}뛑5Y?DY3RQʌd.WG6ԃu7i#W>pIi/LI6N:_=tgzfv0 sa}Z4[ZLJjGUe N1/17( P1R d.k}ۑ#rq4+tCTɆoqs:߄RW>!*ZN '|`+&dVxl]8\.#[pj&)}ւZU9N(g醚NΑӉL3oقBnPΦ5*4F%enP*ү|mޠ ('bd*4&yݗ.q zҕto}_~26p(?KOOVC|+^ZXC$0UW6AItFQQw<3!iHrT'3`k_WV8NLI5\2rU:.zzpkc^'Q©fz** '^;-jH +F&&QC5Ө%M}tZLi֦|a=}_bUR8+ L0dtoIKO!UnK6z)1IN{C;=_e'N{%0 -æ껇!MĐR?4xᱥة;nI`qp39mw@ 5P"R~¢1Nw~ _ߤiK+>HD "9H2+w`=8*e8x{CG24o{oŻGoP@%-`Sw%BQ] ns;.fBQ~X&yzuUhkӧ&'KE ByŻ?iyx#G\?#owC|@>ţw|Ҥ~q֏)FZqIۃ?cFИ#ciq,4PxX1ҋbDD!|"M(#Rn j4`5EhcI Qˆ*۝_vnlҭe |m2e <zh>zr[Sf ɤ3'`8 *$%B*Nh5lmh;7j[%x$ݸba*ފLei$EPKu=5伶Zy$Tܫ:r{)n+E<6u)rDz*(~]I+# kF&Zv&tж4ԎhFE[vk-&$;2FHh:] j+,Tk YFocyk+2#7O K/%B:m8~|Frg~8x|^<=/û hڝ}MF\\]5; u>4kv~A!{y%3Y 0iyts~x^rOO.[K47u:v5xdU5X~ É|"}xkJf?Z6smO΃Ot.=Ņ<}Wɬf3mg0~[x,Vqi,`"NqzbEB55{I̟p0*=t G [:@v8EdK6!:At4ECA."J'㲌Y-2(uDx'4!q2:cJ}zD s (,?ǯ~Cuy&͵$CpJkMinE{-eMꐲ6*eմOC;-E +7@cA( 3`2hR:'R}SB,w8SHa5qJc gEgeE ;c剘ZY[*kmo Di^a"e%]j0 Zƒƨ"bXg * Z0Zt}ݑdƲUdn=ㇿôkQcG:SP E<1g k?۔{dB}z}0Se00-`9 *x>^û?yɗ,/I [݈7Oƛz(anUoקjXCH>+t!T~ьG44eg񧜕>0frz3 wrVic]eg=ٗrJPZY4{ߦB2x0s<ϯx)?1!ovp-O,A.~mS<.lj$+&ܣ4.|<q$ZgWtoS 7JK܎xa YFB بruz$ǔw,m~?Glc7<=۷_CR1ƅC0;&q6yYu3d=zIoWo8k;أ sqHJ*[&4>$;B_]HuzY@:TesA_|~*ρ_",&ui< ,% ms|u^Ŝ!/ags!x[i~oҞ=c-)}T ~!:j8br=eUkZZ*5qj=%mJhEݕl)ٿLyfr%)*hIZ&Z1Uj'sHS&Ψu"%YbK_^!_Wϥfm p.չq%+#ѻ8XqCpBIO \;R4"8JU:,T *'Z.-UyNl:[#&\1|(ٺoث)85Jӊf/%goӚ4Pqt;UA|WfP3|O7İrEoVc7fX<&Ck5Cg}`B<@!Sc%7qVb,ݪ~bo WHx~ZqPE'#y}?tnm-z}!ai(ә`qBFɘwybKG3/gu, )-c^1m?tO[<Ϣh|4kZcw~Mk<|Dec᪯2)%5>- 6µW`` Sqy:2- KB Ac.-P k)ɨj\p UN-]p-t%&eJtPD*pƃ:sd"3d DKX͑T>0MDb ׭8ѕ٘bd>xv &GPQc BŔ[4z0o!Qei)=LRr! vZ#mMd^(KEG,P rnjI3z}u`:YL4 D#Iy-"b<$7ӆlޠ0.qOWjӥ3hb \.U\u>K"WŨxgŸ:^2wXe B.rޯZPK駸"w+WUVh0)@I p_]YoG+\*/ @fl8v^&PU%km8YH-A,:[%7S#Jwqy B*-07 = ޏ=lx!f0Y[qs96)0+*TzLlA^Y*JJEB+RqL=*1*(|HQRZ04̴8A|:4ڝZf(bifpšHDr ђ\(a}LZ6,wN6dd&nqd|& )quD ]9*&$.1ǕQT=Dp}3e 2l`NSap3ġjQ^ V y"!IY׹ <#,Oi$Tqg, g9mʃ:B /%Li˩ ?ad)-tԁ_ZP CUʪ}Ъt4\ >^qzsX7-t\E:TMG'ޅ@͚/*gFOpS',0!&?SkBSM ) OI{4h)EI c=|r<uλ/;1Q iq.{0V3+ޛr~(:$i9Ig+eAu A~rɀ$lNl7*囿rXTUM**bz39  @0Or_<^OWT_뷣Y,G=aS_\sOw-D(צt![$],  PW*1&ߟ=4ܩjA@wխ SfWTE'S+Α>hj6d9JR{PngfxE%&5^-y;8q%;rCcʵ .w?9;@h^hIUeoBulO#-Bi_^=EcU헟}W )(`x=qG%AW""fǝ,Cg#ߓl{7>B\ǥOk>,p1̩\{tuf !E2妠XeBH1 #Rۏ#X{7_Nj>/fߎBlN \ZI t-$]kJ&^1aJhuwɓWwNhjX 9S^1;B5 i}e$kNUDkxNU:$皒Te 5DTgl~iQ|iwiՕ"mS6γ_RF̻j3[_ta6j<<4%Q4_vy_F(Br=B۴3ź6c\]v_IHo)ߏ{$]3aC &fXpKԐBn CdVT["ĻU f~lQR ;1h ˟5^%f=7_/k$Ύ;dn~GwOJ$_d3쎰Y&siizp٘ڦy܊;hhb(Q=Z?1OvܒNӲG>I|FvN>UGڥߞm%["aw{QA!PR:]W||"Z: bQ2Y9QNPu'v]۵)*0bH6g# ;s O%frՌG#88\ t"7PkS_nQoO1jܠVAP}xDq_]oFBGp>Mhz䘌*61J}DA5`w_&"ZJFO{(ugdK_JflwgZH" ##:0]U - gC BKpb%R2,n*1<t!A`Z䈱Dj,-- PUZY3Ѣkɲ.wW7jMӜU(ck!M>@\ʁ |8P{Dd+ v\OvVc*+3@x.s2Pu,e""X n,(q`O8zoJ,0Ŝ#C3\kQ:*tP\t>.62V=DiZK*o2>B;"7dce?yT3ٕ6狁@Q 9T75M1z61cXuRIL*.vsЎ^ 3^__.Dx=VgсL:t-Ĉ7@ˬp6ED%Ee0U-!) fL-U\[bD !~#KЛM7[j\D>&^G"sz-DbH xpWX]iZ ^f5l  Qc C̓L&\e5~88qb{aQ'ݚzFT5Ye"á-rVGgo.0 Tx[XhXꑝ/: _r}NرDяSr#ʦ=^O\ه+뺺M9^3z;F?>tL7VzkoNÌcR |44ُG9Մ Q^ B4ht^l `tBDh&m7Hs=XU_ry J}:LJf Q7E6^oB{{.A\E6svc%o6nJbĜ8Rqwo~4YH<$ /9"DIdUb5 δ.1Ǖ9BEvPjJn45wQYjGr Yzr)aZxIю^C1+E%yjޏ3-Ru\=fTsY]%wMSRqsٛr9cP7'r_`IS`*l1Ʌ\W~-u[c c< VL{4Ϙrj>cki Lg 98{K"ccM u$¿e|qymehD!Ӄ!(7TقJ#TXp8>p(HYh)KT`W&:s$5E af$CGX]#Nr Z+f=&~JaU@*$j gZ}W^"Z9yqΉ%ax,j\aqN*&ctLa0HQ/W8ДSa8*EQY4Oya8Ϩ3Ty:^h/ N9T(dKjIh4s-J qsyM(Od@\A$~\iRt40eEt&P<ތll Owdqt;carp73RW:]]nM?x`$ 8:L_ӬI|s1䚜Dߒ%9\5.O.t= ݟַ5jr̤B$glxo#b~77!-ه]*[(~B3ږthhiEki[G/'%P$֎ED*M dRdI ̀=~GM+9ٛ]g:BLvI}j 82ЧōD)$c߷x#fwWG}Y`+ D& %"^V+O!fʽ}mHV/!^#R3 |ksg/QDA1rByk( FաrOd偩nH9I)wUs߼ǥ,$ht," d(P&rJG3>ڳygYN gt^0[ [YcM0Ew>6:/ m[rB(*7XX@]dw>6:-wp F;S(ЖB1`K԰xrAmPoM GG!. :b3rDM  ҈&O+ ÇI,׎M)^=;T{x,Gr04m8GnԸFmX7n;۔ƈ;j*_ {{w's=^܏^}:a7f{]_i1 fnǓ_רrFxPdBs1ʓS b-gqUȘ~Y6e | Ô&jVNgr4wQ+'Iz9b&ʢ=rČ^H0Ja) 6FدK)+1#!gQп RH)Gp4RY+|J0Ք7=1XIo86T,'j˝ؐb\x9Q0@W:-Z̏mƷbh%jM"aZjB$ℑfƯD$'bB1$,R,)/Zu``ݬeSr6sYW^lVJ©Ap"D*C1 @J!(H"!^2#E4f0J'#Z-|(J2XT6h bSefaSrѳ=^V,܉&ՐI<?>}$kY]biuQpy4SC E-B< Z4- ^"! FvL`,V!7.AV^<Cah^S\ci$a&L:} "[ncretImY:UZ`K0a|SO~kCPRW^I]=HLqe4ryYœV(7\uo'ͭ@x{w40 fF3@ :h_u/04Pv6jIoTҽ+pϩv5 ULi̭\SwEATx{eE}kKs$ޜG > nGQ ACLլek4u -`ZS8j^W5aS"l>E{SN"<928Qt0Dj<(Fb|*3$R#aeoZ1z!D<#۠%,igjR>HD&T(3ud-<Qwũdit.~;ݠLM~uyY,/o&{Jf1u8˓KK5OfS@#E *e˛pN=`W~yB& xCƋzr;Rl*8Ю)A{I]= aEW upsKW)w,}bomi'gA.vYrQIn'#GZ)o)ٮ$V젌 ˱xn 8biBnxW}a#L6RN_sN Z#x3X ?>RɆ6P92q'uԼm1/u0b4awohhq8dD+򛀚RU[0? -9v%%GGHt?n<ޅbN˃]&LfwjvUzNMɼkT3 ~KF|S 0X$MҴF.MϷ/ۇ"o/ QJ|*ߘP+gx"7qJF !̽,FeeK#ub wEFK:#ǘ`.(ʝ?C>ZKFHG wQ1ι:pj4O@LA LTknQ`8b$4i[q;R[RxIOG,QTzfY"XYQ ](ˉCGzCj}0 680h.exE>FNWFKur >m`d <<ysYonI5x Iɏ NJN*/NӰВ*ImyqMsLfIXuiZ+9{u%;iŠ>]UMA%?AUmZY ӭVq[] +xݍdatDje'T daE{,,@RN/lr-aNn&u#SyG.q~BzCur5M/)'D? ӳT"be{yu=3AhU|us>%a2zZ)uo;_8`S=]j%iy< ?4E%?x{~=mU5߸S/\)>eVwS^/FK|.54J$M{{z Kf”rl `RܚkkM5!P\^>}k#m at28Q4`$Æ:mD = i0D $?urȍww%9{TMTjFn'Af]`R El,c!#Z%p 0hH u:@j u l Oo?-Z oWV!m Q5\Ri ܛC21k[v{큥CYZW P{S r^nF70F=<| MM]CFx[S rLiu.Щ 7ָޭ MMÏ8W[[S rLiu>WDWquw4׻a!߸.T&-3mtx9I~N$szuL9m Q@͋PwfzuqrьϮlYOڇ:$N%\.oK$IJzш6ԓU0u$*_3Ў2qJM\zbjN@L84չ %;l 5gPLLyJDSG3 wy^Oʑ\ "FO \;]SzEȫt޻ѓ܅U ǾUH)ynJ٥kR^唈FJdJi kw{S=7.|Ndœ9mccW']N# .֩ -[HOOĻ:77^p'wK$n%l oE=ݲ^|>'^f~~370"W^ZY >kH(NY}%mdji5jD!}8:2g ruƫ0ʫD0"k=,2QwH,ƽ۽s#yZv]t-7E5f/60Q}d2 b=1Q6嵯$ێvayo~𐦄  qVzRY 1hF>{ёFw9I+ezW]6tX#yXZv_(mt"C*mA˄_& =L͝bh뜸FSpFZz aW2]]˴+wYݕ>Ybа6weIzYxh@Omy`۰=5<w$E&ERbV uT*Ȉ/0 S4XJu\[%"7QUڛ=x,_fcCa/:XY*b G*dŖc'sOeRm]be ^D" -"ZNX'h+NdꈴֳU(^oа vW;P"Ļz)@Re+ƃ~<_ ]xpϴV_upm=7ϯmp* V~J9x|JM_~?R9܃Mh%s`?7t.㓅~*ѳ)Jlװ5FV!4LP{BJ*P˄fe5B+ 4kke&\F#Sƈ”O Ƚgd/uJnu Z!Zg:<2=Px9Q )6MRGw?nGH_~ʗtǨMunnqTL3D?q1rB) 󘍌dMPHہQF]PPRN!Ns $Yn e-1G8_:hJԳ Q~[(ƑK7f>C(=3^7*Ǎߜ FY I˕vtxZҮ5h"2i#;kԛAUZ ARPTKVZY__r^tE5zԸ9<XTTC *wiiTbdqHx$hk5Ӆn~: vh)E 1,a`P7^5Vcv6] 1j SP+#7\ qH$ր\X'1-k^{:u9egi36=ù|KƥZ"W>?N_L=ӡw#Gy˩ow>COiO3kqa#m0;:wmn}m Hg^l75YE\^rJ`]dXٙͧG_mn^|8S]^jM bmII{IҐ6AjzIUAe,,{hFB'^&Q~B@CZi3ݺ8mгEIۦܻJѷ"g Jm`Q`hB0khE&w &יbSLx}nC:}zQo/[tw Z`)5Ê{53L`}vnhhS]*+;rhy'*DPu(U]|8=<3p6ܫE Ͽ o4Ej1<=tq9)~3fi1ӵKɮ},s4[OE~3}j=.9w=sz榔/ Wȵd!Dl*rw{bPb:h " [Mʦ+ aCnN=xývJ:hn%ޭ y&ZeS >nD»bPb:ht:_ [MtoR-_-ÑFX=_HF.` X8'_vK\nkvxEyddm(} cHkܧ_aN*K'Mk@ B8T?(| `'G^2g>3/7,x[":{'ɣA'BPI?KbSpqWݦ VSlF:s>Qxufj>E:Q6hSXR!> sT ۀfuOsotB+B۞xvfVBQb&W gYvb,g-+a{-ذuJlfMlLCMLwu,:YuܢW0b f!;8JCTp!`*1 Wy~a*\ ֻD٢8!t--_«_}͖i 'ė4xj*-@ԍsG>*@tL'8 ?aqiR}X/. <'.3139.%kG;K7{4\qeIrJKP^Ӿ"l_4&d@=j !ޒ[-T-Gk":/ȕN~ޤz CҪHH{Wv1@L {"6oꨚ։hKB>uMr2(eBR1kHN @Ѽ̩ST+J}:}t[[k81E+,:UA c 0$V-5#oԂyTo $cfTfx-o{&k4ߍ$Z%Ч7S3< KŠDQV9vuv^ʠ݂s^j߆wT!xZ JLMy)#7Vޭ y&ZaSL#l޸ڹܹ\O]~+FHzdZtNB,ţ1zG"fD4ڼs! ܬ[w_X{S!rhk+In"k d?kCmW4 q_+/B){ W bgr;no1@EQT'/5=9 OA~tӊd[\Q<}!qL-v—~aOp~pl;D7 ~R =O4?_t㖋jЯRtAT4Ȍ`~Tp(C[,B*Nc3-2[^7BK2Ȩϊ;:>L`ih31\go$.c /H`wm#RI.ĶIr 4ݸNfnCrQйdZ )*tp_nttgƿIh0~X h",Luw ? bg.&!UZ%LoBvA#L &žz1CL.9w-t\Bgf:: IT49\]Avc9^M(=U.tT7Lw)=Gf:ळF\;-䓊8|Dj@1'<8#2X{;3vyHq kh{}Kˢb\~+^)45Ċ`yޕƑ#"ev@/bydښjU}U%uL2Y-ud``|}>O:.>ށ/#ϊ`M]s% "nL#Cfb5đWޫJ< e뎵j 6n[U1H- 5RVpEH' ) }u~7Ҋ`*ȧMjEj?k {sL` Lq.wX'v̴I\-.C}vy) L5|]6mO!/ ;ژ4LGٞ^:iPdrHg J;~5CLz_͔aZI]W rOoق[IHb4 { o6F N>~NO}ͪ'Bh8HpHgp>bwK tRۈn-Nh--poSRАXU,T湪b!o*H`Qs e['[ݬ>}YE4FjYf/pȟcJVDZۊ jjԶumeklp Zv&lkSU0ٗ};X,v8’ZHYWasq_V&S/>&ȩ^}7Eڠ`J돷a"Uxc<n[.oVwl~_ծ.jXhNW*4.֮..{n[ 3kx%6G@~_TW||Cc62ڛ}yc'^l:e|&wQpj˱ͽa{7QJD&} D(R]Cݻ(E`ƲiSq d"Ubb= %~3  LDZRXP&X;Il)2EL ^FqOrɇgwap-qٖ z Cտtt8 <ctq#FF9Mij~hqE\!1->|xamQB0Q":r$&\g=`Pd:ΰa{ ]WKe9܉@qr·T{=,)_8_O1TFLq6ZKyPF :,AEҚQgY3cRT 5|f%Ff^]C7oQҨF\3L- +XTZ"DTZTHv)5&ZV&SuQm7P4QeSp`!6,"{WS M;;`זvXR5lQk+YG۪euS#S+l5o릪td44D 433ϰrߴD,2nn.$<7]:NAZb2sHdf(lgk 蔵$`- 9_{\D۸;z6pu;~xߊ9Zr59A4PؑcV,t;bJ6B!Ba Wt8ȝ1^S]y\߅w 1\{XܻS!>xarxF|zbD AĠOM?B|q@1B[RU0N<]\ȕQF(S{;\q- P~D2&EW(p4Y Me0@ĘV(%}ue,%6R f/n#1`{K( }+F^L$rrMR3B[ew'ΖSY{K v>@j`D.T BSۈϫN3}řB^&ٔwj3߁wSL»bb:mxV3Nzz)@B^ؔO[;wSLb했AĘFm]-k MtϦxM9#R˂j{SlFc l)A@D!N]C͘7?BFBhcҐ+1@V}*em9" .=d<##WHFӿ(QNCQ&j:b%XG4XW#a S$L[CA*LeRbO-xK흉=QPJl* &!=jIpiy^pҍ6-јULjmFDkݴM C[MebHb" .EcΖ[jpeKn_j$}7s">%R&˝;MM*IKW4K^v8Ҹ5#ڏ?<\^c&*XtF3eei fyoc%[yLzh'gg蚋K\^4Z'…̪R%OH}9T@㏝ )^U#KFɠHf)JѺ4u`Qp3IT9A/w|PΎhI)NOdl(/!';^iLe'Jǡhs9L%GDK_cٍ LH5۽; O."wxa}ֿ9RHI7Zg0oJ w7wq>jț!Z(5?b-3!U 8xƾR􌐗34a-(% T@w&(GFy(Dk)NE`uT!!T^o_>U/WW:#D=`''Ȱ0[8j;jGƒx!y~O.&{:Uz~~b?剂zkQ~&d{/=⯰Ob9r_~,beO@$8re[/3arĕ qO7/3I0EF hc-符ZQnЖXsDZ˸\>Oߢo/޹unԆ^lqFmTnm]Wrm'%3tc+ןŗŧμ^Zj7_˛rB۶$tHsw>(Xlސs>!F:_SrI? !ՕyD$HMcQy3e򼈼qgC!CFU#dZX@Q%Vvp$W(jki ME^gݗ#.*R`@'ӰWXc+(!؆|6TlU9ү>˘ ;ژ2*n| rWN?($twX,gޭ9pg|@oM4ASv pw]$hzUWIDZ Pnq>a1ªok4u1#&F7}ndHKxj5rZ߶Zk1JW#\kTR]2xөkj.B(WO2H=aQݭzz!LO` <.Q,JUg@Q:+KLT*R-D.B r1i!*%[f LR39L R $)[F)J =D9?7!rZ*5J*mCr)^nS1)FLӍ@iX1pM)0QJ[*!Fwq(#v;rX 7omwKNMZ ʧ?8vn_MȜ 2Wγ%^&yw%$&EF1jAv DDC1ϸmY^N|6~yزq68ɾ$5Gg5Vf? %Q")^ı"_7 t7 vlh9s mx9]7aYx݇pc1ͨ2mȁ؅ ֶ/Ṣwjb'ňj gHA`D10!VRΥ5ĉB[̀_D/| {d{T9:gW =dTM^|Y׻YJ {-PXT#n?ߟ.Ő#x).jQKJ}@WK xg ĿEyRJOd̒1;Wh(/B\Zt}c/8h optxJ <"+ 3G]p8i o~˓C{Aj kp| rQ+;n:dٖq|q l k@|$ ű$3#G:˯ͻ*g/LcDAӒyU FeY&\*P^55] K LZ jڏyz~ws%-c~= HS/%ϊJ2" g8$X6}xݫ[Lvj?bRP u=lyҮqeD()-" !7LAI ssL5 BjVC_#]NQٽKV/Bf3B \a"Uh/\v@?uJ8uݭf[z 3n<$eh gv @̷lΓxs~FEx<|B?6/Z?&CLW0LxG`mHb;#67bJ"h=X?Tkrʁ}(+~)wKT(1+_ߙaaS|n KROusjxɆ9t/u,g6?]&,>66q{glg|4N#LJPq$p{A(.7!(I '`yr} ޺PI^zJj:[ncUũW_~rW{QDNۿ^r w:=$wZx5s}+Tp7>p< W̔*p:\{Pc8[ɗˤ?52O?y6Y2:kb)cX8U@@#"!j4Pi: F  "k`4֣" $f!%˦.r޽lܩ>JvI٬B+!L^F6QX7:Ԭfs?~_g,)pF-f-1Hen_' I>ٰ̂O*LZI᯼ZJN[w6^~M  L~h$NZ-з9F]EdRR>m`3'f5CA^Q; %׫d*ӿMiysJo'[ٞm2[vgx 뇕%Ֆ:<@SDZzg? %0B:}F~3w#3\ |Q7vʯІfaiyJǙ)ǎؐ+KR ~uc2'(7އDϥ3KnK;Sj59w~m76`YiФm:.Yq7a-sQyiAkBA@m(.] Oq9Ki핔i5тYXX<1S\ i:#0C23t';74Q(ASQ wϣV@-8mRޝ"Kp:,fS.rB$ *E!r0%F)ZPB*@5 mpu6'Rj*KA %p =W&wd7!v'쏙>k3Wm1ٗw" xf%E#u`P8! 4%B@WIOqܖGnWfA_}[v<4.K+[{Go9 q`m|BK4)9k/YOЮU譕{m iΙIuˌI>r7[pjGֺZĽْ\-LLWoC}茐T%FV0*zϵXaU JO)V I7.dF\h諸.AxgڭDj&$edpeg&~ܶF-ltBbJz ɫ*]0m79r2aa}*!^fi?-mAxڡo޴J(݂gL/6W^~S*"u|e h\j~Z,F&fwO!`g3:`ԃydf K+vKʖ)ɤ{*]AZn vn{6ׅ`t-2mQҨ4p;r0<$k:CqB]}k%n=:Rs_@é멢OG"Grt'GuH;9PͣH-DZ\Ҽ̪^\?F8k,eJnٔ2&D%:Bs/f ͧcǩKx}+fP>ԮWjFIs. c{ ¿N?٤YGX"%.ʼnbO/]DWcAݑ3B.3󦫱HUcJqXKFwK;/l;:R*%^D\`;xRmث;T8g͕a=pWRŊE _od%aRSטC5p:-Wk44$mA e|q,iK\XsV.as@ƫo6~Hˠ"}*AlaJ~rߛxEz Yr{;ˀ&uv ,2ڻq%cm}u,O! >Ƙ†17{3 H_, ֮xwffپt͠|=,],lo:r-DhxIz DⴭbpoA>4$6sHʐXBJ?&fkq617_̔6i`aT]'+гrJ,ƛ.23CnC [>5\ip^g._lU.JO#?_)魫%%utkp S& 9v KdQ@bAjWlq鱜tyAN@q.ݧzq s(^re3o:=$wd5s=*#`(^=t-Ubz\`B-x ТR2Qg HgoyKǚ?@>&R"yz-V\naju)йA^[D3#.tdFtЪhEyg?8#m`9|Ėְ=뜦4}`WwpKiOGl{^ֈAT/A}ix0PMV2Jg 9#r7'CWQZ捛jVSݥv՘Dwd*fu7D 8kqN*\[S%Y(n޻n|(sw\*Nr5YFBkWtWUMQc`˫ V޶Q#|L:?U|I m0{ sΎ.t|'zca<39wǽ69Wgggm3)ߜcM*AA E'PevsT*6CV!Et;^g5cj_6' ꯎak/fEs 2Zox\O6{JˢS5|E2n ?E0i1m݇o߽̓%2_/.4~ݪ8z5oC};D2X$BAb%eT M# `HE&$ OBGdW Ȉ[M,NkD.`ضip4&E]/t֜DTc@aĨsv(FE!_"< F =^WlwO!)|۟OA=@(]o'[9v<{x4Ci3aA?_S?v=@ -_ +Wa!P=Zu@y zM? p<[3ONń⸷[)^. %ܗE@h5-BjVUԎK-ed rnQ~J7:ƳYf|}87GW1?L{^pZ76}eHňcH}`SLյ֋Kx3핢^xds|=>ITp*t3PS'X+x *3V%Ғ,F?"Ҹ(TF≶I{h:doYfh 363Ip5+;qL[cjA $q œApe{`a9fċ)΃\;zC R"7'E΃3.)vXv:q), Py swO9>( Ajg I>αܯX(L) L{m-a3_徙U26M~"߃^t|߇;CUS_I#tM_B݅yr.qL͍K{;;KzԾ/+Z tE`_ _8@O8_5bnsA U.f誑#pՈ_ =x+˖If% l.jӣupB$ۂ8h_x8UDW3IQ;*u a%=&bfߒGEh ~ I¿{.?cXRՕմ0A5VC 1ѽRc% ibAz˛.>.[e]={XΊSG 7*fϟ5IּDUgΞIӜ'IusԧOb왤(?]Y\6xy3 uʄ9#5z9'\TZs/^v4\\"|78Pɸ8d܉Um{Yļv.?=<pq"w! ᵤ_mxSub &T.l& !#fD:aV5N. uEK1B#*1H )9J %T&7xlPdx?By<3LA d?}=ڋfXC4p;ˤ|RC>mq;z[g5gơ*qD2 $DGIFaXULlxb0Ȑ/~52oП7{?%(6ISA7<->c}ڒObۜԏK?a? Q=eZmpUț3qg4~kY\؉Lj~D'3ţnitxWJڗ&"/k>.t;Y>x!>GjS oaXX`+(N#**(bH'q댋m,~>)P6fs4d{a19E&H 01s1` TβEhA5i49GDZ 6uk8CKg G+At o}HQCpe-@0%&g픐 FᑑNpY10b]aIQ !.rGJ1$05e[xYClN FBw1;[,D3IvGQ&NI0@cijDhLZ{Gs c0skGflKBtbxv8 d3Sߺwf^Xdip ץ9V)_{TV$SįQ\h T +-`H]} AtyQjW$CmJcg䤤狜F&NKNH&.BThLKr.G# 28q9TZI-0o翂GOMr2F>u;1Y-D)dp#GaSRل#bތEBa%D)~&&Ex۟OE%xr鶞T\`)SоNٔo:^tx̄N5&λQze&m_tZݎfEʕXݛ5ͅiPsnKTH푥 )r=S3dzژmERew@e 0=sOnc弻SE;oR{*g̷Un [Y BNgn'Y&H7&J/4TUC^=zOqTU""*yGL$RS|JHaPY"G5{)lDT A '.SJg ${I[$"hDM@Vap$0b)]bbJ8m3\ هzRj$XՓ({'Qa$ NOX1'N`a ïRHԀ$f(" "~+FB7 P5F%ό= yqot{[;{v[70Su5  v^8ͱ L/^遥Hܗ"zwS{+8%E)Jf vx#v.=`OoH߯#t5S?M@ikdw1,z0N?Mg^=kg/?tcaHѰW,B]΂:QXlgy/8s?c͝G~=rp.uT"NvhIQ8 ]c CGᘜB*e|/Dmjҫ;~qZjz見B a+:\ٰ˖N~Wm+ᖼu6Vmُl[=saӶܦ"2ښ6Anhgy^~a\>zQsr/j.-QI!x1jm]9QXy@ĸe+Xc*1.'Lh.|*T,/$i$|[m^&n!XYnor6+!K̼_QҬ2(~QI,*dU!tlٛW_}y2t7Mdzmn t:D^!(gg>-zS;_% }(,f^42EL ^i}-oYӇnKbedXښRarcD&87B(~ b߻t63&XS΃gn,Sg*V KS,?!XۋFx)kQ2ϋ ]/])i\)Rxik§~%"b F2R$(?+.~k.5MK0F2<\ז.)5gX~כ0|M~s*VpXo:o~n f)ϻ W>n͸fb^;oěґ&QwÞݾB_?vO>{qB-xJr~gy1d5 s{-:CyND)fuݐQd8bqn6qt"#J8鸦\D  mcCX 9`-|"Q`@+%8jFc UƶGlb!!d,\%&ƎZa4Rc. uEK1B#*1{As3KWLL!W!p:$K#!1 ċ\Y=Pi#j:Ͽ؉x0pmʧi(e3 xYޏc=fZ݀W֜ 6y8:n(rJ {\= Λw?}pwm~Y$H,2= dF/1(fE{ݚQ_$>t4=kk,W,Vu/+SK:󩓟0:!ӻaG7@H!0@ ,BCո64X,XVi3x%TZtNӽw'cn@> |D^eIw͢.UDU|Z;³/oE~]ߜafʿ"h'&E3CM֨MjXonVNnE_Psf/,_~@AV'fS;+EufcOߪvsuwPrF~8EYLY3ZDKIfON9kVwrZ}"LvwV>Ӭk"3}u(|~YtΠ_zNh}ysڃ2FUoɕӡm+'p;'Pv*>[hh?ޜWxR 0262DS{N!4Bt-zcgZkA)m%AIlΒwDap YrR ܙPFvZLa[=-vkmk%kvjKCs&K zy꒑%UFp(:gv-N?x-g[SlP,VೈDJ!84:œE/UO3kACAi\uJ ;N%~a3pGlo0 ,75Vi5(;0Z "*9Exl[vQ}%U#j+(U RJ1 ? ժ1B9ZDq9QOQvjg9D()Um;vKP,BK.RDQ[i7u)m5Ę!!sy8kGxaNkA ҹdm x=Ũ@5Zmā˺ zqu/٘u,Ĭ֘bdis1BмfO,Dh%J:K V[ d액J5p11R9DXƔ#ˆB@5ODwGԈT< $ YrkB2Zi/Pj'ݨi8)XxFj$ؖ*Zʱ3Zvdn̅d+gG^XCRl9td3gq2 @62 +hBl{@5γ@? \UvU4PGDueL*zmke#=`)$fT;Wyْ-\#n=@Oޝ@/ ?+3jx7E-wGۿez(0} /m-U{>pf|\@ot~Fˬ 3aʹcO3i; r!f`3:q-CApa} Mm5Eb;(yq^v7>ZXjt4AF)u!#;:"s'P&B+>{!`9zv@s4 'oT’8ztB@DI @G萷/vȟ Ur ~Oԩ.-r/_^YF-^?+P18g?,|a`Ef]wU.U|וL$F!gRVDu_@v5/ xUu5?.V߷gxr~W:ѷggmMxSil,ƿ[yv!@s@QMRjr-A1䄱YdYd-grbM{`b{*E;I"H)߮Q ;i -h P9Nr>{5<}c@C(9tա7P o %"Ԙ1\UjrKR^:[k}.OHU/8_ :RD,9,{q 3 =SKxQ˨Ӗ^?{G,* `1n9Jޣr|%}4;=bw(??^DhxAw0?U:1?ˇo6p:['(L ӿwx&)﫮` /n+_pL8=qoE;Ogt]\E[~26*-)n"}tqI 3K[hr@8xRN̩:1,9 xǮӓ;}}*{?އvWw˹œn.ooWBn^=KW! /.\:XVI)((Es6c, ּSeVe_6.j1T&I3PHZ_kHs"kTl葳N50PϬ֜œe\+";MkNd?9 Ϥ@~9ұ`AkM Gd]L7Z؅lh6kkOvndЯ^?};&2C <x싵'/s7XE%_O=p^t$]ѭIQ>[4jV++MFŐ"x~7K:LS3vL[ZZ%nnЏg)fߗ.e .%YeCn|ku_'G;\p DZأqx:al_w|]N r#4Oبp$^Ϗ㬕 ].̖wUUxWE]WyJWʨZO6Gp )9Č&IhKPB!%q]sЉ|gqwyyF5o̎r}IݢZcwc;-D @::e4hIރ+![RLRa2UH>阢B8XA)r_o %sO}6# +(l$Т$B%9Zas` 5Vil$`*(iAk$X`Zgxnn꛿$wDZr(vW|/Պ?^~occAS8 KFY@c mxQҼJ灡 |kw/XaK?~26{vQR)rվ(kK7>wgb5%;-wK:aguxdpp]6+BB\zkhzm9^d6U>VQ/dqzR@_v7D< =Di3֕idlo>w`p"|{շ +*8L&7Z%9%v|mʨl=kX*(K]xI_L"{nXu$'S)_:L =t: Cn|23# ĩ'd$ikzAZQ[<|={H)x񥱉vSЙvq`Qm>5%$9%pa68ր^X^ٻݢ Q[X\\XaHFiL¢ནH!_ ) K c+j}JtfOһ7gZGB(>hgkv//#we5H2kɳ nNu^A&a_K9rV^|~66w=fs6tq^=q72F?b6+cIػ7 n2?7 nx[OaYؽB x%.vW%._dTZ*-lX~wW Щ{?%c4Z^I؋]84*=uG"k}eE볫_gUQzf7o>7o*܏DK+.w5c9y귫T;s6)z!B>EUv%J6)U0WƓ&Ym Ȣ"Dd`\HR ˰;F=*ߋjoAEH=D:H6C»hr&joT6ive#I>Hd9Y|h4xCٻ6cWH0ھT_>, 'OW[^LI俟!e ilrf(>MsեŇ*҃9$irQJԪXÎB}юXjBKW>[DMETUJRG "j%TdUͣsٿ_KayNKHͬOTw:n[y>=E P*9d-A,طc gI\u6[aBCbby%NMg$a+t)v"}ak?O/汶_3:rM>0v q`FjWoQ}6o?o:uLQ7*܀ZJL*3OI7Kphj 3;j޶xD$qkK3\YJ| b3k8 (YW'×Nr cF`vTh'z(%#('&-` AG &j:\G 7lʶiJ\Z[&ВZ Bh#31 &}o!wV(W_vgM i!} __4A~S4M7]͞ & 0"wFHY3g"2ɐ D.zfK_׭\^H*n,'ˠ0Gg$s6,h!t6yyqƦ,`^zxQd%c婰IՂVzhyڪ0ΣbiJ<,XsSMr }sO.֭˽U/;UW{Qj漣#]M.mn @*lok+L޹Z? cKt}BD6K87UbRfJrf;F3Q9kSg<8($a#'1Vuꭗoba,7쿥Ny{OiŢ}C\IV+=0ajg-uas7l8i)V4z/cmRYC#YR!^hҢuzCRthS{сj5ETA 괴Em=V:d-/YU\lt5EԨ:9ΣEOzȽKue4Ch ]BSOԶ5є%^6kPѥ&?nGx;bs3Z(;mE59h8$0%}om*92<ƫ} ZCLz8%bYm0p#`^ tU@F*bҒeH6D!$W@_dL|L,Z)J7>o" ͍\M|C/4C-~θ$6_1+ ^2=sF 66`4oH *N&&f%es 'o/U C;㢕)%dGhFo.}B; :8@A,2畈1VK>e.0h9 ɰfea݄xL L|A3\HMz֠FK׌xYI|^d&LYLi){^h)@c찵TU=e~huZڢFcYKKVQku>F}JG-=h-ET;e= F4&i3>\W PUTA@ t7!5xDap{&Ͳ$?A-&<.Scx} #z9-1s;6`BgS|:e l#t A42>G=TdcCQ3Bx B M8)|#5͎;nMoɊ![c65@rega3|55# @e+NL s!71wlJˈוSFVIQ1,1ç L阘5FjxO*l>}1!Jvh#<ō66IXZ$b $ffeSySi&,5"4֛$EYGZ12+3Ɏ)aǒ h>7;8~ ^ΏwɦhʎVoeTC&[Xf%#XM/ϚHf VjQ+dA'N0Jlِ67EDRF   'Yb85Yruo-JkYf1?A(B)E- ʹ:h@hEQ.OKpŶ!X"92]@a/Ǩ(hӎxdԨ5IWdzF 7c) T As*2i)2^A&m PxLfM&mIB dJ{F;07B-<'pʤ@ey1*EjVky E!As0–LkI:P).%%Hӑ)p!- =Peٖz$-VyZ0E*Qw{N 6Bagv68@^+a 'DR9LΊ`k$֓K\ Ȝ1SԺTJْZ),6'Ǒϴg"1*# Ks6[e]։Az%F9Z2 -|Hڑ`SL5v&eUGZ!ƀFT`Y4y!_3R| 57@;g' *ņ-O3" gKd;kJ{c1Wq~; EfU6'_ΠW)e5{ۄ%WhV.UzxYEhT`4lo}4L޹T1H${5eWP˛6E ̮k}el,ع_c9 ɥ{KxlSbctl]c/Qd7%"O(odi4( 8Zs͂q.hEY)ԅkIZIhr@ea}߈ͨ}[zSׇ8m͏;%Gyj]f4tq}#3R2ywawQ=ڸps8]ݑ=hc1?Hf5׸P2]VH~׉U.S2@6ОRĈQNbO\&w" F"uq''77MLr7 IV+)0H4.k֔ʱ 9mkoU v0I{utI/ zUw |w4w[%'6ˤh,|"x44ւo]R6qougO6=Wr4H#XGʓtp*2=d"a;w~:"j3oﶇ)E3 9=f >qԻ{<$ CAB(R1hT/A 2P^󬾿ZXyYWS49`Q:*ԎiPlMNp4p1;-;&>ք߼)ĕI2me6M>ROt4P˪LH8{ ђDjim;įB}raB77U/;8ۻ~ oZT(1˟­4VJ`/N._gWgJj4<54 aOG\L1ͿJ%(UCՋO.{}bRC*A*Jj#3-b0u$ZԖuNo-bUu*-ގ{>-]lu]$sEK{uUsȥrgo{yURx9(F BGF SʦROd##O9 m2:-\lec;f-nY Ȟ5LރIӢ#@XLrRi5gFn[$2(3dJUyH (z!rDB9S!3uHnDv֕8;2/4 %F(ϴb[gZN.Ѝ3Q7Wwd2_ոf'H($T5ǽջijB5:j Ycui< kbhMlMӒ(\D:e8Zt6Hqbt`ZoD^ACcFrLAXc*FxR}\I/=@+, Og`}`c{%93r,nvKvJUXUl{Qu e܀V7x_?z|U&O)umlAX"DFik\ĀzTKg-(T Ԍa"MS,jB+kF Ir!1 19rµF/ʚH4#pຶb0BT3x,Z ENh:OYq}un8wcpI|MͮCA$cЇ2}1F PhcA&P ?LzfQ@2O2C#*E|;iwymK6׽,EqBQDyzΫԉbr:azeEIkF^AZwrE0JNOL.b,_e.=WåX|2`ӨsfBkgzS[M~z*~Jz3q>]NKAWC+%ֹz 417k3p-댋gDM47Qu{^WЯ4FqFR-c!q :yIQ= @tUY DTݍtA a.Yc4 `m&gnU_G5IfY_%unH{LZp>^-^5>Р#єhuc9A1w|˜t7Z@a-rxxQ%zsc1/ ԯDEx?湻c)wr%y&a;N ,GFF{;Y' U,H86.,WBAގ,l" _Nf3IA|:v X/<тu^zolPIHV8.>hb]Hx"BDS0"M  .Q0KL%N[c%}CZԦ:nQu+q{1fp'L>U_p4P9r~y9n.Yvܲ4 h L@dM$DNhIt&Y[n%ʥ?Y3M@s=JEaQ͖Eyr}.cnji=X.dp «\ =ߕnV2ˀϷ뿚p|5@2rH01A&e15~zW'Uݜ(G-`W#OI"ע9wk{usq$z]y/=zP&Te| @(1.<qZJ*r#V}}x]"ZШ_#YÉqe6B~~ Uﶊah= .-keB"#iKPy*̼IR\#]̀4IaO)/yC 晖ODAmOZ62D iKPy6"\IuxBN._eC JiK0y$s ||sFLV чl{:x0N^Aj|eT9qYK'pѩu8,!ȓ!GJɖӫ5Ld=0AH>x kѸbRiM0TY"ehdB1v@`rIc48 cI+ *X JQ6IcR  XL/_Jz42ZHPd%QjkKN[ | G=<2iye!$'h hZ)Ay7fn%ݍ2YhCc7e(wh٤M@NȐ!6ϼ`F`]QXFVn~ժG$kw3O-8U6_ժъwKZ8PTi{F61Fb.Yewܭ>]j*AE?Cx]hV-u/0@.(>77JopPCl)wB z{D˻5ڐ-p/q׎agx3V 2^\ˌƒesݻgџ[>?`QrN wm\se1'푮NQt|9Ň'N{o)?S[a0l]ɨ1F591rٵK;RTGfh[8ݙC39 Hb|y袢ġI/ꢢz y2}SjѪC5lL`NQ(d~ܙ5 ww,5F[>c8R2O–ft>_a8!M\KZOWE_eivxNxhz6K)JqQ 5%ɟ$N}`8|^ DBG])2X,ALaT{AH WU$H!Pƺ(HR)j7éhI57(肀 (hıDF79љ'ĊJ'eIrn8ՊH[E5Ƞ.uh52n ipPfңGP-dM6ct`iJz" "j[ŨG k8聻1eP4FH+Ld APIp6. 1~ZehPD K"pq(!:#)-x HKP5rjVXr@'P5ʭS/qW 6g?" Yn/[?uvmNLv7ogy%M:E-nЗz};jkeMU~>å:[۳OuD*1ݭ۞MjMp!3BPoŻYs7G}8&=7.{EjyX$&xé捡4)$Dž^yZo;y Nt"3 5֌4jLb?8cc$br)XaaS[GK8M/Y<p'` Kް׃dfO.+F%fY៸oo7׿_JI'y )  tP8Կxxq~Miv*:jN'lcz-En#UG˥n3;o#-UY*Y<pP6uaH͹Tӄ21])M3,rsgM9 9#xrլs+;)bRFHk>%~ru}bvmzf="1;=N WJaT8rpwmzJ@JK+d &jsaUg4lb`-&݊m KڮN6u4wcq/$UH5~^{u$Àţ<*)OT(ۻ>k{_*z?y4)P1:ƻ|>ۓcS~A[kWˋ-јأ)k-̆Aj:@Tw>Vu hc':fsYأJN~ #,N?q]W+V)-/vG|ljs#Mn $]E3O91]Rޤ>ɌMǍˉfs^Y ]6vY̹ځp=ڽ_0[YQ b/aTn2޵OMQds.ZH-E9)35;L`I&yCl8%TzeU ޹-!gwѬqֶV<_C+? bWQ}'ӤT?6t'iQWmoU+LsN94Ji7 ^gmN] :Muԣqsll Tx9YL]Ș^PJ0')g)"W2)%kd@`PJ"h{4 #=TKHA-` 7brŁQa(ZnA*5)(inL"4<,8GȖR7.(qމe (ؚ7 #QrMH]hyF.WeǨ2 ~]NyG B4E +Я ՛g='%'g_KLJ\3۩❢t4wd^F_ΞUzjnVE/ QSQmhjQLm:Xm4W#{^t椗tA{tKFIs")=V}K}DSE;Q{z-z-S;KPյ&=; 03+hQw`j{ȑ_e̗"Y4.%{{7`o"`wc4F3,ْԲfwKv4&V*Y4& {%E:dRw! {ՙz]g$G}1V9{R-mhXvjv6f?"'6lgY$< Fx{WxuP4L;%lnr @"/ !-JnʕJԠL\K V4B.$Z9$PEQ,+V KMln 0K#){DRi"%=EY Z/=T ùV Q4 j98A _-T(kyJg@NQ( H)>c(Qj!XC5S>'BTsHKoӇRg(%GAMg|[/lqkVOٗ_/ *^JQ ]ATMnh+?2ukO͖_ڟ端bV7cÐcq6 wfjА޿.=(p 99c|g&'\7EĢ#x݌B{S^tXc J@v cE+Rc/(J={/Z(+}vv`KF'%?A,J^K/MQ YV sc28KM I4dq^R^ Si&<,j$Ah%y1CaaA1@ 0! ДHhxw8tGJd(e't}(H! v[Q[tR,33^PԥPn,CJ $9 :0E8Sx&/>s lIxZJIs.2fL0s''LJ?MIs)s0pG!CXc0MD>!Sk'VSM(pF#ӡ$ȚF9"e<4|u9*D0"|CQ!Pe\8oӇ="~Iiҗe1* oCwd$Ä祑 ЦIAR~ĭ))-3b)S]vʋvTҺ=^#(DJ2KDFR* mr.֊<e Ҏj"Na3<Ӝ֡E^̋AO !U֩} e5..~ka3gF)]=Bz[n],ݳQY/hG?.mΜ+?#50|~I7U{fȿrՍ 7uލ25ljB^r̻1w+pv0_7N&ů_6?5sx/瞑u<ڛM}ɉa#7e5v\_N^۟nFF|>}iW^8+U0/}] Ȓ @Îh9li`\rcpfˢ@YVlO~aɚ^#H/7Bx_'v]{;s(a_ۭ}c\˵Qf8 PN3N*E.tdAr B Z投Ys]UjoJ {{'DfjlFMvп96!t[hM;!A LSdslqa?f#rI?i`9b^ŕPWM=\_{~zc]6foái~k2:I(Pю4Q>;9bcfuܥs;sn t45, DžW06!B!}ps!U2$62_au,\)ɫKYf "}UExaaCwݚSz]+zV12ariῳSpX)l!W~ZO?۟m֑bǿN统]%^QNJD Zz=8,~ ,ӳdIBKIX5rĻkl+\@WLjobO:j*-G B4ʙ3׷p.7ꖎot6NWيۥ\]okBJ0[ݗ$zQcz?Vqb ggzkri; i+csf3[ 1``J]KM].yy\~S0lIV_r ݨt#0Qdd/Ԋܸ( Pz ΈUGjq-fBospp#. 1!ws0ӠeF}lإ]xw+'QRx(Ev&ͱyV2Q33]h(&? g6{Jџ l{^ϔdgiܐQ]r6?Ǚ]l鸻:<'/ϟ|1s)5 L"ӒM)$Fs㌟gJӮJFXq~3+}:6ݩ/mj^Ŵ6ë[_;-_Z}CaPK y!SRQ=H b *Q\ Jit&M![T9EqA1굑Seh!P[S\v;4 馑l%u@ny^LЙUh!r(n# G{?'ɌZ+C26tZDnҧ'<} ަ;Zp&ŵ"gngf1Bιs59sK+I˒;mB39e ,G509Li'wP7wZ|ؿDXʈE_(Y'A0GdȈU2_D˜141ڎuiX #(j|[C5[,=Ek}zgWA,j٢ nĜ`^3a+E7-zE_Vk7e/bQue;CE:'. &2tHP-Dh!XXy W "Ca,2FR ݏE^ɔ> , Q?YTRd4OPS*1<LhAƥ$ӆp0KQKAYnLsSׁ0hTQ!2PR+EI'#9QҀa9ˌ/ V䠋"g|f7FA{11bu7/19-C!omVS3{M9ӟru1ԏ %+y(1D\9!)S2mQ"LR * gDPUPM șa10 J͌9x)HRV DƒRN)QJ{qQ҄kV^7^?=8Yu Glb/(=ݤQ\H@$?df tHp0_̣)vxJtaP[Iuڌİ8?VxB햊A褾#Fy*`-=c.Pu!!߹&ɔRgkC g)LA T~,B|wǯs h̖1ղ_fb%/jA$= %g4>v3a{ڽ̛.E,/?o&[Ic^V 0OI؃${r}a(%E=IߗduiT7) -vb琽|ByuΖcz_w';.S>C2* 1r 2i,.(ְO7LNk\1``$0N9]x& A^Tu~Z`;^t֝:/ @$, Bz |a~;Jp N(tQ(E;!%0ic[#Pz#wzBQsۑ .RLE!J̀:YU"k!FjROBՔQ%BIkEr8֤1brFUa5ut 9%L4Q5#uńpM̶K\@6Dm{?6Z֜LINGsrF})Fb[螑LaB)soFS@'=+5r%],{SLQtF|edC ݚ~@>7 $yٯy8^LVfc܈'߭]fX<ی8zV&gɈmm_[%`cVѵ1d &֚[kn&`3B9Ompe]5~iI?h33zŝ*a^ ޺˯@ 59YΝ]dK/ 0aWTiu1GZsD~3!uӉt"9c3"[*y>svMjۂIqz5nMF/u~HY[p:<˻gay%1U)(.as j]>B zϱN2G1@YS@.aDxA`\`R* `ɵ7 !:ϮT/ͽL2?1"?o/=ec'CVa::p t^IjS,霂8B,%?AId8,w_l'1B6'u8'"䪷C`;0ׯw[AZ̨$c@2ҧT.8cg-x:n<YB욲:0u:2~>qӜ-~ ʖw%X%pY} H$b0GXFʨ,g 9$W1XL/C~NIB.3r)Edɟ6#F6Wx17'=5beHeIAb  tq`<ށqx•.9'Gs"!+b#,+x -Q) ͆4]/4s;r&^x+q1'}8J9J\8a$y$$/.rs=,T@as7D)Z;II&_o/OobMîk2zJADv"g x~Gp ~RLV~a~a~}6T0h܆\.&W Փ<.d ) clJ=_QǬ$^#]mN{~ȺE%s1Ҟ CB\e#a-̽aV+C`8y_Ou"Jh>tέG cIB4/7+DDpgHƂn\<^QWRH#̃lE$8Z.\iE3U\|nQaOwB3V 1 밄_VAcZaNW$0D\o%CirH%>WRRž@({B(d'TD*2/>Ē$!p$:CH\(dQ !!gJ 2uPxFo~T /]e° u5 ԰LfT: `RѓXmS mv98]' 1W7dB(j;]gm)$*A1|5颲X6/)N1Pȶssg]qE]zLƢ*-%A@Q#ṮĨP9΃#ca$8V*PT< UׂIY( TsT /yrMqeĭX!8y٬nii‘aJ[O3#fN7Կvb\`0e)0K&]x,(xlZ"k$OQdm %xfރb.@2Gy#LF3YKۭqe` .4M,䄅}n%J +~LX >|0 l Fq#hV4 \W_e{1=i5k8rSIeSL(tũ }?[U\`2y36yVD YSa8kt@ƭ л'dȀJiEw֙-ȲghFQtR_u?sٓa8R'~` QABA8o۫c|U: >=?ِS)o 9s$0褶qb:A߯||luSoKu2U76f *pځƯRl7Nv0_Y $$*ߏX)E"%J+ǿD;mg"C1xk2!YRVRtȏR_DWPC]Dћ|磃FH|VGEW;ߍQQPpeO?I%̾@^zI(IH;y YjsV 8ttqDEbz+с=BKU_ALsi7ևoNA$%8Dbz;с=FLlt)Ys_+/ODGձqHSVVz]&ͨ[oZj(q\O/̧4rwHɂ0Jb#;6|tzP`.\#ʛeS"2u\]RʣM$(x\ݍz*! b!L߿hoP[tr=9X{f 5tK{hEɪ1jJF F S!<]0 bhgOm|l/Q3FQJ ,^*,vt] Ή8獪|m03{4DJi>أ ,CӪDgM95?UDq%\1Mϡ9(e{15n 6(Xq@G'Ѱ=42 ޼c5v+.JX fh}>?ak~Gmdoiɪٓ9{Y=oHtcr h˸{}[|Oژbyb I6eq]c,v`Wl 駟N?V9lwfnك]ۆzĶ`=|vSϧ}R2w\-f|QeqY4i/qNWWne9m ;J#9'ݿ!v伾jTM?̏Etu|84غ2\h$ܬ84V?~x9{3rܽ*t-v-ӪvASW9nuB9oZ(y]O9nCq->.W-P+rVћxfW7n3*-Nby](n`@*ht/&3Q9id| >WiajfsiG7 6?dWdKV',OW/; )-%꯬˄T-RЅu!H$,v#h$rq$JKED8QHuI^Ydmە@VsBh{5Lv _X(ሁ䔅 c/I4R(1rb II$QeCՀ8/rc.Y$E""Pt "$ V(;(^ROxBOC;і՘NӟtR̈,"sQm;%iJeJQ2+vtUlf9z|ęLKXj>.~nn 3V[񣙥lޚ +2zz ;|3=k2'3AUĐo8Sn8ջ١@w+}F|s/iNޭtC}{*S[?ݠs>NgιࠚnoVCqnp!%Egi?n'&pyw]hqamL՗j2r瞍0 νP.5cGan,ZvuDw8w%_(C`H>\veXqY2IнKO<>o?7Cn>^_]Q'k |TxLzh["%탉[+L)_uOsqNo'JΔ(ҳ4lh}$HL#JQi;U r -$I *tCn*j79Ibu@d\oZ ֝ے^bgnoir*BcYy|la17K4yL'%zf7#(Y͇?6<sM*J.(Ҽߗ\PTT\pk^^Pg~L涛zrkTK6JeL̝t7(ٱwel6pj;`ѺJX g-|ʁ}VK ̝;כC^*;U<{uzv@xk:`g)HY*?j<*\@Iݢw5mPse B%֓]g3n4֭f鳫` OicVU[W}ׂ$x8t6}IWmXۥuؾD KhWY ((n;3U }z`1I<×rZ%X2^8{u=GW,J> ?]l`BM62h*Iű;8aHY\HZo?J``@ Pv0R 8{L8ϹdCvu=LjgAfqƓc^x{=FDs{M)i]?0%8Z5?osΛü.;r^M$l~>U.ط5$JBN6t,[0 ஗6ˮKn7Z#gATk~HHKfǎ u8+h=i*-۔rN-6`SՐsDMw#qK LU q7RV iè . [I B ~1z sLo*шCT>qddB!&B3#C%U*@BLKX$p)sT[WeoϷ4a&1&0]u(mÕ " Cœ' bM2E{vF-#]CK2߅ijuVOhqu/~Jmo:aC( H{lF-/보mf%inL^C:ݳod4/9-j; CWuշ{7a Q;c0&lB~5%B$J~q/^aZR~nO 9Ap^\=w7]wMf)7GΝԯBWy*%4%EK R/ j$zK/xhKգM-R*XJE*CVؕ$pjO!TR]ػqR.ogASj: =C<\STH_i-ձSh%/i/\V-M6+^6b|Q )ҳJWBݒ (bmU/f6Ro{ʩZ %$")a0DD m #~y$pbyp>?Lg|C*B۫ԣo7 Xydz۾fͣxr:`lFAuk2k lHցS<'K@9"QR.XE:/դ,:q߱YWRAs8TKFtֲPIN#E !&ZDak%ُa "9eJq$2^*į tzݐ EL=t`2NL5B*rZ|0}zL;0H4%eZ~;IQe7LuթJ4jP"iZ%ZRG"e#QA%G](GuZj4_I( b!* G (kP&Yh. J"e b#C I er-BJbf`fJHR$48 hHTZXƷXN,Z)ޭlӰ[#wG_󰿥q_N#X-暪 `\x +3?qf< bcw4!}۳ܘv^U>>憔p~;=zʇX>>Y]S)CrL*?{Wȍ/;e d?0`7ٶJNGZ/-b7ն Lf얺 "uz`.XT+wsg@^YM Z||2ۄEf;! (2U:m=)ç2h rx8O򷿧r ;SyaWJRߍz :L'ߝ-ts#s3#H@&™Ωq }/jY._}pwq#>$Cx_C\I7q1`Ƹ.,Ҁ(XD'pߑ{szWqGWUo B@؛/YZp(-7}.v,,qT~ SĠ,1!.]G="i}[ĕd tvBhc5nD2,.Ò^7%F2ZDXL;Sx:9h)S(q\ՓSt&ZӑY;|5'}|&9DŽI.ݵm^' Ubܨw^lD x.~( 63pD}{DXt]?:%AǗ~~m&IiWjFjgXefu?>MO\bB]e݇aYfn,v (ݏl^fg'H/O%5oGhfy|]~ vJo:/lhFݫԊ?FƂSLr~WmF[Wi2V|v(j/t$ڮA?Subb_kWB>10FR$8T.cfڞɭBm}hm U?D@ޙbf<.h2yD:m+)ؚ/1r کڱ#8?4]F]ɠg,&Xg80#LdbSAkVy0ksv$]h4a=^y7'ut bsI#)"Y{lfSl!34E1Fg H0?(æٔ|V%HN'tLA$QI+s2qW];(rQE5qN^FiwJ5$%yЎȻ_ڠsli<6)D\bT.غz iIGRQ^Ү~ ֛ ^R&PvwN횺떬E꘨_즳5/"cHQǛV1b!9 Dv*^u>D;ɔS3{n >O (me WiC~\l1)V%@ U|OL Bx.($P+@5 RBrfɒ1 xV:dH!32ٹon~bA}ۏ:F2ݜ P~t& x}݁sL!Z~BWעf:#[4e})+O :n7@$-s$ o0_N)uխxt޺*2rYi]0\ I,sY0%R\` ˛@yFF&/=i:jo1ɯIUޱ L|G-FNP!A~F q>B0 ڳSM^7z] ICƦ'MCVePn^}izħ%фΛioi/Na1Zi(#ȞͿsY)=CD2f>f0LD|M~֤.ww0m%?uw6eV!teb+rDzg wk3+ΈJ)RZC +(XriN%ĐopVV[D}tm$/d )5evAKSK=P_ֆ|p)BzqY7 kwf/hb1QwXK"(d$uukCC>S .jq}^gcNu2xN5o]H+<| <CN;Sռxfq7qLOabI=ũ7z)F};)Oµ||-yw`:y?? u_zݏl͆FOUBC8@Iڥ(Zu7 /ACا!J6abHAH^vaAX_`J0!vL5%ubr@wsV=~V%uI&N2VQiѡcpuKr%~"yx7O9-EQIBڟfˆY g] ?+T։KrY*!P_}H qѲ@O60.kl ]tY;۩MG+JX89j"b`U*` a9d |<Q{wq4(y{UQH%H^B5ʔXPPxn@Q΄ں!ga)݁\);PĐT2g4-eN(]hs͓({ݎ8v٤`BߪdFn?#`5JFjR\5Wzm#)ym B)JIRήq an|a[go~^庡(S%Of:/N:J(n-o(X G8 xԟEwKs$vMV;Cbp\ck(}Y'@`@¢!h uZ-T?X-ejfĞ! : ڧJ`&ah!=#+xA!ǠQOa&2,pkl1/Q)ܔ A@ U;{W[<\J0F%oeÁ`8~v :fĂg|=Fj)ɛ߻aY&6{sD8"El|^1n3)ZW+iWjIp/bs9!/g<5x~U4F(2oW.Y .[*w1qY>ZАQ:Ex]"(WwqZ+yؾZА:E!=TE)IMIZ.yss坅Z}K j&[V!kښ\=VmV[➃C! ޜu7FVfV|+퓝gvApZ_2ѯ_6v&:[O0biՑϛ7">Np)ng2Gj<w35Zuj^,#@V ɅV+P$+1%Adv'Bbzo>q(Kj\ U>dW\^3\P>! Z9.SDqs3T nq%" *s%y % @a8̙ hUh KS!P&JKAܑlu2}~2'O67W/[zD=e;>̿2ѰdS w Ŕwx`al6{~72pj723z> 5cukjI;Lݢ^\/{RwgũƟ-/kY^Z{`>c/Qw6}e`Fmb5^)~ ֫u+z9a`uPgeMk_k\ ӬiB]Zƍy402pSIc0iaRμ =,FKwBe^,#Fǧγ;BrG\ (jۃ {3Im{⍰ `]]G"Lrv ]B`0{$IN@iۼ:5OUC70`䃽|ɽ!y: ;uByN-i%2ĝC@ՓrGvh{Rf];Eh%F[ҥKe2;E֯gښ6_a!{_TYJ8p @B2R)LO=====ݍ9Q?)*"!^Sw`lG ݍxKT$k{)*@ "ʜzՁQ1sHy(o=+um1Ie I/&,pv@"ɡR!ޅǐДܣd (4qrJ'v:B[,$=;͔CҒho\nƖ镱U8wก1F޴4BSj+z%tz$4L-PK;3c[dh.lYdNj$;ı.˾pbgێϮ4JQ,)pXimXE8}z ob+-J 9l 4Tcm䑮zJG25et2x%5}XSmSO9НοY?5lG^GQZ߻O}iĀ&ӁùQRJB .ic|S!y(.  2XzxN\oBZAZ4hi! %kh@u;597;y0&FWu5A]]t76;j~!x,R@`H& U9amhvK3  nM9kkAZ:?#2SP<XkǕRf~(bBa*0ߞbF݇O;'>T,[tQ ;b.fk8#No70߂6팦}dDw = {~0Kx.0|ƨ'OFD!. U4iYL%=z29g~oxle;}H>)G,/b՛/MҼGĸH4i,UTv*U w |we'$tzc(ӥe94c?ċs)7&~2YC[Y,qL\#@.]ɑ8P<8(c(#zoe߯S)vea0.z_0&Ն1Q 3)wb&7`,]Y{ kjKKdQbAiH:65`  1fOS+S_ 㫥^QHTtX8y,KyTcπMHA7\u<7kQをLcaCc}9ghH]}n…P2dU̬t y= )(At=rh%bҶ1 q0Hf/lۉRX3/AsjD>8a%nFxn<Ұ]gX^0VCBW %FP1 ;hF;T>BZOW>ὬK{~smB D Z'K{,6|'ѝ$Q<i9ECGuV;^TXn lNb:fS' SyߪOx BH-\ 13/ I/4Ji<.<5Bb]!4rB)4Q ԗ."ldr<؈1u"@qr=;Rhr9s]$s=hh@)ˤ̞E.nlR=b,>IIRT>s"<ưbd4z j"0MUR~RL\bFŒJڥ k,l$9HX:;4c`4V.F4'@@|&s#p!pPpGC#y|Wzw=5,FL5EWaxJbm'CN8Pi*)+B!BJB.nxJ8$"! 獧@HLjÏORC\JGݝ/Rr &Ҍإ2[%U&+D|EC&_l\}0J+*ʫМ+c pȪN= c{"E}}[ڞ\JP4$LP ;g|8]`w]9ysAĒ^ty/.1/_Z?V+C:X(D4FBR@e*@!^zN5 (5wSmD T 0(#F[}2e3܂mat4Ecod889q§j&r&L<{=L"qt1ĤMo({qR8<5~ѩ`ց Nl8ؔr÷i`N8+2?K  y;<"r jV>!Qk|V~a^7f_<TTMm @]Eհ r”ٻ7Ztp&^C6:VNޕ\jVoyh ٯ+&q\qQ&Tzq5YlUo:M_8-;ZQN,Lq+l.Ԍ jJKecwMMttYn ퟹ16gJ&dܐ/x>}P֬~Ff^FZTwQ|6L]x Lte A,ޅǐGI? RT퉀ᤩVzNe gP{tB'42z؁Q{|oZ1JԮt9$<oK<$Uȍuřd܋{8|;tz2}Oʘob8i)RijeZ^tI ޜ7{, E4vʑ.y^͢tKQ%UYllgB2Z{A썎[."- ' p H3oֳ ϑbUN]űu0Rc$w4\$ Im VEO 1;vRճyTڼ r_`~6LqG&3u|zN"29{ɺJwJj%5񼳢Anp'f3SwAq ;LX4e`c`ʧ{֠o)M9PK(P٤-vpI_>Kd1G[׭==HaqڵEAbwkݼrXwk@5ǝMed513@9 ) %./]βuOT= :u|y9bsD1ӓҦg99p}.ęx~&5@7ń&ݴYnf0jL  dqA'3)-g%X!\Y 1s GL%b6C>VF{?ݏJhTO ׏.C_őIpo')!ؘM Gwx, 7!<b< |\y#ʬGp&i<&U6;y-s]c2KsU.$--'(A=we'$TKlŪe=2.b\&O`>l ֽ%8AUR\ &)GpćXJD|)oe߯Sinve0.z_0Ն /|'fkr<i8s4)ʩzw3K{ ;KKdbAiH:65 ?Y,2UbfH~#=Ka`1ꬽ(5FfV?5=h%weck^d+tTX'Q6$|'=)*)rWpG"wYE%hgl\a7=㸥vBB o׿Y֪z; g׉!UGvRF~غFMLxCח^<2~x; f^xכ%%J:>&^0Ʌ_'wE"/+k?YMyY2 '_-F _d1]5ax'{/Q(e[gẑ]mo7+8f|gQ8g+Ļ8 _ضveIrWIdǚH4VXX$OPbo꜁owy*n?m?>ֿq\}.>.~ [ xc V^r͊[>xmXΰ]_>!aWjW˞UcĉPRKӂk!nD(o_74u&ԆvzN0`"P vqirr>4 (ɶ'#JB7nzt]/U[cWfn̠WVQqȹ176KC@(P*<'&FWRTjsJCi%K4np"6F? QF5,3S,ͦFt ΂eӭ/I-ik&Exr?xvl{ ·y2+/icRR QȀO `&p0vg;N<8*zf`c6EVZچ3x:?oPP.̖7:ѷR7ŨopE) qw_jmՁ7ߟfDzFf8'W-޷j6xy7:lrƣZ+]vk[P@-Xv } ;ϼ_>a}T*ejpTqu[OSޚ&l:{s׻:Rj6ym-49Qz{;G7vdufR/_ӫ{|2ZT1ypkuZnӽ[qYtn-*bdocfkz]H}FiB '" qjZ@S$]M+GMMWjwu5 ^*+aА`D7{>_v;SodmAt~@BCW!qnC y+іo/59$u`{%#  ]| ʨ)Z8ڒ87iΊ E܈xs3MWTs4҃0ՁFq ݛ*!"Zrס3U_a48FSѦ狭$(Я}~&]{ڐ}](4>lOh Sgg!C4SИ:Ooy7<8л:mxb/g Cz3OB9D0ֻqn{7Fbu6Bې#L[OB9D[cJk\_;ʦ=gxp⎍nW/VMw^;/mux4UL gaυ@ ?]9+?2?Қy(PJJ+A38QʔJKלQ)?VRUңF)~(f܌\q{[J:Q*Y=P 0dOH})5F> LjRSCb(Z=-N:L*d(}LM^kAj`! ߬tA|EH#{~*~G;ww?Wv&8vI$$MieFM'x)(צeн̃anu9 '9 = 8iw_)KgTSABRƦ.s%^56|y$7%QJzj/בr3ꢁ^vfAiON^-6բ{. CHWL5ǂ0|]g94&i &ӈ}FKC/҅"@VG63&Yƥ--8E)MB8Y 2O N<5c Um$7:혗Q:l*dyi5MqBT Ҝ)m9)Feg*J õsm`vKS:Q %5ZB*Ɣ$4RIs80)1E.(GPHrT̖)$aKFH÷Zt;0 #NXbY}-gN^ZSgg"^/ӫ՟(͍r:G݆$SliKc+ cl`бB̨SRjrH2wJe5b($zNbw'lVd-OV>'[N3F.&iB Hf*OVk>H!TpdKKj[S  5T'{@Gݝ/EH!ac=xuwEqd_qϣJjzWsq @2^>+)+=VG=0 #cFil޷XTҋ s" _J!=1ܷT_bU t`%ibmR;:ĖkݟR1Zl{Lw~:Z.Je"QRފp=B3nt% \ Q@mQ~JZK9x .!Z?ݻwE Qg)iQ!#%D!JOs#`U@Qp%5\նa"]5`9rKx7z+.ʇX ;$ktww~QA/ ԗIϷ$&j۴a+wbYq2㗂8($RLDl7@I aCh6cDrL&;`p${N|TaB%M^/%J-NM42e2|n6v|y .{h5y/89NP$=]?0P"v59:,|<⟝w|װOZ9z(_?ƞhCFhB[ ZjA1t)uܕRCXTm#w6B_T`Ϛ] C4 S#6Ļ)8л:mxmݶM7JFһųxz.!)08n:Nl!mUFznDOޭ nP`+6i e<aRa1mxVJ`3m Q*h\٥ܜ&5 v2/SSe{^d-pKie ?#/=ɥ( :cV84uF)˙z4B*ĉH e5NTys1ϤFZ!tI%LV_%.iJ2*{y\Ҝ.IJ,U\4-96q=O3eKғJ``,{}Y $UG˺Ɛg CI { L.rPF`0\,d6+(4gJԣ™A*TH_ĄP$8?UJHI%BySS-QIѧE42F}i H06364(|F/-aeGĶx %P>.QwK WH9N70ɩd=ҽu3 :At r wj3sbjE'~A|{w5g~?m_MPkW/$!Đz6e%sh+ F c{z#Z|* \sB!SXH;c\|A*B _7Rͻ~ s/1/g/_R3^֙ez5緓'Ζ9.Qʙ-;)M%+"vzo|I4ͺ- U;@ w>w}9i&!Miw|BoGsM"$6 !LҾ &$kE6n=`.:qEa銁]QPQt2ulaiL㭋PY4RMџ)K3ͭH3j“,1j#_nS \RQ2юdc6 7 e3"fl,2/i$6k$x~ʰU'/N~)&.̳,!(Co4<"6h$U7w(UoWgSf۹;8# )Erwu2/?N&{1ŏKh^8h^8h^8h^4xA>jJ ea.rHwH_Km @>%^`?bl>v79;grC#ej$v_Xr?X_1,w7La[* i υr%PM4͋Տ+\?.V>[XR_t&+PDl_mM3{U&6;M k~BAJe~cD( @'RHZ7[Er 5$VgS"fԡ*,`^&5RT#)k48.tjs{6>m7G}^=ܯ6OU޸ޑ_6Yn޽خͨ* % ZJ:wق (a'dƮ,7) x_*k[:~7_?@>?#MEl&PrjȲ#ƭ ,-M#yDP u, [`CzcTqMqb -CA`B+KbY$ݷB6(o c0BZ*@P1BY5~NT,DhH2)QkS:#!B+#3UYy@|0:rZOЖI PJvݩG޶BBc-<5 6cwh8>0WXh>MAi!E&ßǩ>9^.;: E` 0e' E%BZD{<%m@ӂqS̃#-Tg -鬛cQƭT[t'¢&-b:,*EY_=ۇ1T5’ͿD8 `7q.]#PuyLBaQuGc0s*X0yO;r7](m:Y \%YZ앗uI4'`5~ <2^"$i1G^ BgesE ;&]泪N8@x(7BCpM)R?nIHH|8! vyL?Ih ų_g=+&IZD"`1걳Phd"xdW9:+;oT>&"7B8@fo!=Ci|o F孯G,4a'@ʛGk|*3F߽-sQ0fd dYL gF'|pGg%'|@wC_KQ:!2Ds8N|r$0=h(a1y-PLщO pbB9ϳj(8"3x 0aT222 %.c6Zz=K(@ c0;OtAHmfmŶcbXf?ۆ0YZՓ}ku8KޙһꂻӕdfL `UM걌|1k~0׉EI0>D^!6O"I2u)yw΁3 BLV|"Dqp$EBffb>jI)47Hgp%W,PRbNjbZWuf$;"2Y3|uE>;it2.$ y~"Rei8 WՊӊzp>h\Y9F"C+Nar /K(.4N c wK7/n{O߮{aMFݚc!V]Grk$iّkt> Bvb|H7䩽YJh?FʱBK3ZH%* Kdv^` LT\FiT՚S\WϛLR+BDY-- p!JArWz %,Pʡ$3U 4NByʂ1QR0ʵX! CL"m ϹE$|ΠΠyyg0@Fdr3`Xwc2 {~hHY^AG+|& ϩnݯg?I"!r~tYfR}g?q,.eA }}qފ\mqخZm~s4:߽;-{n~[I{-[ѭ?vnQ/qvEqDO*˿{3{3f;K,PHax:s=I=L5J;:ٽ Rp6/`3 @"8-F-g-}ZJ#@ n`_?Nz+[e_h&t0ǩ(s@tnr8gcQD͓$H>Cp)<Ȏٷ{) Ex\PLf<m݃낺C b^Њ2>9IIPQk s!?|/Ёw|a1Κ$Ygi)cJl"%0|8[4H C#N-T)pL1‚F-Rp$ ܐHIfސJ3Vٝ:V#mmw TD7WnȏӽL~m"s^/{|mD\SQ)eH"gy es-Ls,R%~|_IqO}l]Y^ѯӫWΕM3L~yoefS1%cqmY,@9cXQ9H[ R0 4bRY:GX 921Q;xNf ,t(l>HI\c;O_w[tPv4y}Rl P$6g2nѸnr)60:W]0ǻ2GtLΐIјt{=֙Rڰ,$ąAC]Y3*$c k‚,N]\~NNIc pe arc7$QK~+mZYf=_Sm ⪘]Q;zyM߬/Ul t]m_|*W.ĭw]"O k.[ra1)eTB]EO|dQ`Qֱ4 W$tsƄZuKAꤾu&n m U4I_?ܺ l5uKAꤾuqdTm+F"hJaK+ޱ`fGV'/8v +I1Žۼ1vEs5Hwq3w͍Z݊u{|^#޷hQH>ژ՚<|ŒyN\NDʿ{ےxf?߇T#Sjza wnk.⎾5Q8=0:Iw`[45K2|1Xf)6]-[Wt-[}rl@l$)$S@9M/TH48Bɭ+y|o1kfX!B˻/yP\x/>&Ё%ZllJK@&( -IGsG*L>yj!q1Y֡Sn6NBS-X$?vR7cM*!حƼ#KaҀߨ `z2LO1nmTHQ&vԨQt4-4&0 73qr;9 On? 0vOzB%ald"_|e$rM0 _`fsɑ K d.Ҙ@ ?9]ѼJ?lƳvjU)ewHN.hXf #:arrl֌E`ӐۆS/c$XOaW{QKWS R;}>Į>\}7NK6T`3\$rHt~iմO'MJT&Ji~Ztk.mvЗIRhԟ !8E(4,ݜ^(fJm̓`*mx'8J T;mǧ?ITxeMu{[ڸA}-ןj\G<7{7ނYY5, w)C=i;b̿do-~ǒ[GY}xd@Pn2KQU.^4O09ȄӉ_)1\ N5qe`v*<d$gTRdNa)] ZN_+k 3 旇O`1 *?%F^bt < '0;/q"X%I,A&+D~.FcZS(KqVe!&00B4uȒR! 3: g\g~cBʩc!s- RD(mZ$p0PFGjFL`J=9vzkEdpzN;@=w'a`-AUuzmͿ4ANvۺޤM'Iw4 ʒBNl} ie4rvZK|/ *Ƃhĵ`'gLII4oQBXJTI"9O-J*b^3vԇנþgL8)WN۠>\`xR)f wO< .=laI.˭@ 0(b__򝝯ɰk}:rvE3|=6筑٢3AJn/ fv(詳L8XI7R#0oqZ5Η]+ƒbt"yﺌ c<\&PEn[GDqD4k[#KXK^(z-n%xrrrr˺g(Q V2XV:yǽR)-Sfb(Os^fɔd&z?|^,8zG% :A.1_*i]\Nhv|I>H4>uu*&bӕW"68440?2F5 <1QEJ q.\Ds, Y GM,5Y3KUjy%0q$"Q (M6DD?2"S2„S.K%yyA~K&3[ߍ"I)sz-l2i|O'8)9g'__Ƿ~'~)`0@JMHy~7Y矁LE0 φ šlXԣac`ӧ_<=`t{T-S( /.{16@q;> ؜?/| TFGiKKP__8plӤV4{J_ρįxϋBڸ+ K@#p_ߓ\ oO.M*x?_)@'|>z1μ&շ7>?OˑfT-,9Նfpm @2 /J_!l-=ۤ||v_jY~~ L͛?,X9ç* nTY]q9Ieby%\+-Z-h[Ęk-.꠷T痚UXuPY:>!۝[=WUݦe/ZEZ Qq˴ t9"/. # ]bTбKGJ3x>seD{]Ji9s5}0|R#Cq1u%Yv) Rc$);yd2嵣\F|*R'4$s/2j9 7Ճ m2̛JU i檜wֲ{d+qЛNr\z mtE{䊵‚}Ѹ2B :|qx3ݷ i+#ꦋؕ,{ϝ*`4{)>ѳ ##S;5fEzt7_Moۙf@ Wý{W-/WԌkAP8evO{ߌ2{oP[nBoZ2ױ^([o]vWq0_Pe, Ӭ{9-=Ҧb@%Rm2t\.ȉ U{Gc!c7w:Ջbj5S[UfVG'kݵN焈-*&FfrN^2 SZ EF%SB:LojOt8S_)=g(0y< L5ܬJd0e Ii{yc41$-1q<>hׁ::ې ׻[Ę9Wې*q[;G@mh qM(Iʽ4I2, XiWrf"s2"R`SIךBci-H@2fTtӏ3zSFʺV{*[*[J +Hbz/H ea}"a_p29a(n~ߖ0 C!iP{a12Ԛ(h1>hGBb)%kjr4dY_Es)/^(tuv`+х}Qbd QЋ'Q*idO(Jǐf"|*$z`C-#HTwP+ H? 9d\,L䨀 N0rrfNuČ.\7Ć7wyFM33a!s{Q*-w.0K* Io1lf>IAЍå^:#)a JP[l(0,d!$ˢ6/=A鑡 k0GҘq;1"E΀+H:JqPdvVM(${D45rCAD:=lNz;Md{tg eʵ;GLb1O;#7|GI~52t`^vB u5IGpXQZk@B^2@u oܸIW?8uŇRM8>E NJd$E`kD#pD g2|uM3F~6^"|]i2M"(Q;H8 b4E`Qr5ll%Vi2])%C}$)J͊("|Y2„/U%B,;} _N[E v@NQp . :dYABF^t,y)Yľa+H-Nf6E.*mOߎH| % 1d\ Fg`hjNI{P )E'[D,4\P68a!.nсQ$ (ިC#(NK=$R ϰZyȈ&+k!W5o3dz|;22iER$Myqi h *$owELɚwq;/TsGߎL ;@b$zjݰu])"!!ךQnCjsnhjF`›,0C<kb2Ǭv=Mc*(ʳ/+EPB?{؍`iV$<` @&Y*-˲$n/oQvKr,YGJbPdxįE,SE9}=r_|B rDku!o9c$̬% %͂ÒJќ[v=<ŝz8dJ0wF f )4j"@,]&D wxTؚlBh]FcT^ :ȼ(E%qPyt,(9Fp[g 4_ 0Ȧgh?ާF8-, |1@fH1<#e]3j)EH7Jx'+\@B\r1Ʊw`OЊdQydc`y>q:fཌྷc;CkxCZC5q 6%J 5hIL[~l|{_\eK +\KD\!iPe|ʀABWl,n+nVj2 s hq8rVD¥kKyG}Y!sD<KqgJhy&>b/h=0]wMF7|M0vYT@` kȽZxZ\ 9t:ar5աEITISZ%xTץUdԾ`pM+.}b n9J4h%ςc騽q)q:LjD9xc^>6@!am4#NmcfH FOߴ)[v%=@َN|#;y>IOICQ ƺp؛͘-7O[umw |v=zyRĒʮ%%2e qsAND&]6CֲFAq ׂw'-<cʞt !gG@}2 Δbxvzޔj8 o7@;@R><ܷ+2nK %)OƤ ȵX U=`~4t}ur3j+vr"/LCN$.^Oހ{0^mLRDf9)&`\nOޛ6uk}Wod}vRAǡtz.FWEï x!0c`ɥg鿠:ڲ#2,h퀅5Qf!st!FZΓgφذ`'mJxbf;D8!"Ip*i:ċ9J-g1'EϳZ6j!pN~+N*^wygKdt ctcBIg DF[$b RH. xX1pBcn x5]퉬;iu#:~dx3N~?2X#@vjOc2<i&pgؔk鎈[۴kTڡǃ%*q<ȶ(B} q⤮{!t]*4Jm%;Ӫ,s-Ԯuŧ,MmjEmu,G[T/e@*:Pq,őǿSP KB0LT\TR6}kr3KuAr!OkDŭL"-<2̶0}nUј{Pwaݩ/LJ ̴lBGA9<;ފN{ghr)-kDL k"ϼq)QjKo9`#!ڡb.ll/g`o@JsF@nuno8.m9B ʠ dyFUtHKio,8OE#-Ede=g |0<^5e@ F鎛|k2o'_ O&CM|O &J.|GI.{qo5 i7,sõ~[!IFS\eas沬'sZj[gyN[kj{AfM d\̐֞%Lt+B.l# q߱@V,ֆd<ʂM+pk.MmJ2X&S+2q B*y8϶bELytM!SH 'Ev`n``r5Г#dHV씪 JmT` Xk )+wLd1Euݐ~g~>92|\;h]03m20U*]rq0_}F_-qib^װ/s@/ }؄6ab2(`NQJ:Ȅuu!H94}%&T,l ow~Nj(,H>MQF52{-2嶁rI"%ޛbI!fq6NjX@n"d\B4WVoى=EqmӖf~5ENJ GZ-r-*O?_L(f,$τLތavzOFj7lQc0ڬ]Tݺ"%N,*#I./F3V7Ww9Џ;OOKKl02?pR5a;/b)Lӓ6\*N6MW`u=sSZ82kSW*(v,)3~,m~>[Weڙǃ+zW$xzEcQ5 O;fs lJGX"FW=ƃ.xkW=v˃7ϲ'oHw˛77~hK7x>cL[n*wfjIzM!g)b! ;xoju<..&]wڈ@=RPYKB۠4Ģ뭞2ʐ1z|)Z짫>au}&B,|_ƙcҙ6iP`gJzqVGNrqlaŋ)I/uB"lu-A 'Gp]D[$(#CYrJ4ij_,.bvL9kR7PH (hlya%}JT=!' Υ-tr7 RTYDH*\4J~4TD9+[?#"nW*!H_n+gY-hW_&)-4 OAO//`7"7o˸\FYYYU^ ' Nh* >I`Gz^ŽHO:hl`9crC?5b(܋R5贗#.=Aw\tfMŜyL)sOe9S*e9;ÊpiIl"Iuc!Tsж9Cf?Jq4׈E{¡1*hJAۆRZywMmi;rT/mϤJ .V R'jum7N PlC(7 2.b\2 0JfTD ?5Ш8yb`xxigqs+og;[$ut/̗ "Zp3K)s4߿﾿D,ϨM_.]N`-&g ?_D+8NK*ٟ1Z1cNzv:IB>^W&7)k`_EsN΂SbgD\52B @8B^A;WlM`QlD3  i!$ =gCr7Hq|9 hW=O]Io P9MũI+5._h/g_ہ0B>fq+0 RX( /x.ű-^ݞkWYBhnj\S~@-@޻ t|4=um38KX{<`nJj.h[s]OaLQPTon^z/3&Հ9D1\ 5,~YfvO igJd y:Ncljh'_mj??gqqKI jײ-Z?wu~-go9_/?l@- 7P[)Sq2-Vڥ:&kk+-ן;8pxdT-+*W(ejOQotVqa -l'~;!r iADż1!\J5wn`s|aMmG:ázp-41w_dUyhU,2lO,wuK-g{E2Dm>t=糳*GO>O{wn9<%Xi.Fam֘Jl-k3Z?C+aiUFmUSg2EA[v9K%Mzۓ[qHᴓ_dܳ]}+p>* 1~;9]MX<tq3h*stʆ!Kd-z->LUN,ܜ^z08ryrmSBtR?n3%t:Lӵԏ}j9ڭ y,ڈ꼇PJen qP: 稥ݞmדiZ]5.:ZZW΢;TK ˄Ih3ǽ|{>$_=|5تm.`5[ 'XdK__nJKAsIgr2#h0rd ȺTY b<=r5P&W>j*!#7 ZJ{ɿ"'I2r;UJR\$+j "R/'z4\r^99ˆN|?wV}>ovbs}.)5`絖/ʳh{ shKUZXӼ9\a77zd(grؕ[y 0 ?WFusbnIgT \L;-eU jL&VFԒTmj!qW-VhZn\nUu/Yʻ 6A4`Y{P (WRu)koDL.U KJ'G-A-BK']->87/tzޫ{W.\7A[9\?ߎh**}9|K]x9rmSpPݚ⠄uQG=h*[ڭ y,ڈ1&]nMqP:M稣ݞOJFѼGvkC^9S15YjA?\S&r֫Pw ѱÆpwR88ls:bho *yְ%~R + +P*uQeXx(Y+JX\XZV-?Z LpH8  :!(H i0NTnBkAZ /\ldt^taz.ƅH3'1!FIkqRS62XS '/蟣;V AeHH Z @}5\E%~櫡6Cm 6c&1ߏVo};FS>Sdf5Wg׹—C &JJ&P?~XH@pw a\hm``X8\)0Bs>}2<KYD#yd4B|ks`meN0 >Y}LWW~g&xcc)YIL:5X{)HLN84j hsYc+,n6JS 5Ri{6LrZk4>}oyA-{5)L1`} B-8$p(()s(fkoű903Lod'U[_5ƥGͥpKeNm5LҲ~^PTVʴSf9TGyX 6E`mjk-kƦQN$C>:jd Qܠܩ,+!'7LG ҰD\&W 6vًxJP&Cً{hjQ {:!<ҫn 1=s-TɫƘG89*0"c4^AdR`O~ g5t2N q`Ȑ‰@}w9t/t+MABp:lzdBoc8TacdN ¹H-1eK*I=)T?| >O *5XhdbTHc/4}(PL+#"„9"P>8BĘ7ΙGe6~`ُ)K+DcMaC`1E- YO)ѧ!VKH2w GX@&cY(t#p,:rD`LG"d#rfᨃ {c?75WJ`?Dlj> s{AEBRl,) m>rk(euUksh^{./K%>zKԚ2'/__]Gh=FCLu1rdC`|EbtF'}2tk3Ck`j^uFے !x5*]%^/|Ƞ^ u-ElvUiO|<$raFC\ 9H)qE!"G}3Wkh_1mcH<1ˤZ/֢$leGj =#@kj`-T}Tk˕:ț,VWBZ(U-3w`d6:Ґ`,!A@FQ٨}Zʉ-ɬ3KUJcQ,V e#kPTUQIG}SQLie7ߨEk L&7R#S@ASK4N DDdo'LxSυ$>xL\LVenkZ z'2|] L|T=b1.hA"䎍/{mլpֵ:h?,`閧T~o(!͂ =7/ۺ44ڲ80-}\-NIƹII'KU<+yMh 6GڋD[R:yE[rYJȨe3 &&/L݋kGyntJh+: h6Mp۞`j{ڢE6\ٲ|2ї[,R~i=fwFbq7>LMK{.JZRG=M!즣GJ7ZT.s3ߚ L4dRэ}F6p 8`,ejZiYhQ}QljG@BU?[.Q`8j??V􎤛@iNǹԓuNO{fw:-ֳjy?'Tg'3Њt0 T1#Z=8rv. {m[lRgdgv6lw?v2$ؖ$Mw/@J2EQ@Pʹ898888/~=9dۻgޑPlf35_R@^xWnZ~WkF(0t>y5fYw5ۼhg1p+x^8 ;#;_,?E.K8bKY%`i$̼΀2/ ~*ǀ?r-3>Bj1RwvY":=bQ~BD [p+r(sdܙeL;+͖ZpA7 ND [-}*0LUqq2+#W1:jn:WA+L)>Mm'?j^>y3+`m'ĵ% S1yޘb쉙y.ܜnu7{S̝d|pkԤ/^]hcegg3/_?Mմl_Wif3\i1OѭPB*^\&$شښO3_^o@}O LUPRd I͍vS? jy!/]Z+n#!A$ȞKݛor{pC7f1$ЊU?:OG8G 7 ~({5=s̃X}ଙ:pei%MVB9jAq*vE_haҸiB]ΘP`*1yAbTrjh68&BC~ b3ܩlVq0Yx12`шC"ݳ 2JAn䀋"RC2gc, EjdS>J .sGuNR$) Rk P*ÇpW|}סO*}UZjD?T;`7XwJT]rr B6Eo߬? "_M?o}`[_g?^_L|\Czu>Yge|\q~g쇇;7Opgn+~P, = @LNA.&Ʊ>0!!<cF䦑i>fϜ+b<$182=Z>~L,_ @g! 䙤gNCiLrƌɪG2 46ytQ!{<!7'Kq^ MAo_>6qm1>%$R<#_Gۏ6h=2UƜgWeήy`R GY[ua6?fu8,L3g:WO(3w3Sm͙n߅EwKZP(F= GșFɿ<=C%rv>`z;3/E*rBߟs_JĔosbeΉcr3Z'|sF$bj4EC|+ *wZh0X&6W)QPpKO#BJ 0US ֫ )˩qhw :J:78mxHk"5[ 9a@ `t[#d:8R= yYHX7u*3\rͽ\.r+P` !+UpV8̅\QSp*|1YPP9':?2=f)Z$9R i|dL;5<$9kHΏCA O}l.(0 ) K@ =Za*Cr I(J Aަ)΍R/m-a˟W[.VRwnspW5fGBK .$=꒳&bk~ S`6?0 D4Ko!䶰ũQ}]fk|m(f^sy^+ت<|^,_R 6b~._Jޔ+:ߕd^ΠsSa'DH|`#%Ƨ`?EhG9GH>AۛUVzGfak4xOy6EpK̄]}>|1Β'^^6 FXRӼ]5a ݱ?y\@ҥe1JQh}B0b{vA( ?c{! a.4YnN\vr|~u1/_fso8jeVm=誉US2?]\6oAw<3(!w":íuˍt`wnsQ2rwzƙnTSE>eYOhLQ;͙[*!6x*n ڭ E4E'oqh7kmƴ1ex lPۭ E!SGp [a!1;Twv#%AWP:YTMͳkk/ōV]^\3"߿ћCRviJ'<;F 3H};E|$ 928-ltt4>"LNB KjqcutD Kdh8H:ᰖu]Kz. Kg֫@XyGv,T|*X/p8S\gʇdA\ءF% 1%$rg]ȩg}CQQ>.$0q9jzV|(|߿6GE!ps !2Bʐ,xtrxբCZ\tDdZ/ᑅA󎯀C?{`58}pa(%z֓=`*bl*4րA@{|Ne9/x!Gpލbŀ\lKGˋ` {,g@v>Dԝ1 X, (elnIKf d Lrn j,# DtRHPTP87`99DQ,v(Z#{'F`K 7迦W m9S 8핾j*}աJ_S&iu;+vmD8vDES!nЊ]5ޯpl{7ˀ5D Ƞr#;@P|5v\[O9%3c?(ꯤ+[VWvU할{kl+꟬9*K/.2  7|PݷPź}酡b9)K}uV{ٸ=jpm)U#P4ܦ|4}䐨g?{F$)M߻k$X$8yJ(Pm =$E]55,K7竮kOUXmT@⨞QRxEoFX#?X{ r=-|:w};@B@5|{q+63"5s>$C &:3)߮ϟ5^◦ž38ٜP9OVqq\Κ}խVy0O>!s)ɻ)&5|@==쭥ydM+(R+Pc̠EH+~;fjSL+X!kaxj9$z?$H=FI gbA譵;h3ei[%A0¤d,+νѮdoJq#B,'EӫW-bYد5g ogNUnY3r{0__ЖYϼ/[\ˍdՓv#8H9lκbA>lۧsQ0UO׫/K]V>^G)5 tIk=](y?#Wxc?dۧ|^o W>On͗5aUn6ҧҜlԃ1۵4nfe_bú_OkwݔKK0ͯ~0m%L[ V´O lC|ZHsUvVoFH[LK~(k%)@-V K&F1ә%JzQJ;(in =%bY R&9<b)z¹$<:bc޹!g~/ *﹃PBD-|v;#kU:Ӈgub鱆5j_IVN*$_] dAa$ P6'byϋq vN?~>nO  (e͑@^ojNl:tjD[S3` \Uy}X}Hhc>.[bf#t)PڃA%D4/gUv(P'2c|VR^Lj}gda6@іG 2nfIH.6Csă)c>21O뼪-bڠxљeN $k"5.s]dD ~ Pt(" WB]7q߂hzB4L+!ďXο~ŝrd͏Փy(# k@li"Pt\rX8b(atncR uL]mَhx͛.8p1攛cs`~z0Jd#Z G/Vs1E(ujMjI)40ΏEy׋Hw8[QG(U&o/Y]dUs܃飹c#'ͺkZ+RB4f*VF/=d ? `vȌP'3 hq|Ϧs`@.Vmɉl1zm b D% o}?zn̘ .~3x*jvJPsþI-@dd4VBlx>B|P $?*u#RA'BKp"Ta61juGaRM[e;&b 9:\MCK"c*`(':M pE6C%,iS 2~"֒%`"TIs,:íɇR:Y.Q$l)X$D|L*9zɁԋXZ4ͼ6'm9SFsvbZ9F>֪B/-33X( xm yu)šW7\8, M֍:Հ9d2"J$=-$b+e}CI92Nd%Eb]Q"t"KELƂeaB$iR2{|ZXSHnVc }ϋBR$ejF:Ipo8! 9#5kO|O-w[ x1 HJ c}&}҉=A]{QW͇ˠϥĢrli6H[)o@ٔ(U"Z\_uAN#'ՕJ4ƓEҺ1xdyO*f͚cwW{ъ E=C9?CkX/OY2(cGC[+,nz,r$$ #ICK0'"yFQe3{<}܅/h85$2ΒrU\pB^Ao(9#.sX͹ r#yJ!yhv*#hzM Lh۩  ?%;JI^H{ UTJ27ߺ)2}TСc򳇼Aw&ε[&GjiIaۋQo?>!cojiK8 Cqћo9ġȬ;nG'-SD 0`=r!^!suir(5IԘ1LUb x1޳b7F 9lNzvsv#!Ύk?f095 U9$(DM⊫cYM۵'٨$WWqqK Y}Haؖ@+X RLj{Ɠ5͍׊][t!a>q'Ij/EQ3:6h4sc/}veEZ넧!OJz ƺtyu.rCB.ubA 0x00}89SsƔ"yfx\󌓽DTb%nYr~Jd5S͝H8z$u7o:B}YP@b9;Od IubX6M<2J 5vOeN_ WeXPARdL L%Wk ޳豕r^a#}r13_[+>2'AXŢOa31Vڱis;+Fvy.ZIF6D'fHޕdeu(M,cK]l~?y'鿻JXA{.ٽG.{/(ɛ?j(z4L(ȶIp11pugg-7B ,< #M*Γ5Ɇmֳ˄kOI䚍?!@o꒺)^6HUPkPsehS8-$;n5_\cyB`%R,osR6M^MUK2(O2pN; .WLYG넠)E`킡DLs-.lyWG,_=}%[qltםҺ+ē-y'Q&S6% Tivj [}2z 5>9*/2{Z]6rA_:*$բ )x+!] FY2(2}%ӽ)oNsW-w OrzE~#$M=x@yk_dКgaq; c|D2 N/F3?67䯫ᔐrꞏ(d [vzI`3Y4UXL[ы5hoeQO}Ym7+jtA0wח7m .+ v:DD$V6ZhSbV75"9~|⢄Vgfy ǘmX) L Gaw% *j *Е?#YB/gݦ~1 {6qbWI$%>8T -i,Ȝɘ![_ݺFJQxzۛ5Пg܈]9mtI%r~6.2"s$?LAZ}]a2Yo^sT`LU:Ff4pta>X=7r*nfj B,PDURg豌MG2k" 6HiaR)CcwL * Q`KxpmoJn46$E OG X<$^nP[.-:IsgPEtVByGDcn(y(יwòQ+9RoJd~}X]wZ )=0ۊw[u3VF4QZWl3IW2b${քh!mW0e46b HKe5'*qĠ D9X$zw,YtawXz*2&krJHk, >)BsJJWP]߿#F@[X8.*RHG'"8 ~0iWPFK-:XNiJd2HK\CksW1 z6Cw5`%\q v vI)Cl D6$h *R,az|b@UR\fiQYy"9(w^&“i,OQF]^@|aOsx N0q},wEk׽==Z,9_R!855r e_pa"|M4 ʫ:5i|35h&=_?u\AmS߼:^LWr2&fwe+Gw9^;rY'~9݁ȝKlLH X%4 f:z{Qz0 =-E4}FtɘO{8-Sc>#.}v9š{H˰nc^,o`1m9cc%"yQ̮elI~ר1-J\ͤﮕn=/9VYuѸ:N 2@ B4f6X\Y[2rcfA%q+?8E]rm;&Jn(h#Bا ?i>k9MͳYVٷ y֧Zsv 7a$[ո2)T?<]/藓,?~{?^y xIH%䔤 !"Q'dQJm?2+oYzARlBb#< (ZyFafXQ@XV\PmC=)\c B[Sլe^X1jzzU ]\'ގr5+uwӓ,!~F4I#N&w~7/j<e;E_,o~#j:q憈BJ{%Ԋˌź6K[cQ};Ż[+2C J>]9nb/6$_rvz5ܪ%Pp4ܷ+T:NK1N`]j;ez#s m 5G~]4sqhk,Z k'f~3 Tj=qţ#[:J+z{,i<4l4f U1X(WsTRWj|YM [[RFVc"<;_F6J.٣߀wYh)D!a(gHyCt08y/=ף_9ǖqjk}8k+9/-wwgUoz+,r21+|+8c %dx"ˡSǏVHÅ4|D-LwJY-vP !H 6Y i3uCm _sa>^êZsh-;a7ᬭl\>X}ܢ!pwR%~XP?%K*D}7[FniR\2^Z|`տzA q?kgK֑\U.T  Bk]6,h:,D~p~yyA5Ed\ o*I](\΁.[,`5?l)iED-[jY*-Bsj:vU9"8ۺQ6@һr2%)D RX$VI<^?z%m_ZFm=n2ʈ(\J &,yjCrE`|t[ P2XDee 7VT{5TXD,qb v>2l ؀w\-Q6Zk\sBm'\sSB IV ~Vw.a ՌRvB fݴqח]'!;jǗ>m\P_\-G쯦[ZCu՞n Kc1SvY8O@u!z *˻iS%sXwHGܒ\kwփ mkxV)v޶д1JR2#}*1Xeё|7gre#:Nb5kY4&l=w JyR!d&g|&meEt(GͅHb(&Q6tQxr$U9?dRQZ]!AO3Q_R&eJU2E去eTJ>D~1@BFA=| D< bh2BT̒-[vS?k PL814Ӈ6ҸÉ^Kbu8vP ;U)iAML%&ei)222(M%#IX(rR2Y`Yr6XVYUW!ZB Z='j=!$$bȌ$ ;Dc:H)˞"-i!^ DP"l!IR9d*7S8's4L\q/"ZDCd&alZPP{)uRH]+[5L=B(yb-Tq2=TUJg쓻t|@VrħKezupmJsW)YhtF0]4;#ݵLS+| aYv,̰zcGYnQ˅:HLP圊`"~EIQ &>ikFR#yNlM%Y %B,f,+VU*7#T:J0r ݵѭhӪx$JIc2;{1NSc+WB@PXp \2cR`Jl1޶`JF}T˕^]UXN*Wsuml.pkN[ʖms8[߹hzO.<Rw-$s(q[JY> -ܶRnQJՎHY> ? ( Tݪ8kvhF /}>^udO?l36ήWNJ), x$k-qu5 v!{yI{mH/)muIԊFH /@zu^wm /FU_3EpvOI& 3:Ȇ,M2俟jRiJQL5ăh& IZ|rv GRLy,sHIٝs_ õk7|\~d'I[{v뻊[k}yxu+-M1lM4*-ie;4h=![@Roa-?A&5E;KR3l絴1xYd"YdaO(__Z7Jv}m3C>y?0Tq8& /{QHyN|N2Ify t(:gF\:Z*< 6`0J0Z0C&=1,*)4#JYPRZ#mjϙ6!6(bsQ}^ZR_csEJ)K@)R5yyzA@+]>w]E&u QW )e@ա P םrka.w| c.Çۣk ĖdS9)"^N JϷ|u}OjekG30Ĩ5k..ӥ!2zNIPfE+N'nU:#ڮMV緷CDqd%eۿPZL37Q~} ;F:FBѼ-nUvuy&–_ʷ'6Jhqdr,oS 5-$73 pK(sF|cmO2"j IH߯~6nPsƠ( 1Bx{ 3;r}k*@^le'noy' D3&RSb5_\ZzMfWN'Q) f~;5YZeM2DR\̦W^uޜˉ-yjGizn.Tܷ C _}b(&- Z+@jY'75vS[ogzrhh6Bb<`PL^`?4j bo0 8J5 F<!G$|xs|{'oWK(%cYF6+q{d|^wl^~c'_O1s?+|0-έ>Nmzp.n'1ޡ+G #Čm6㝼?78 "TɌT2;sWzauo觎 A͠d b4!2Sm)q*P2'HTBBJ҂jK#R)!!$|J3i 4cLSޥxbEɹ i.RLt,{Bmj!$,:Q+Ls[,q2i٫'}~뢐{*X?: L#UיTk Z _pU-#Ed\fm3rqz5l?HRvK`#[ׇ| @R^JS< A9Z@ 1]Q)tEy{7r8OWO$ Gf0EKS\yV)d:VT瞳PNˬD8ϥʡY9rM:wl }ɲ7{Ug4耡!o1aQ=ԥ t=q0G8%ڋk*UԘ+ZЈsprG.0rFI%dqinh'|eS?nJ,y\ ո_okMC\$CJ:7C61RQ~?ڡǗ*3\gS{:iP:CY! NR̘/CkfpE!~="% ©MRy%Ҝf s2vPpw/K]Ÿt{_'h*([4R`Rnt@(*o$6+4Bژο05lxq\Toţw*oLL[8\ZKr2*(wQ'Xf-'7\% cbCx bG]lHDCeu^;MFWbp&_ߚ"J!/`0/qfs2S,D{QVue({oی=BHlt1B6haSB ' x<6 bm>TF+4Bx*(0kn AK1]P 7F&ZB t pFБ"^\eI3O.M=;%@O/>>^氎3Ȥ|\ym z`?3 vɠM`Ob@d4 p2Ģ6\F\}O Zp]+]4Z¢'/h%]<`g 4팢w-7;-߬.1y@q;Aۧ#70v9 tCB%l>7 ^ڳ0 GR%vѲ; J Su<%3=|F3x`uu8ظ+ov=U3>Fq ZMQC3vgbݖ!n _ezxW[Wҍ4&"^iGw)襋L#2fh%o믶3eX'`ɔڽ?;}(3 {D。5 T$xOh"ZLXˆFԓߐ^i^ڝ3S- }v7ڀnsٟ4j(-ـ5ĉakC|( M/NޠI;4:#DJ)$QÆ G3(`GR\*E"d$LNaš[ƝvGAnRPxi(T24_0[1{ͯwYQ3V#`!*n&mjzxQn!i?j/GeL-nok4(72ڻ}P?p$##í}Eӈ8\7O%Sй=1 NahԱoe*oޙ]\yS9+fFmBbesyT O`^oPm*Gh4z #D17C|qqvݖ=T׶C&{{P4hX^.RO1.^W|%aBapwoYf[©w]CRa x]+(pNnE7bn÷2R֓$䙋hkR x8_c1#Ȉn%53z42Bh Q5.vPd?e5Ɓ,%V$4cN>OJ3VBtL'R*h롱0ZxRd4<͌ׄ@K>>7TX\t6Tvar/home/core/zuul-output/logs/kubelet.log0000644000000000000000004666305615135726703017726 0ustar rootrootJan 26 16:47:12 crc systemd[1]: Starting Kubernetes Kubelet... Jan 26 16:47:12 crc restorecon[4689]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Jan 26 16:47:12 crc restorecon[4689]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 26 16:47:13 crc restorecon[4689]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 26 16:47:13 crc restorecon[4689]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Jan 26 16:47:13 crc kubenswrapper[4754]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Jan 26 16:47:13 crc kubenswrapper[4754]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Jan 26 16:47:13 crc kubenswrapper[4754]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Jan 26 16:47:13 crc kubenswrapper[4754]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Jan 26 16:47:13 crc kubenswrapper[4754]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Jan 26 16:47:13 crc kubenswrapper[4754]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.607117 4754 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.610214 4754 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.610238 4754 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.610242 4754 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.610247 4754 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.610251 4754 feature_gate.go:330] unrecognized feature gate: SignatureStores Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.610257 4754 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.610262 4754 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.610266 4754 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.610272 4754 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.610278 4754 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.610283 4754 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.610289 4754 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.610296 4754 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.610301 4754 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.610305 4754 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.610309 4754 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.610314 4754 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.610319 4754 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.610324 4754 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.610338 4754 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.610343 4754 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.610347 4754 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.610351 4754 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.610356 4754 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.610360 4754 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.610366 4754 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.610372 4754 feature_gate.go:330] unrecognized feature gate: Example Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.610377 4754 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.610382 4754 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.610387 4754 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.610393 4754 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.610398 4754 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.610403 4754 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.610409 4754 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.610414 4754 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.610419 4754 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.610424 4754 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.610429 4754 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.610433 4754 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.610438 4754 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.610443 4754 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.610447 4754 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.610450 4754 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.610455 4754 feature_gate.go:330] unrecognized feature gate: OVNObservability Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.610459 4754 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.610463 4754 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.610467 4754 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.610471 4754 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.610474 4754 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.610478 4754 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.610481 4754 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.610486 4754 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.610490 4754 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.610493 4754 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.610497 4754 feature_gate.go:330] unrecognized feature gate: InsightsConfig Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.610500 4754 feature_gate.go:330] unrecognized feature gate: NewOLM Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.610504 4754 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.610507 4754 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.610511 4754 feature_gate.go:330] unrecognized feature gate: PlatformOperators Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.610515 4754 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.610518 4754 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.610521 4754 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.610525 4754 feature_gate.go:330] unrecognized feature gate: GatewayAPI Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.610529 4754 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.610534 4754 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.610538 4754 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.610543 4754 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.610547 4754 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.610553 4754 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.610560 4754 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.610565 4754 feature_gate.go:330] unrecognized feature gate: PinnedImages Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.610661 4754 flags.go:64] FLAG: --address="0.0.0.0" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.610686 4754 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.610696 4754 flags.go:64] FLAG: --anonymous-auth="true" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.610702 4754 flags.go:64] FLAG: --application-metrics-count-limit="100" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.610709 4754 flags.go:64] FLAG: --authentication-token-webhook="false" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.610716 4754 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.610722 4754 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.610730 4754 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.610736 4754 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.610741 4754 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.610746 4754 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.610751 4754 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.610756 4754 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.610760 4754 flags.go:64] FLAG: --cgroup-root="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.610765 4754 flags.go:64] FLAG: --cgroups-per-qos="true" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.610770 4754 flags.go:64] FLAG: --client-ca-file="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.610775 4754 flags.go:64] FLAG: --cloud-config="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.610780 4754 flags.go:64] FLAG: --cloud-provider="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.610785 4754 flags.go:64] FLAG: --cluster-dns="[]" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.610791 4754 flags.go:64] FLAG: --cluster-domain="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.610796 4754 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.610802 4754 flags.go:64] FLAG: --config-dir="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.610807 4754 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.610813 4754 flags.go:64] FLAG: --container-log-max-files="5" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.610820 4754 flags.go:64] FLAG: --container-log-max-size="10Mi" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.610825 4754 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.610829 4754 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.610834 4754 flags.go:64] FLAG: --containerd-namespace="k8s.io" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.610838 4754 flags.go:64] FLAG: --contention-profiling="false" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.610843 4754 flags.go:64] FLAG: --cpu-cfs-quota="true" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.610848 4754 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.610852 4754 flags.go:64] FLAG: --cpu-manager-policy="none" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.610858 4754 flags.go:64] FLAG: --cpu-manager-policy-options="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.610870 4754 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.610874 4754 flags.go:64] FLAG: --enable-controller-attach-detach="true" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.610878 4754 flags.go:64] FLAG: --enable-debugging-handlers="true" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.610882 4754 flags.go:64] FLAG: --enable-load-reader="false" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.610887 4754 flags.go:64] FLAG: --enable-server="true" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.610891 4754 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.610897 4754 flags.go:64] FLAG: --event-burst="100" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.610912 4754 flags.go:64] FLAG: --event-qps="50" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.610916 4754 flags.go:64] FLAG: --event-storage-age-limit="default=0" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.610921 4754 flags.go:64] FLAG: --event-storage-event-limit="default=0" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.610926 4754 flags.go:64] FLAG: --eviction-hard="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.610932 4754 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.610937 4754 flags.go:64] FLAG: --eviction-minimum-reclaim="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.610942 4754 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.610946 4754 flags.go:64] FLAG: --eviction-soft="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.610951 4754 flags.go:64] FLAG: --eviction-soft-grace-period="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.610956 4754 flags.go:64] FLAG: --exit-on-lock-contention="false" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.610960 4754 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.610964 4754 flags.go:64] FLAG: --experimental-mounter-path="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.610969 4754 flags.go:64] FLAG: --fail-cgroupv1="false" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.610974 4754 flags.go:64] FLAG: --fail-swap-on="true" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.610978 4754 flags.go:64] FLAG: --feature-gates="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.610983 4754 flags.go:64] FLAG: --file-check-frequency="20s" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.610987 4754 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.610992 4754 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.610996 4754 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.611000 4754 flags.go:64] FLAG: --healthz-port="10248" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.611004 4754 flags.go:64] FLAG: --help="false" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.611008 4754 flags.go:64] FLAG: --hostname-override="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.611012 4754 flags.go:64] FLAG: --housekeeping-interval="10s" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.611017 4754 flags.go:64] FLAG: --http-check-frequency="20s" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.611021 4754 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.611026 4754 flags.go:64] FLAG: --image-credential-provider-config="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.611029 4754 flags.go:64] FLAG: --image-gc-high-threshold="85" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.611034 4754 flags.go:64] FLAG: --image-gc-low-threshold="80" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.611038 4754 flags.go:64] FLAG: --image-service-endpoint="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.611042 4754 flags.go:64] FLAG: --kernel-memcg-notification="false" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.611046 4754 flags.go:64] FLAG: --kube-api-burst="100" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.611050 4754 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.611054 4754 flags.go:64] FLAG: --kube-api-qps="50" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.611058 4754 flags.go:64] FLAG: --kube-reserved="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.611062 4754 flags.go:64] FLAG: --kube-reserved-cgroup="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.611066 4754 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.611077 4754 flags.go:64] FLAG: --kubelet-cgroups="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.611081 4754 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.611085 4754 flags.go:64] FLAG: --lock-file="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.611088 4754 flags.go:64] FLAG: --log-cadvisor-usage="false" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.611093 4754 flags.go:64] FLAG: --log-flush-frequency="5s" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.611096 4754 flags.go:64] FLAG: --log-json-info-buffer-size="0" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.611102 4754 flags.go:64] FLAG: --log-json-split-stream="false" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.611106 4754 flags.go:64] FLAG: --log-text-info-buffer-size="0" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.611110 4754 flags.go:64] FLAG: --log-text-split-stream="false" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.611115 4754 flags.go:64] FLAG: --logging-format="text" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.611118 4754 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.611123 4754 flags.go:64] FLAG: --make-iptables-util-chains="true" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.611127 4754 flags.go:64] FLAG: --manifest-url="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.611131 4754 flags.go:64] FLAG: --manifest-url-header="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.611136 4754 flags.go:64] FLAG: --max-housekeeping-interval="15s" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.611140 4754 flags.go:64] FLAG: --max-open-files="1000000" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.611145 4754 flags.go:64] FLAG: --max-pods="110" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.611149 4754 flags.go:64] FLAG: --maximum-dead-containers="-1" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.611153 4754 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.611158 4754 flags.go:64] FLAG: --memory-manager-policy="None" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.611162 4754 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.611166 4754 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.611170 4754 flags.go:64] FLAG: --node-ip="192.168.126.11" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.611174 4754 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.611184 4754 flags.go:64] FLAG: --node-status-max-images="50" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.611189 4754 flags.go:64] FLAG: --node-status-update-frequency="10s" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.611193 4754 flags.go:64] FLAG: --oom-score-adj="-999" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.611197 4754 flags.go:64] FLAG: --pod-cidr="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.611201 4754 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.611208 4754 flags.go:64] FLAG: --pod-manifest-path="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.611211 4754 flags.go:64] FLAG: --pod-max-pids="-1" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.611215 4754 flags.go:64] FLAG: --pods-per-core="0" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.611219 4754 flags.go:64] FLAG: --port="10250" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.611223 4754 flags.go:64] FLAG: --protect-kernel-defaults="false" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.611227 4754 flags.go:64] FLAG: --provider-id="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.611231 4754 flags.go:64] FLAG: --qos-reserved="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.611236 4754 flags.go:64] FLAG: --read-only-port="10255" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.611240 4754 flags.go:64] FLAG: --register-node="true" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.611244 4754 flags.go:64] FLAG: --register-schedulable="true" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.611248 4754 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.611255 4754 flags.go:64] FLAG: --registry-burst="10" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.611259 4754 flags.go:64] FLAG: --registry-qps="5" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.611263 4754 flags.go:64] FLAG: --reserved-cpus="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.611267 4754 flags.go:64] FLAG: --reserved-memory="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.611273 4754 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.611278 4754 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.611282 4754 flags.go:64] FLAG: --rotate-certificates="false" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.611286 4754 flags.go:64] FLAG: --rotate-server-certificates="false" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.611290 4754 flags.go:64] FLAG: --runonce="false" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.611294 4754 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.611299 4754 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.611303 4754 flags.go:64] FLAG: --seccomp-default="false" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.611307 4754 flags.go:64] FLAG: --serialize-image-pulls="true" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.611311 4754 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.611315 4754 flags.go:64] FLAG: --storage-driver-db="cadvisor" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.611320 4754 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.611324 4754 flags.go:64] FLAG: --storage-driver-password="root" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.611328 4754 flags.go:64] FLAG: --storage-driver-secure="false" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.611332 4754 flags.go:64] FLAG: --storage-driver-table="stats" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.611336 4754 flags.go:64] FLAG: --storage-driver-user="root" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.611340 4754 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.611344 4754 flags.go:64] FLAG: --sync-frequency="1m0s" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.611348 4754 flags.go:64] FLAG: --system-cgroups="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.611352 4754 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.611359 4754 flags.go:64] FLAG: --system-reserved-cgroup="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.611363 4754 flags.go:64] FLAG: --tls-cert-file="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.611367 4754 flags.go:64] FLAG: --tls-cipher-suites="[]" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.611376 4754 flags.go:64] FLAG: --tls-min-version="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.611380 4754 flags.go:64] FLAG: --tls-private-key-file="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.611384 4754 flags.go:64] FLAG: --topology-manager-policy="none" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.611388 4754 flags.go:64] FLAG: --topology-manager-policy-options="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.611392 4754 flags.go:64] FLAG: --topology-manager-scope="container" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.611397 4754 flags.go:64] FLAG: --v="2" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.611403 4754 flags.go:64] FLAG: --version="false" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.611409 4754 flags.go:64] FLAG: --vmodule="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.611414 4754 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.611419 4754 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.611551 4754 feature_gate.go:330] unrecognized feature gate: SignatureStores Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.611555 4754 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.611559 4754 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.611563 4754 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.611567 4754 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.611571 4754 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.611575 4754 feature_gate.go:330] unrecognized feature gate: PlatformOperators Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.611578 4754 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.611582 4754 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.611586 4754 feature_gate.go:330] unrecognized feature gate: NewOLM Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.611589 4754 feature_gate.go:330] unrecognized feature gate: PinnedImages Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.611593 4754 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.611597 4754 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.611601 4754 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.611606 4754 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.611610 4754 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.611614 4754 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.611617 4754 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.611621 4754 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.611625 4754 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.611629 4754 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.611632 4754 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.611638 4754 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.611643 4754 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.611648 4754 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.611652 4754 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.611656 4754 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.611660 4754 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.611677 4754 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.611683 4754 feature_gate.go:330] unrecognized feature gate: InsightsConfig Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.611686 4754 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.611692 4754 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.611698 4754 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.611703 4754 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.611707 4754 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.611711 4754 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.611715 4754 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.611720 4754 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.611724 4754 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.611728 4754 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.611732 4754 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.611737 4754 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.611741 4754 feature_gate.go:330] unrecognized feature gate: GatewayAPI Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.611745 4754 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.611749 4754 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.611753 4754 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.611756 4754 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.611760 4754 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.611764 4754 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.611768 4754 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.611771 4754 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.611774 4754 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.611778 4754 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.611781 4754 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.611786 4754 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.611790 4754 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.611794 4754 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.611799 4754 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.611803 4754 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.611807 4754 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.611811 4754 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.611815 4754 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.611819 4754 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.611823 4754 feature_gate.go:330] unrecognized feature gate: OVNObservability Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.611829 4754 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.611833 4754 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.611837 4754 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.611842 4754 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.611846 4754 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.611849 4754 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.611853 4754 feature_gate.go:330] unrecognized feature gate: Example Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.611866 4754 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.620266 4754 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.620286 4754 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.620365 4754 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.620372 4754 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.620377 4754 feature_gate.go:330] unrecognized feature gate: InsightsConfig Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.620382 4754 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.620398 4754 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.620403 4754 feature_gate.go:330] unrecognized feature gate: NewOLM Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.620407 4754 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.620411 4754 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.620416 4754 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.620421 4754 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.620425 4754 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.620429 4754 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.620436 4754 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.620440 4754 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.620444 4754 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.620448 4754 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.620453 4754 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.620457 4754 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.620460 4754 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.620464 4754 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.620468 4754 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.620472 4754 feature_gate.go:330] unrecognized feature gate: PlatformOperators Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.620477 4754 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.620483 4754 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.620489 4754 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.620495 4754 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.620500 4754 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.620505 4754 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.620510 4754 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.620514 4754 feature_gate.go:330] unrecognized feature gate: Example Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.620518 4754 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.620525 4754 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.620531 4754 feature_gate.go:330] unrecognized feature gate: SignatureStores Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.620536 4754 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.620541 4754 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.620546 4754 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.620551 4754 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.620556 4754 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.620560 4754 feature_gate.go:330] unrecognized feature gate: PinnedImages Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.620565 4754 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.620570 4754 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.620574 4754 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.620579 4754 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.620583 4754 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.620589 4754 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.620593 4754 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.620598 4754 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.620602 4754 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.620607 4754 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.620611 4754 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.620615 4754 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.620619 4754 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.620644 4754 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.620649 4754 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.620655 4754 feature_gate.go:330] unrecognized feature gate: GatewayAPI Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.620660 4754 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.620683 4754 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.620689 4754 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.620696 4754 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.620701 4754 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.620705 4754 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.620710 4754 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.620714 4754 feature_gate.go:330] unrecognized feature gate: OVNObservability Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.620719 4754 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.620723 4754 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.620728 4754 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.620732 4754 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.620737 4754 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.620741 4754 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.620746 4754 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.620750 4754 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.620758 4754 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.620912 4754 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.620920 4754 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.620924 4754 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.620928 4754 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.620933 4754 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.620937 4754 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.620942 4754 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.620946 4754 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.620950 4754 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.620954 4754 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.620959 4754 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.620964 4754 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.620969 4754 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.620973 4754 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.620977 4754 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.620981 4754 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.620986 4754 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.620990 4754 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.620995 4754 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.620999 4754 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.621003 4754 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.621008 4754 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.621013 4754 feature_gate.go:330] unrecognized feature gate: PinnedImages Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.621017 4754 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.621022 4754 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.621026 4754 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.621031 4754 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.621037 4754 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.621043 4754 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.621049 4754 feature_gate.go:330] unrecognized feature gate: SignatureStores Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.621054 4754 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.621059 4754 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.621064 4754 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.621069 4754 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.621074 4754 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.621079 4754 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.621084 4754 feature_gate.go:330] unrecognized feature gate: GatewayAPI Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.621089 4754 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.621094 4754 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.621099 4754 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.621104 4754 feature_gate.go:330] unrecognized feature gate: Example Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.621109 4754 feature_gate.go:330] unrecognized feature gate: NewOLM Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.621114 4754 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.621118 4754 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.621123 4754 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.621127 4754 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.621134 4754 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.621139 4754 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.621145 4754 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.621150 4754 feature_gate.go:330] unrecognized feature gate: PlatformOperators Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.621155 4754 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.621159 4754 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.621164 4754 feature_gate.go:330] unrecognized feature gate: OVNObservability Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.621168 4754 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.621173 4754 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.621177 4754 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.621181 4754 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.621185 4754 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.621191 4754 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.621195 4754 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.621199 4754 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.621203 4754 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.621207 4754 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.621212 4754 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.621216 4754 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.621220 4754 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.621224 4754 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.621230 4754 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.621236 4754 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.621242 4754 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.621246 4754 feature_gate.go:330] unrecognized feature gate: InsightsConfig Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.621252 4754 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.621363 4754 server.go:940] "Client rotation is on, will bootstrap in background" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.624300 4754 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.624378 4754 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.625034 4754 server.go:997] "Starting client certificate rotation" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.625057 4754 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.625497 4754 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2025-11-06 23:32:29.133289197 +0000 UTC Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.626096 4754 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.633408 4754 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Jan 26 16:47:13 crc kubenswrapper[4754]: E0126 16:47:13.633594 4754 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.102.83.236:6443: connect: connection refused" logger="UnhandledError" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.635275 4754 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.650725 4754 log.go:25] "Validated CRI v1 runtime API" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.675005 4754 log.go:25] "Validated CRI v1 image API" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.676788 4754 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.678895 4754 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2026-01-26-16-43-02-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.678927 4754 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.691374 4754 manager.go:217] Machine: {Timestamp:2026-01-26 16:47:13.690272048 +0000 UTC m=+0.214452502 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2799998 MemoryCapacity:33654128640 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:bfbb6c33-ce46-4e81-a8a8-b44409b03821 BootID:d8506764-e7fa-45cb-a13d-6f527164f548 Filesystems:[{Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827064320 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365412864 Type:vfs Inodes:821634 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:4108170 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827064320 Type:vfs Inodes:4108170 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:6a:f0:e2 Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:6a:f0:e2 Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:44:16:b2 Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:cd:e5:68 Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:1c:1e:48 Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:a8:8b:49 Speed:-1 Mtu:1496} {Name:eth10 MacAddress:26:b4:0e:f4:02:86 Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:96:e3:39:29:48:ef Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654128640 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.691588 4754 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.691784 4754 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.692149 4754 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.692329 4754 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.692379 4754 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.692701 4754 topology_manager.go:138] "Creating topology manager with none policy" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.692717 4754 container_manager_linux.go:303] "Creating device plugin manager" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.692944 4754 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.692991 4754 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.693349 4754 state_mem.go:36] "Initialized new in-memory state store" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.693440 4754 server.go:1245] "Using root directory" path="/var/lib/kubelet" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.694177 4754 kubelet.go:418] "Attempting to sync node with API server" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.694201 4754 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.694237 4754 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.694249 4754 kubelet.go:324] "Adding apiserver pod source" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.694259 4754 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.696066 4754 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.236:6443: connect: connection refused Jan 26 16:47:13 crc kubenswrapper[4754]: E0126 16:47:13.698746 4754 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.236:6443: connect: connection refused" logger="UnhandledError" Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.698871 4754 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.236:6443: connect: connection refused Jan 26 16:47:13 crc kubenswrapper[4754]: E0126 16:47:13.698999 4754 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.236:6443: connect: connection refused" logger="UnhandledError" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.699113 4754 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.699458 4754 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.700281 4754 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.700773 4754 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.700794 4754 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.700801 4754 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.700808 4754 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.700819 4754 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.700826 4754 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.700833 4754 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.700845 4754 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.700853 4754 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.700860 4754 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.700870 4754 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.700877 4754 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.700892 4754 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.701765 4754 server.go:1280] "Started kubelet" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.703024 4754 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.703370 4754 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Jan 26 16:47:13 crc systemd[1]: Started Kubernetes Kubelet. Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.703998 4754 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.705331 4754 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.236:6443: connect: connection refused Jan 26 16:47:13 crc kubenswrapper[4754]: E0126 16:47:13.705323 4754 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.236:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.188e55caf003a91b default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-01-26 16:47:13.701710107 +0000 UTC m=+0.225890571,LastTimestamp:2026-01-26 16:47:13.701710107 +0000 UTC m=+0.225890571,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.706590 4754 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.706652 4754 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.706852 4754 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-20 09:19:37.954981818 +0000 UTC Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.707209 4754 server.go:460] "Adding debug handlers to kubelet server" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.707274 4754 volume_manager.go:287] "The desired_state_of_world populator starts" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.707287 4754 volume_manager.go:289] "Starting Kubelet Volume Manager" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.707447 4754 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Jan 26 16:47:13 crc kubenswrapper[4754]: E0126 16:47:13.707755 4754 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.708304 4754 factory.go:55] Registering systemd factory Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.708348 4754 factory.go:221] Registration of the systemd container factory successfully Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.708865 4754 factory.go:153] Registering CRI-O factory Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.708910 4754 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.236:6443: connect: connection refused Jan 26 16:47:13 crc kubenswrapper[4754]: E0126 16:47:13.709003 4754 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.236:6443: connect: connection refused" logger="UnhandledError" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.708960 4754 factory.go:221] Registration of the crio container factory successfully Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.709200 4754 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.709279 4754 factory.go:103] Registering Raw factory Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.709343 4754 manager.go:1196] Started watching for new ooms in manager Jan 26 16:47:13 crc kubenswrapper[4754]: E0126 16:47:13.709423 4754 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.236:6443: connect: connection refused" interval="200ms" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.709975 4754 manager.go:319] Starting recovery of all containers Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.714621 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.714718 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.714737 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.714751 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.714766 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.714780 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.714791 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.714803 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.714819 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.714830 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.714843 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.714856 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.714868 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.714883 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.714894 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.714906 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.714995 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.715011 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.715025 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.715041 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.715053 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.715067 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.715080 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.715092 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.715105 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.715120 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.715137 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.715160 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.715173 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.715187 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.715198 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.715212 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.715226 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.715239 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.715251 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.715266 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.715283 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.715297 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.715310 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.715323 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.715338 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.715351 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.715364 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.715376 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.715390 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.715403 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.715416 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.715428 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.715442 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.715503 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.715518 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.715532 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.715573 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.715588 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.715604 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.715618 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.715633 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.715646 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.715658 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.715688 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.715704 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.715719 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.715733 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.715746 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.715759 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.715772 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.715783 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.715796 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.715809 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.715822 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.715835 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.715847 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.715860 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.716504 4754 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.716530 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.716547 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.716561 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.716575 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.716589 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.716601 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.716614 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.716630 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.716644 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.716704 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.716718 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.716733 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.716747 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.716762 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.716787 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.716802 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.716817 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.716835 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.716848 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.716863 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.716876 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.716892 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.716904 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.716920 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.716936 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.716949 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.716962 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.716975 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.716989 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.717004 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.717021 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.717042 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.717057 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.717072 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.717084 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.717096 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.717107 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.717120 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.717132 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.717143 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.717157 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.717172 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.717185 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.717198 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.717213 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.717225 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.717240 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.717253 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.717268 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.717282 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.717294 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.717307 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.717319 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.717337 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.717349 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.717362 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.717375 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.717386 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.717397 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.717410 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.717422 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.717434 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.717449 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.717461 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.717475 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.717487 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.717499 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.717510 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.717522 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.717535 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.717548 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.717561 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.717574 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.717586 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.717599 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.717613 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.717626 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.717638 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.717651 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.717681 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.717695 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.717709 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.717721 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.717732 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.717744 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.717755 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.717768 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.717780 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.717791 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.717802 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.717813 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.717824 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.717835 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.717846 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.717858 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.717870 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.717882 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.717894 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.717907 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.717918 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.717929 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.717941 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.717953 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.717964 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.717976 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.717987 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.717998 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.718010 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.718022 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.718045 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.718056 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.718070 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.718082 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.718094 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.718106 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.718119 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.718131 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.718142 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.718153 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.718165 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.718176 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.718189 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.718202 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.718215 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.718228 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.718240 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.718254 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.718267 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.718281 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.718293 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.718306 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.718320 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.718333 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.718345 4754 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.718356 4754 reconstruct.go:97] "Volume reconstruction finished" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.718367 4754 reconciler.go:26] "Reconciler: start to sync state" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.735142 4754 manager.go:324] Recovery completed Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.744338 4754 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.745874 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.745909 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.745919 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.747587 4754 cpu_manager.go:225] "Starting CPU manager" policy="none" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.747613 4754 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.747646 4754 state_mem.go:36] "Initialized new in-memory state store" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.758597 4754 policy_none.go:49] "None policy: Start" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.761176 4754 memory_manager.go:170] "Starting memorymanager" policy="None" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.761233 4754 state_mem.go:35] "Initializing new in-memory state store" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.762401 4754 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.765998 4754 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.766067 4754 status_manager.go:217] "Starting to sync pod status with apiserver" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.766102 4754 kubelet.go:2335] "Starting kubelet main sync loop" Jan 26 16:47:13 crc kubenswrapper[4754]: E0126 16:47:13.766182 4754 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Jan 26 16:47:13 crc kubenswrapper[4754]: W0126 16:47:13.766828 4754 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.236:6443: connect: connection refused Jan 26 16:47:13 crc kubenswrapper[4754]: E0126 16:47:13.766899 4754 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.236:6443: connect: connection refused" logger="UnhandledError" Jan 26 16:47:13 crc kubenswrapper[4754]: E0126 16:47:13.807934 4754 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.818887 4754 manager.go:334] "Starting Device Plugin manager" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.818971 4754 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.818989 4754 server.go:79] "Starting device plugin registration server" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.819413 4754 eviction_manager.go:189] "Eviction manager: starting control loop" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.819443 4754 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.819755 4754 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.819910 4754 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.819928 4754 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Jan 26 16:47:13 crc kubenswrapper[4754]: E0126 16:47:13.825983 4754 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.866747 4754 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc"] Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.866929 4754 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.867821 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.867846 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.867855 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.867969 4754 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.868306 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.868337 4754 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.868910 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.868926 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.868934 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.869033 4754 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.869137 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.869180 4754 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.869392 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.869421 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.869433 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.869800 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.869842 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.869854 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.870033 4754 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.870077 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.870112 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.870143 4754 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.870140 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.870231 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.871528 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.871565 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.871576 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.871818 4754 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.872108 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.872320 4754 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.872507 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.872562 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.872577 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.873126 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.873153 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.873182 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.873192 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.873217 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.873227 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.873471 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.873511 4754 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.874542 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.874585 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.874597 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:13 crc kubenswrapper[4754]: E0126 16:47:13.910841 4754 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.236:6443: connect: connection refused" interval="400ms" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.919598 4754 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.920042 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.920116 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.920159 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.920197 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.920221 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.920317 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.920487 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.920518 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.920541 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.920560 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.920580 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.920598 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.920619 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.920681 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.920753 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.921091 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.921127 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.921136 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:13 crc kubenswrapper[4754]: I0126 16:47:13.921160 4754 kubelet_node_status.go:76] "Attempting to register node" node="crc" Jan 26 16:47:13 crc kubenswrapper[4754]: E0126 16:47:13.921719 4754 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.236:6443: connect: connection refused" node="crc" Jan 26 16:47:14 crc kubenswrapper[4754]: I0126 16:47:14.021551 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 26 16:47:14 crc kubenswrapper[4754]: I0126 16:47:14.021655 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 26 16:47:14 crc kubenswrapper[4754]: I0126 16:47:14.021738 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 26 16:47:14 crc kubenswrapper[4754]: I0126 16:47:14.021781 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 26 16:47:14 crc kubenswrapper[4754]: I0126 16:47:14.021825 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Jan 26 16:47:14 crc kubenswrapper[4754]: I0126 16:47:14.021840 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 26 16:47:14 crc kubenswrapper[4754]: I0126 16:47:14.021869 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 26 16:47:14 crc kubenswrapper[4754]: I0126 16:47:14.021841 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 26 16:47:14 crc kubenswrapper[4754]: I0126 16:47:14.021922 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 26 16:47:14 crc kubenswrapper[4754]: I0126 16:47:14.021998 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 26 16:47:14 crc kubenswrapper[4754]: I0126 16:47:14.022014 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 26 16:47:14 crc kubenswrapper[4754]: I0126 16:47:14.022042 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Jan 26 16:47:14 crc kubenswrapper[4754]: I0126 16:47:14.022071 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 26 16:47:14 crc kubenswrapper[4754]: I0126 16:47:14.022085 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 26 16:47:14 crc kubenswrapper[4754]: I0126 16:47:14.022117 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Jan 26 16:47:14 crc kubenswrapper[4754]: I0126 16:47:14.022132 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 26 16:47:14 crc kubenswrapper[4754]: I0126 16:47:14.022162 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 26 16:47:14 crc kubenswrapper[4754]: I0126 16:47:14.022174 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Jan 26 16:47:14 crc kubenswrapper[4754]: I0126 16:47:14.022188 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 26 16:47:14 crc kubenswrapper[4754]: I0126 16:47:14.022216 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Jan 26 16:47:14 crc kubenswrapper[4754]: I0126 16:47:14.022245 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 26 16:47:14 crc kubenswrapper[4754]: I0126 16:47:14.022041 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 26 16:47:14 crc kubenswrapper[4754]: I0126 16:47:14.022259 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 26 16:47:14 crc kubenswrapper[4754]: I0126 16:47:14.022286 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Jan 26 16:47:14 crc kubenswrapper[4754]: I0126 16:47:14.022303 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 26 16:47:14 crc kubenswrapper[4754]: I0126 16:47:14.022202 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 26 16:47:14 crc kubenswrapper[4754]: I0126 16:47:14.022327 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 26 16:47:14 crc kubenswrapper[4754]: I0126 16:47:14.022416 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Jan 26 16:47:14 crc kubenswrapper[4754]: I0126 16:47:14.022466 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 26 16:47:14 crc kubenswrapper[4754]: I0126 16:47:14.022419 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Jan 26 16:47:14 crc kubenswrapper[4754]: I0126 16:47:14.122392 4754 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 26 16:47:14 crc kubenswrapper[4754]: I0126 16:47:14.124135 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:14 crc kubenswrapper[4754]: I0126 16:47:14.124209 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:14 crc kubenswrapper[4754]: I0126 16:47:14.124233 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:14 crc kubenswrapper[4754]: I0126 16:47:14.124275 4754 kubelet_node_status.go:76] "Attempting to register node" node="crc" Jan 26 16:47:14 crc kubenswrapper[4754]: E0126 16:47:14.125034 4754 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.236:6443: connect: connection refused" node="crc" Jan 26 16:47:14 crc kubenswrapper[4754]: I0126 16:47:14.193980 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Jan 26 16:47:14 crc kubenswrapper[4754]: I0126 16:47:14.217703 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Jan 26 16:47:14 crc kubenswrapper[4754]: W0126 16:47:14.223755 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-5f83b2946958b1f51ca79ebbb68504ffd12298c614070bea73303b8743ea8ba6 WatchSource:0}: Error finding container 5f83b2946958b1f51ca79ebbb68504ffd12298c614070bea73303b8743ea8ba6: Status 404 returned error can't find the container with id 5f83b2946958b1f51ca79ebbb68504ffd12298c614070bea73303b8743ea8ba6 Jan 26 16:47:14 crc kubenswrapper[4754]: I0126 16:47:14.234795 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 26 16:47:14 crc kubenswrapper[4754]: W0126 16:47:14.243543 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-455e07cf009c247c6c690b701edd9005e99820ec348cf8abd98fdd735178032e WatchSource:0}: Error finding container 455e07cf009c247c6c690b701edd9005e99820ec348cf8abd98fdd735178032e: Status 404 returned error can't find the container with id 455e07cf009c247c6c690b701edd9005e99820ec348cf8abd98fdd735178032e Jan 26 16:47:14 crc kubenswrapper[4754]: I0126 16:47:14.246901 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 26 16:47:14 crc kubenswrapper[4754]: I0126 16:47:14.252271 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Jan 26 16:47:14 crc kubenswrapper[4754]: W0126 16:47:14.252501 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-027dc212fe30a734bd5807ade87bb99437982732e20a409d945d845f8c208a4d WatchSource:0}: Error finding container 027dc212fe30a734bd5807ade87bb99437982732e20a409d945d845f8c208a4d: Status 404 returned error can't find the container with id 027dc212fe30a734bd5807ade87bb99437982732e20a409d945d845f8c208a4d Jan 26 16:47:14 crc kubenswrapper[4754]: W0126 16:47:14.268177 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-5a9232a8050a5a5222b286b99c19093cc31832d7c0fbd8ff8992ddcb4ee7ec71 WatchSource:0}: Error finding container 5a9232a8050a5a5222b286b99c19093cc31832d7c0fbd8ff8992ddcb4ee7ec71: Status 404 returned error can't find the container with id 5a9232a8050a5a5222b286b99c19093cc31832d7c0fbd8ff8992ddcb4ee7ec71 Jan 26 16:47:14 crc kubenswrapper[4754]: E0126 16:47:14.311907 4754 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.236:6443: connect: connection refused" interval="800ms" Jan 26 16:47:14 crc kubenswrapper[4754]: I0126 16:47:14.525955 4754 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 26 16:47:14 crc kubenswrapper[4754]: I0126 16:47:14.527569 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:14 crc kubenswrapper[4754]: I0126 16:47:14.527612 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:14 crc kubenswrapper[4754]: I0126 16:47:14.527625 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:14 crc kubenswrapper[4754]: I0126 16:47:14.527654 4754 kubelet_node_status.go:76] "Attempting to register node" node="crc" Jan 26 16:47:14 crc kubenswrapper[4754]: E0126 16:47:14.528206 4754 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.236:6443: connect: connection refused" node="crc" Jan 26 16:47:14 crc kubenswrapper[4754]: I0126 16:47:14.706953 4754 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.236:6443: connect: connection refused Jan 26 16:47:14 crc kubenswrapper[4754]: I0126 16:47:14.707044 4754 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-17 17:13:37.309533628 +0000 UTC Jan 26 16:47:14 crc kubenswrapper[4754]: I0126 16:47:14.772462 4754 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="23bb333dc72b17ef056baa2ab67c8964816ac60cca2b875e6fa5f853cc625715" exitCode=0 Jan 26 16:47:14 crc kubenswrapper[4754]: I0126 16:47:14.772542 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"23bb333dc72b17ef056baa2ab67c8964816ac60cca2b875e6fa5f853cc625715"} Jan 26 16:47:14 crc kubenswrapper[4754]: I0126 16:47:14.772631 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"455e07cf009c247c6c690b701edd9005e99820ec348cf8abd98fdd735178032e"} Jan 26 16:47:14 crc kubenswrapper[4754]: I0126 16:47:14.772805 4754 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 26 16:47:14 crc kubenswrapper[4754]: I0126 16:47:14.774023 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:14 crc kubenswrapper[4754]: I0126 16:47:14.774053 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:14 crc kubenswrapper[4754]: I0126 16:47:14.774064 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:14 crc kubenswrapper[4754]: I0126 16:47:14.774319 4754 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="91ba0a944e0cac147aac51868c09fedf144859e7db28a129acd17895172391b4" exitCode=0 Jan 26 16:47:14 crc kubenswrapper[4754]: I0126 16:47:14.774390 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"91ba0a944e0cac147aac51868c09fedf144859e7db28a129acd17895172391b4"} Jan 26 16:47:14 crc kubenswrapper[4754]: I0126 16:47:14.774412 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"5f83b2946958b1f51ca79ebbb68504ffd12298c614070bea73303b8743ea8ba6"} Jan 26 16:47:14 crc kubenswrapper[4754]: I0126 16:47:14.774486 4754 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 26 16:47:14 crc kubenswrapper[4754]: I0126 16:47:14.775490 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:14 crc kubenswrapper[4754]: I0126 16:47:14.775536 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:14 crc kubenswrapper[4754]: I0126 16:47:14.775552 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:14 crc kubenswrapper[4754]: I0126 16:47:14.776378 4754 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="eb7a3b85cca4436068c8ceefc3f2358cfa1a2c398a8acecd367b944528b11489" exitCode=0 Jan 26 16:47:14 crc kubenswrapper[4754]: I0126 16:47:14.776437 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"eb7a3b85cca4436068c8ceefc3f2358cfa1a2c398a8acecd367b944528b11489"} Jan 26 16:47:14 crc kubenswrapper[4754]: I0126 16:47:14.776454 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"5bba6ecb2d8a4a4c8ede381d93b3937c688132e6f40695532d1f9540853e67ee"} Jan 26 16:47:14 crc kubenswrapper[4754]: I0126 16:47:14.776517 4754 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 26 16:47:14 crc kubenswrapper[4754]: I0126 16:47:14.777354 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:14 crc kubenswrapper[4754]: I0126 16:47:14.777398 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:14 crc kubenswrapper[4754]: I0126 16:47:14.777412 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:14 crc kubenswrapper[4754]: I0126 16:47:14.778849 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"a74507c941b9de4d0401c8f1d53bb79ed11359628d3405cdab7f5474a75c862e"} Jan 26 16:47:14 crc kubenswrapper[4754]: I0126 16:47:14.778915 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"5a9232a8050a5a5222b286b99c19093cc31832d7c0fbd8ff8992ddcb4ee7ec71"} Jan 26 16:47:14 crc kubenswrapper[4754]: I0126 16:47:14.781218 4754 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="22ba7649a1308815bfad52016000987a7a9a5bc214d7ff61e137ba099b20fc6d" exitCode=0 Jan 26 16:47:14 crc kubenswrapper[4754]: I0126 16:47:14.781257 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"22ba7649a1308815bfad52016000987a7a9a5bc214d7ff61e137ba099b20fc6d"} Jan 26 16:47:14 crc kubenswrapper[4754]: I0126 16:47:14.781309 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"027dc212fe30a734bd5807ade87bb99437982732e20a409d945d845f8c208a4d"} Jan 26 16:47:14 crc kubenswrapper[4754]: I0126 16:47:14.781425 4754 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 26 16:47:14 crc kubenswrapper[4754]: I0126 16:47:14.782361 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:14 crc kubenswrapper[4754]: I0126 16:47:14.782395 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:14 crc kubenswrapper[4754]: I0126 16:47:14.782407 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:14 crc kubenswrapper[4754]: I0126 16:47:14.785262 4754 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 26 16:47:14 crc kubenswrapper[4754]: I0126 16:47:14.786636 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:14 crc kubenswrapper[4754]: I0126 16:47:14.786702 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:14 crc kubenswrapper[4754]: I0126 16:47:14.786718 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:14 crc kubenswrapper[4754]: W0126 16:47:14.912972 4754 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.236:6443: connect: connection refused Jan 26 16:47:14 crc kubenswrapper[4754]: E0126 16:47:14.913093 4754 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.236:6443: connect: connection refused" logger="UnhandledError" Jan 26 16:47:14 crc kubenswrapper[4754]: W0126 16:47:14.951636 4754 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.236:6443: connect: connection refused Jan 26 16:47:14 crc kubenswrapper[4754]: E0126 16:47:14.951734 4754 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.236:6443: connect: connection refused" logger="UnhandledError" Jan 26 16:47:15 crc kubenswrapper[4754]: W0126 16:47:15.071780 4754 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.236:6443: connect: connection refused Jan 26 16:47:15 crc kubenswrapper[4754]: E0126 16:47:15.071889 4754 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.236:6443: connect: connection refused" logger="UnhandledError" Jan 26 16:47:15 crc kubenswrapper[4754]: E0126 16:47:15.112805 4754 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.236:6443: connect: connection refused" interval="1.6s" Jan 26 16:47:15 crc kubenswrapper[4754]: W0126 16:47:15.136857 4754 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.236:6443: connect: connection refused Jan 26 16:47:15 crc kubenswrapper[4754]: E0126 16:47:15.136938 4754 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.236:6443: connect: connection refused" logger="UnhandledError" Jan 26 16:47:15 crc kubenswrapper[4754]: I0126 16:47:15.329141 4754 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 26 16:47:15 crc kubenswrapper[4754]: I0126 16:47:15.330354 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:15 crc kubenswrapper[4754]: I0126 16:47:15.330382 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:15 crc kubenswrapper[4754]: I0126 16:47:15.330390 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:15 crc kubenswrapper[4754]: I0126 16:47:15.330410 4754 kubelet_node_status.go:76] "Attempting to register node" node="crc" Jan 26 16:47:15 crc kubenswrapper[4754]: E0126 16:47:15.330854 4754 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.236:6443: connect: connection refused" node="crc" Jan 26 16:47:15 crc kubenswrapper[4754]: I0126 16:47:15.707911 4754 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-25 18:15:36.934094961 +0000 UTC Jan 26 16:47:15 crc kubenswrapper[4754]: I0126 16:47:15.785619 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"f772a003566198519fe1849fd14f29d4346004b2e655810662507c95b4d5ffa1"} Jan 26 16:47:15 crc kubenswrapper[4754]: I0126 16:47:15.785696 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"4402182b3e2e35a7d5dd4c44ac53e152915b312b7ccb076c4f2b41b925a6e7f7"} Jan 26 16:47:15 crc kubenswrapper[4754]: I0126 16:47:15.785704 4754 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 26 16:47:15 crc kubenswrapper[4754]: I0126 16:47:15.785712 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"23017e73f8450839b4c2c02fa61235ffe4604bdce11556def27f25bcadcf89dc"} Jan 26 16:47:15 crc kubenswrapper[4754]: I0126 16:47:15.786957 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:15 crc kubenswrapper[4754]: I0126 16:47:15.787007 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:15 crc kubenswrapper[4754]: I0126 16:47:15.787016 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:15 crc kubenswrapper[4754]: I0126 16:47:15.789042 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"8356f76fbc3a4e636c12d1da5455d3f6767d9b83ce2f2925a11435d7d7261e9f"} Jan 26 16:47:15 crc kubenswrapper[4754]: I0126 16:47:15.789161 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"27b74e2e65a547e0e28fddd4f15feacb17f4d57596eeb8da82df9d39243c85e0"} Jan 26 16:47:15 crc kubenswrapper[4754]: I0126 16:47:15.789182 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"4f45445cbf1d180d6d4dd9588e85bf678f043bfce8cff0812b78ca3510904f9a"} Jan 26 16:47:15 crc kubenswrapper[4754]: I0126 16:47:15.789196 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"085fa7b8d4fef8b2722fac8c346107699b5a6bd33fbfc74ed35831cb8fdf9dad"} Jan 26 16:47:15 crc kubenswrapper[4754]: I0126 16:47:15.789208 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"b9c5f40188fda14b14128514fe84bddfe8dcede3d6c560283ccb2ac0c66895fc"} Jan 26 16:47:15 crc kubenswrapper[4754]: I0126 16:47:15.789308 4754 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 26 16:47:15 crc kubenswrapper[4754]: I0126 16:47:15.790042 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:15 crc kubenswrapper[4754]: I0126 16:47:15.790074 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:15 crc kubenswrapper[4754]: I0126 16:47:15.790085 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:15 crc kubenswrapper[4754]: I0126 16:47:15.791101 4754 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="16f0eceee04af7372c3d260219fc9b300ed79a52c19cc704db61abe37c813248" exitCode=0 Jan 26 16:47:15 crc kubenswrapper[4754]: I0126 16:47:15.791179 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"16f0eceee04af7372c3d260219fc9b300ed79a52c19cc704db61abe37c813248"} Jan 26 16:47:15 crc kubenswrapper[4754]: I0126 16:47:15.791302 4754 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 26 16:47:15 crc kubenswrapper[4754]: I0126 16:47:15.792062 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:15 crc kubenswrapper[4754]: I0126 16:47:15.792101 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:15 crc kubenswrapper[4754]: I0126 16:47:15.792117 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:15 crc kubenswrapper[4754]: I0126 16:47:15.793460 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"750d9397eeb93fcc4e6bfb4e1417933e459f8e5541fb80b58306dda8b12a5a8a"} Jan 26 16:47:15 crc kubenswrapper[4754]: I0126 16:47:15.793531 4754 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 26 16:47:15 crc kubenswrapper[4754]: I0126 16:47:15.794210 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:15 crc kubenswrapper[4754]: I0126 16:47:15.794241 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:15 crc kubenswrapper[4754]: I0126 16:47:15.794256 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:15 crc kubenswrapper[4754]: I0126 16:47:15.797117 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"8f9bf9e4a705cd849595e59e1123ba70773bfa5f8a0d0b79195e4fc7e447ca26"} Jan 26 16:47:15 crc kubenswrapper[4754]: I0126 16:47:15.797172 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"9e5fbff42fdebad6a74c21d85931d3c8d983973db82fc7f262ba5383bad76172"} Jan 26 16:47:15 crc kubenswrapper[4754]: I0126 16:47:15.797183 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"7da43edaa2a81f1b647b96c234d795b1ddf5fa23173b78043123f76dc677c517"} Jan 26 16:47:15 crc kubenswrapper[4754]: I0126 16:47:15.797288 4754 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 26 16:47:15 crc kubenswrapper[4754]: I0126 16:47:15.797988 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:15 crc kubenswrapper[4754]: I0126 16:47:15.798026 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:15 crc kubenswrapper[4754]: I0126 16:47:15.798038 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:15 crc kubenswrapper[4754]: I0126 16:47:15.834256 4754 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Jan 26 16:47:15 crc kubenswrapper[4754]: I0126 16:47:15.849890 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 26 16:47:16 crc kubenswrapper[4754]: I0126 16:47:16.157378 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 26 16:47:16 crc kubenswrapper[4754]: I0126 16:47:16.708106 4754 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-06 18:07:16.921680505 +0000 UTC Jan 26 16:47:16 crc kubenswrapper[4754]: I0126 16:47:16.803518 4754 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="fd852d56da224ada06720a2639b665b6ccc7a44cc804edef8bff761604fa5b2d" exitCode=0 Jan 26 16:47:16 crc kubenswrapper[4754]: I0126 16:47:16.803590 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"fd852d56da224ada06720a2639b665b6ccc7a44cc804edef8bff761604fa5b2d"} Jan 26 16:47:16 crc kubenswrapper[4754]: I0126 16:47:16.803865 4754 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 26 16:47:16 crc kubenswrapper[4754]: I0126 16:47:16.803918 4754 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 26 16:47:16 crc kubenswrapper[4754]: I0126 16:47:16.803974 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 26 16:47:16 crc kubenswrapper[4754]: I0126 16:47:16.803930 4754 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 26 16:47:16 crc kubenswrapper[4754]: I0126 16:47:16.805658 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:16 crc kubenswrapper[4754]: I0126 16:47:16.805735 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:16 crc kubenswrapper[4754]: I0126 16:47:16.805761 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:16 crc kubenswrapper[4754]: I0126 16:47:16.805774 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:16 crc kubenswrapper[4754]: I0126 16:47:16.805791 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:16 crc kubenswrapper[4754]: I0126 16:47:16.805791 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:16 crc kubenswrapper[4754]: I0126 16:47:16.805922 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:16 crc kubenswrapper[4754]: I0126 16:47:16.805968 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:16 crc kubenswrapper[4754]: I0126 16:47:16.805995 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:16 crc kubenswrapper[4754]: I0126 16:47:16.931478 4754 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 26 16:47:16 crc kubenswrapper[4754]: I0126 16:47:16.933029 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:16 crc kubenswrapper[4754]: I0126 16:47:16.933076 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:16 crc kubenswrapper[4754]: I0126 16:47:16.933087 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:16 crc kubenswrapper[4754]: I0126 16:47:16.933118 4754 kubelet_node_status.go:76] "Attempting to register node" node="crc" Jan 26 16:47:17 crc kubenswrapper[4754]: I0126 16:47:17.709652 4754 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-08 20:19:42.282723773 +0000 UTC Jan 26 16:47:17 crc kubenswrapper[4754]: I0126 16:47:17.809781 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"8028452fa6e2b6788c037928f263bd6613d43f61186bea9c4def1b94696c2a8b"} Jan 26 16:47:17 crc kubenswrapper[4754]: I0126 16:47:17.809843 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"635f5505215f6f7ddb85d114a6b4e652e3bea2d3205aed056ec2c1ad7666eac3"} Jan 26 16:47:17 crc kubenswrapper[4754]: I0126 16:47:17.809859 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"87159b411947480ec737a3c0f239343f68d1568ec29ead13bcdcb08813a1e5b6"} Jan 26 16:47:17 crc kubenswrapper[4754]: I0126 16:47:17.809868 4754 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 26 16:47:17 crc kubenswrapper[4754]: I0126 16:47:17.809868 4754 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 26 16:47:17 crc kubenswrapper[4754]: I0126 16:47:17.809870 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"f27ed8c3bd03027a12862af0081bab9f1f3f377cfd7b2705b4b7a6418e200ad9"} Jan 26 16:47:17 crc kubenswrapper[4754]: I0126 16:47:17.810727 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:17 crc kubenswrapper[4754]: I0126 16:47:17.810773 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:17 crc kubenswrapper[4754]: I0126 16:47:17.810789 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:17 crc kubenswrapper[4754]: I0126 16:47:17.810857 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:17 crc kubenswrapper[4754]: I0126 16:47:17.810890 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:17 crc kubenswrapper[4754]: I0126 16:47:17.810900 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:18 crc kubenswrapper[4754]: I0126 16:47:18.211068 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Jan 26 16:47:18 crc kubenswrapper[4754]: I0126 16:47:18.211271 4754 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 26 16:47:18 crc kubenswrapper[4754]: I0126 16:47:18.212719 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:18 crc kubenswrapper[4754]: I0126 16:47:18.212761 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:18 crc kubenswrapper[4754]: I0126 16:47:18.212775 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:18 crc kubenswrapper[4754]: I0126 16:47:18.710824 4754 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-08 08:36:51.722818783 +0000 UTC Jan 26 16:47:18 crc kubenswrapper[4754]: I0126 16:47:18.818861 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"99d19ccec26867e8bc15abfd93970c15b9fdd9e7ff04f7d59934af432885efc7"} Jan 26 16:47:18 crc kubenswrapper[4754]: I0126 16:47:18.819005 4754 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 26 16:47:18 crc kubenswrapper[4754]: I0126 16:47:18.820077 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:18 crc kubenswrapper[4754]: I0126 16:47:18.820105 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:18 crc kubenswrapper[4754]: I0126 16:47:18.820113 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:18 crc kubenswrapper[4754]: I0126 16:47:18.826484 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 26 16:47:18 crc kubenswrapper[4754]: I0126 16:47:18.826740 4754 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 26 16:47:18 crc kubenswrapper[4754]: I0126 16:47:18.827774 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:18 crc kubenswrapper[4754]: I0126 16:47:18.827821 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:18 crc kubenswrapper[4754]: I0126 16:47:18.827835 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:19 crc kubenswrapper[4754]: I0126 16:47:19.712013 4754 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-12 02:49:26.879463976 +0000 UTC Jan 26 16:47:19 crc kubenswrapper[4754]: I0126 16:47:19.822583 4754 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 26 16:47:19 crc kubenswrapper[4754]: I0126 16:47:19.823904 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:19 crc kubenswrapper[4754]: I0126 16:47:19.823974 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:19 crc kubenswrapper[4754]: I0126 16:47:19.823995 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:20 crc kubenswrapper[4754]: I0126 16:47:20.500911 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Jan 26 16:47:20 crc kubenswrapper[4754]: I0126 16:47:20.712439 4754 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-30 03:03:55.125245315 +0000 UTC Jan 26 16:47:20 crc kubenswrapper[4754]: I0126 16:47:20.825963 4754 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 26 16:47:20 crc kubenswrapper[4754]: I0126 16:47:20.826993 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:20 crc kubenswrapper[4754]: I0126 16:47:20.827039 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:20 crc kubenswrapper[4754]: I0126 16:47:20.827050 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:21 crc kubenswrapper[4754]: I0126 16:47:21.713089 4754 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-09 12:10:59.772351742 +0000 UTC Jan 26 16:47:22 crc kubenswrapper[4754]: I0126 16:47:22.714191 4754 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-17 23:15:00.017952854 +0000 UTC Jan 26 16:47:23 crc kubenswrapper[4754]: I0126 16:47:23.011780 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Jan 26 16:47:23 crc kubenswrapper[4754]: I0126 16:47:23.011975 4754 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 26 16:47:23 crc kubenswrapper[4754]: I0126 16:47:23.013195 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:23 crc kubenswrapper[4754]: I0126 16:47:23.013227 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:23 crc kubenswrapper[4754]: I0126 16:47:23.013237 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:23 crc kubenswrapper[4754]: I0126 16:47:23.714704 4754 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-09 15:03:46.052921276 +0000 UTC Jan 26 16:47:23 crc kubenswrapper[4754]: E0126 16:47:23.826150 4754 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Jan 26 16:47:24 crc kubenswrapper[4754]: I0126 16:47:24.714984 4754 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-17 05:39:41.392683923 +0000 UTC Jan 26 16:47:24 crc kubenswrapper[4754]: I0126 16:47:24.729387 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 26 16:47:24 crc kubenswrapper[4754]: I0126 16:47:24.729586 4754 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 26 16:47:24 crc kubenswrapper[4754]: I0126 16:47:24.730542 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:24 crc kubenswrapper[4754]: I0126 16:47:24.730588 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:24 crc kubenswrapper[4754]: I0126 16:47:24.730599 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:25 crc kubenswrapper[4754]: I0126 16:47:25.457001 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 26 16:47:25 crc kubenswrapper[4754]: I0126 16:47:25.457192 4754 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 26 16:47:25 crc kubenswrapper[4754]: I0126 16:47:25.458450 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:25 crc kubenswrapper[4754]: I0126 16:47:25.458508 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:25 crc kubenswrapper[4754]: I0126 16:47:25.458526 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:25 crc kubenswrapper[4754]: I0126 16:47:25.707650 4754 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Jan 26 16:47:25 crc kubenswrapper[4754]: I0126 16:47:25.716028 4754 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-14 00:39:55.462909386 +0000 UTC Jan 26 16:47:25 crc kubenswrapper[4754]: I0126 16:47:25.754369 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 26 16:47:25 crc kubenswrapper[4754]: I0126 16:47:25.759632 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 26 16:47:25 crc kubenswrapper[4754]: E0126 16:47:25.835946 4754 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": net/http: TLS handshake timeout" logger="UnhandledError" Jan 26 16:47:25 crc kubenswrapper[4754]: I0126 16:47:25.837539 4754 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 26 16:47:25 crc kubenswrapper[4754]: I0126 16:47:25.838571 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:25 crc kubenswrapper[4754]: I0126 16:47:25.838639 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:25 crc kubenswrapper[4754]: I0126 16:47:25.838717 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:25 crc kubenswrapper[4754]: I0126 16:47:25.841632 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 26 16:47:26 crc kubenswrapper[4754]: I0126 16:47:26.492448 4754 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Jan 26 16:47:26 crc kubenswrapper[4754]: I0126 16:47:26.492520 4754 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Jan 26 16:47:26 crc kubenswrapper[4754]: I0126 16:47:26.496267 4754 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Jan 26 16:47:26 crc kubenswrapper[4754]: I0126 16:47:26.496334 4754 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Jan 26 16:47:26 crc kubenswrapper[4754]: I0126 16:47:26.716960 4754 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-02 19:39:49.921917656 +0000 UTC Jan 26 16:47:26 crc kubenswrapper[4754]: I0126 16:47:26.841184 4754 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 26 16:47:26 crc kubenswrapper[4754]: I0126 16:47:26.842074 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:26 crc kubenswrapper[4754]: I0126 16:47:26.842119 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:26 crc kubenswrapper[4754]: I0126 16:47:26.842136 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:27 crc kubenswrapper[4754]: I0126 16:47:27.717553 4754 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-17 12:29:28.112029878 +0000 UTC Jan 26 16:47:27 crc kubenswrapper[4754]: I0126 16:47:27.729896 4754 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Jan 26 16:47:27 crc kubenswrapper[4754]: I0126 16:47:27.729995 4754 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Jan 26 16:47:27 crc kubenswrapper[4754]: I0126 16:47:27.845132 4754 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 26 16:47:27 crc kubenswrapper[4754]: I0126 16:47:27.846023 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:27 crc kubenswrapper[4754]: I0126 16:47:27.846084 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:27 crc kubenswrapper[4754]: I0126 16:47:27.846099 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:28 crc kubenswrapper[4754]: I0126 16:47:28.717740 4754 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-18 07:55:54.772985684 +0000 UTC Jan 26 16:47:28 crc kubenswrapper[4754]: I0126 16:47:28.832175 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 26 16:47:28 crc kubenswrapper[4754]: I0126 16:47:28.832727 4754 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 26 16:47:28 crc kubenswrapper[4754]: I0126 16:47:28.833916 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:28 crc kubenswrapper[4754]: I0126 16:47:28.833979 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:28 crc kubenswrapper[4754]: I0126 16:47:28.833991 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:28 crc kubenswrapper[4754]: I0126 16:47:28.836507 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 26 16:47:28 crc kubenswrapper[4754]: I0126 16:47:28.847419 4754 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 26 16:47:28 crc kubenswrapper[4754]: I0126 16:47:28.848270 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:28 crc kubenswrapper[4754]: I0126 16:47:28.848322 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:28 crc kubenswrapper[4754]: I0126 16:47:28.848338 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:29 crc kubenswrapper[4754]: I0126 16:47:29.718813 4754 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-18 09:06:38.036912391 +0000 UTC Jan 26 16:47:29 crc kubenswrapper[4754]: I0126 16:47:29.868960 4754 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Jan 26 16:47:29 crc kubenswrapper[4754]: I0126 16:47:29.882409 4754 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Jan 26 16:47:30 crc kubenswrapper[4754]: I0126 16:47:30.530830 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Jan 26 16:47:30 crc kubenswrapper[4754]: I0126 16:47:30.531070 4754 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 26 16:47:30 crc kubenswrapper[4754]: I0126 16:47:30.532507 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:30 crc kubenswrapper[4754]: I0126 16:47:30.532548 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:30 crc kubenswrapper[4754]: I0126 16:47:30.532556 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:30 crc kubenswrapper[4754]: I0126 16:47:30.545549 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Jan 26 16:47:30 crc kubenswrapper[4754]: I0126 16:47:30.719857 4754 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-20 19:29:45.151063845 +0000 UTC Jan 26 16:47:30 crc kubenswrapper[4754]: I0126 16:47:30.852834 4754 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 26 16:47:30 crc kubenswrapper[4754]: I0126 16:47:30.853931 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:30 crc kubenswrapper[4754]: I0126 16:47:30.853962 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:30 crc kubenswrapper[4754]: I0126 16:47:30.853973 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:31 crc kubenswrapper[4754]: E0126 16:47:31.495311 4754 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="3.2s" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.497143 4754 trace.go:236] Trace[1689716203]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (26-Jan-2026 16:47:17.191) (total time: 14305ms): Jan 26 16:47:31 crc kubenswrapper[4754]: Trace[1689716203]: ---"Objects listed" error: 14305ms (16:47:31.497) Jan 26 16:47:31 crc kubenswrapper[4754]: Trace[1689716203]: [14.305483063s] [14.305483063s] END Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.497181 4754 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.497143 4754 trace.go:236] Trace[334245773]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (26-Jan-2026 16:47:17.767) (total time: 13729ms): Jan 26 16:47:31 crc kubenswrapper[4754]: Trace[334245773]: ---"Objects listed" error: 13729ms (16:47:31.497) Jan 26 16:47:31 crc kubenswrapper[4754]: Trace[334245773]: [13.729203635s] [13.729203635s] END Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.497246 4754 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.498532 4754 trace.go:236] Trace[748717530]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (26-Jan-2026 16:47:17.640) (total time: 13858ms): Jan 26 16:47:31 crc kubenswrapper[4754]: Trace[748717530]: ---"Objects listed" error: 13858ms (16:47:31.498) Jan 26 16:47:31 crc kubenswrapper[4754]: Trace[748717530]: [13.858136719s] [13.858136719s] END Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.498567 4754 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.498716 4754 trace.go:236] Trace[1731947122]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (26-Jan-2026 16:47:17.669) (total time: 13829ms): Jan 26 16:47:31 crc kubenswrapper[4754]: Trace[1731947122]: ---"Objects listed" error: 13828ms (16:47:31.498) Jan 26 16:47:31 crc kubenswrapper[4754]: Trace[1731947122]: [13.829151959s] [13.829151959s] END Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.498749 4754 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Jan 26 16:47:31 crc kubenswrapper[4754]: E0126 16:47:31.499304 4754 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.499819 4754 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.526700 4754 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Liveness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:52932->192.168.126.11:17697: read: connection reset by peer" start-of-body= Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.526724 4754 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:52948->192.168.126.11:17697: read: connection reset by peer" start-of-body= Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.526769 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:52932->192.168.126.11:17697: read: connection reset by peer" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.526902 4754 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:52948->192.168.126.11:17697: read: connection reset by peer" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.528091 4754 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.528128 4754 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.705191 4754 apiserver.go:52] "Watching apiserver" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.707745 4754 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.708060 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g"] Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.708509 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.708575 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.708522 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.708579 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.708719 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 16:47:31 crc kubenswrapper[4754]: E0126 16:47:31.708816 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.709103 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 16:47:31 crc kubenswrapper[4754]: E0126 16:47:31.709348 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 16:47:31 crc kubenswrapper[4754]: E0126 16:47:31.709406 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.712881 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.713164 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.713408 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.713494 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.713519 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.713464 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.713586 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.713493 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.713621 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.719996 4754 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-07 13:41:21.578973978 +0000 UTC Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.748775 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.763638 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.780301 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.797322 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.808756 4754 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.811826 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.836068 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.851078 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.856847 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.858388 4754 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="8356f76fbc3a4e636c12d1da5455d3f6767d9b83ce2f2925a11435d7d7261e9f" exitCode=255 Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.858424 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"8356f76fbc3a4e636c12d1da5455d3f6767d9b83ce2f2925a11435d7d7261e9f"} Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.866461 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.874443 4754 scope.go:117] "RemoveContainer" containerID="8356f76fbc3a4e636c12d1da5455d3f6767d9b83ce2f2925a11435d7d7261e9f" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.874627 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.877001 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.886869 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.898800 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.901920 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.901968 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.901994 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.902012 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.902031 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.902049 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.902089 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.902106 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.902126 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.902145 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.902161 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.902175 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.902194 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.902211 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.902230 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.902248 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.902315 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.902329 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.902346 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.902415 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.902445 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.902469 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.902493 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.902521 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.902516 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.902553 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.902577 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.902598 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.902625 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.902691 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.902713 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.902719 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.902746 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.902772 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.902796 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.902819 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.902844 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.902866 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.902889 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.902911 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.902980 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.903007 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.903030 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.903053 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.903073 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.903097 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.903120 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.903142 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.903166 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.903188 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.903204 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.903220 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.903260 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.903276 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.903292 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.903307 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.903323 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.903339 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.903354 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.903373 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.903390 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.903405 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.903511 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.903530 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.903552 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.903573 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.903588 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.903604 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.903620 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.903638 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.903657 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.903704 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.903750 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.903774 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.903801 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.903823 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.903844 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.903867 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.903889 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.903909 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.903930 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.903951 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.903972 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.903990 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.904006 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.904022 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.904038 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.904055 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.904071 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.904086 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.904101 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.904116 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.904132 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.904147 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.904163 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.904179 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.904195 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.904214 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.904231 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.904248 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.904265 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.904283 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.904300 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.904316 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.904332 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.904350 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.904366 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.904382 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.904398 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.904414 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.904430 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.904445 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.904460 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.904479 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.904494 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.904510 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.904527 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.904543 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.904558 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.904573 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.904588 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.904628 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.904643 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.904661 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.904700 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.904717 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.904734 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.904750 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.904770 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.904788 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.904803 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.904829 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.904846 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.904865 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.904883 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.904899 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.904915 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.904931 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.904946 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.904962 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.904979 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.904998 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.905018 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.905034 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.905049 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.908043 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.908149 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.908210 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.908265 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.908302 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.908356 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.908387 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.908443 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.908496 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.908537 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.908594 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.908623 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.908661 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.908722 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.908761 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.908796 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.908830 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.908871 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.908906 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.908934 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.908968 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.909002 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.909032 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.909054 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.909084 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.909111 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.909131 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.909166 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.909192 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.909214 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.909237 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.909261 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.909293 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.909314 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.909343 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.909376 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.909400 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.909430 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.909463 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.909492 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.909519 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.909546 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.909572 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.909596 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.909621 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.909650 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.909706 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.909748 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.909838 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.909877 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.909917 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.909943 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.909973 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.910133 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.910187 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.910232 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.910331 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.910364 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.910394 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.910427 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.910454 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.910484 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.910516 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.910546 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.910576 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.910598 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.910619 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.910755 4754 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.910781 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.910794 4754 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.902715 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.917093 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.903082 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.903959 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.904324 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.904492 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.904643 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.904783 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.904844 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.904906 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.905028 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.907061 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.908810 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.909276 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.909322 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.909352 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.909629 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.909781 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.909998 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.909991 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.910942 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.911016 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.911409 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.912269 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.916221 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.916569 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.916602 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.916622 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.916927 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.916976 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.916987 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.917395 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.917696 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.917482 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.918205 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.918439 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.924552 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.925418 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.926615 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.926722 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.926730 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.926917 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.926951 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.927109 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.927120 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.927197 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.927219 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.927197 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.927218 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.927227 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.927428 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.927493 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.927572 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.927626 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.927639 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.927888 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.929018 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.929162 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.929243 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.929243 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.929440 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.929450 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.929607 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.930186 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.930220 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.930240 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.929929 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.930508 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.930565 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.930882 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.931263 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.931724 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.932237 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.932699 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.932760 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.932869 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.933043 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.933238 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.933298 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.933319 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.933346 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.933748 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.933920 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.934032 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.934128 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.934263 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: E0126 16:47:31.934578 4754 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Jan 26 16:47:31 crc kubenswrapper[4754]: E0126 16:47:31.934641 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-26 16:47:32.43461887 +0000 UTC m=+18.958799304 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.934948 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.935295 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.935352 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: E0126 16:47:31.935552 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 16:47:32.435535104 +0000 UTC m=+18.959715738 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.935745 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.935821 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: E0126 16:47:31.936307 4754 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.936491 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: E0126 16:47:31.936508 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-26 16:47:32.436493181 +0000 UTC m=+18.960673815 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.936317 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.936825 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.937023 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.937100 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.937178 4754 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.937612 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.939076 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.939322 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.939388 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.939556 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.943449 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.945031 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.946858 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.946996 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.948496 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: E0126 16:47:31.950466 4754 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 26 16:47:31 crc kubenswrapper[4754]: E0126 16:47:31.950498 4754 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 26 16:47:31 crc kubenswrapper[4754]: E0126 16:47:31.950515 4754 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 26 16:47:31 crc kubenswrapper[4754]: E0126 16:47:31.950606 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-01-26 16:47:32.450576847 +0000 UTC m=+18.974757281 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.950939 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.951784 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.951918 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.952709 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.958532 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.960249 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.961410 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: E0126 16:47:31.962248 4754 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 26 16:47:31 crc kubenswrapper[4754]: E0126 16:47:31.962279 4754 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 26 16:47:31 crc kubenswrapper[4754]: E0126 16:47:31.962298 4754 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 26 16:47:31 crc kubenswrapper[4754]: E0126 16:47:31.962412 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-01-26 16:47:32.462378252 +0000 UTC m=+18.986558896 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.963168 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.963564 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.963609 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.963639 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.963717 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.963880 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.966227 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.966328 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.966774 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.966792 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.966909 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.967147 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.967166 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.967213 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.967659 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.967727 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.967894 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.967941 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.968018 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.968160 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.968179 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.968558 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.968774 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.968841 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.968868 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.968960 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.968967 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.969184 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.969640 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.969916 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.970228 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.970301 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.970236 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.970369 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.970463 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.971111 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.971311 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.971624 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.971680 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.971723 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.971833 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.972107 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.973339 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.973387 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.973417 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.973487 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.973564 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.973990 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.974177 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.974610 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.974644 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.974955 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.971710 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.975555 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.975911 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.976755 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.976844 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.976944 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.977139 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.977495 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.977569 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.977651 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.978709 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.978718 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.978791 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.978797 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.978955 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.979051 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.978979 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.979093 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.979319 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.979438 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.979484 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.979635 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.979772 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.980498 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.980579 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.981595 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.981833 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.982118 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.982309 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.982955 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.982996 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.983188 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.983641 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.984340 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.993260 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 16:47:31 crc kubenswrapper[4754]: I0126 16:47:31.998876 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.004461 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.011136 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.011181 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.011229 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.011240 4754 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.011250 4754 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.011259 4754 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.011268 4754 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.011277 4754 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.011286 4754 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.011296 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.011308 4754 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.011319 4754 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.011330 4754 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.011340 4754 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.011351 4754 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.011361 4754 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.011354 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.011409 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.011371 4754 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.011479 4754 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.011493 4754 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.011513 4754 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.011527 4754 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.011540 4754 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.011551 4754 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.011561 4754 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.011573 4754 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.011584 4754 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.011596 4754 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.011607 4754 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.011618 4754 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.011629 4754 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.011641 4754 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.011655 4754 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.011692 4754 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.011703 4754 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.011714 4754 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.011724 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.011735 4754 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.011746 4754 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.011756 4754 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.011769 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.011781 4754 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.011791 4754 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.011802 4754 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.011816 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.011829 4754 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.011840 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.011852 4754 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.011862 4754 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.011873 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.011883 4754 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.011895 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.011905 4754 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.011915 4754 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.011927 4754 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.011937 4754 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.011949 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.011960 4754 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.011971 4754 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.011988 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.011999 4754 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.012010 4754 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.012022 4754 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.012033 4754 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.012042 4754 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.012052 4754 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.012063 4754 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.012072 4754 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.012082 4754 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.012093 4754 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.012102 4754 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.012111 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.012121 4754 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.012130 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.012140 4754 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.012151 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.012162 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.012173 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.012184 4754 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.012194 4754 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.012205 4754 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.012217 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.012229 4754 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.012240 4754 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.012249 4754 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.012259 4754 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.012271 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.012281 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.012290 4754 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.012300 4754 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.012310 4754 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.012321 4754 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.012334 4754 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.012344 4754 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.012355 4754 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.012367 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.012381 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.012393 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.012404 4754 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.012414 4754 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.012425 4754 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.012438 4754 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.012450 4754 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.012461 4754 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.012471 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.012480 4754 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.012489 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.012499 4754 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.012509 4754 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.012518 4754 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.012529 4754 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.012540 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.012550 4754 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.012560 4754 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.012569 4754 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.012578 4754 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.012594 4754 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.012603 4754 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.012614 4754 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.012624 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.012637 4754 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.012648 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.012659 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.012688 4754 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.012699 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.012710 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.012721 4754 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.012732 4754 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.012743 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.012754 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.012765 4754 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.012776 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.012788 4754 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.012798 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.012809 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.012822 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.012833 4754 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.012845 4754 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.012856 4754 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.012867 4754 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.012878 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.012888 4754 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.012901 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.012911 4754 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.012926 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.012936 4754 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.012947 4754 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.012958 4754 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.012972 4754 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.012981 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.012992 4754 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.013001 4754 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.013010 4754 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.013019 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.013028 4754 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.013037 4754 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.013103 4754 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.013124 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.013134 4754 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.013143 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.013152 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.013161 4754 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.013169 4754 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.013178 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.013187 4754 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.013196 4754 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.013205 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.013213 4754 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.013222 4754 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.013252 4754 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.013261 4754 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.013269 4754 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.013278 4754 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.013286 4754 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.013294 4754 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.013303 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.013312 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.013319 4754 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.013327 4754 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.013335 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.013345 4754 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.013353 4754 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.013361 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.013369 4754 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.013377 4754 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.013385 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.013393 4754 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.013401 4754 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.013409 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.013418 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.013426 4754 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.013434 4754 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.013442 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.013451 4754 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.024442 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.036153 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 26 16:47:32 crc kubenswrapper[4754]: W0126 16:47:32.044645 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod37a5e44f_9a88_4405_be8a_b645485e7312.slice/crio-b03d492b71f1f82f5f444d5caa72876d450d01d7848dd97c42d61002bdf33707 WatchSource:0}: Error finding container b03d492b71f1f82f5f444d5caa72876d450d01d7848dd97c42d61002bdf33707: Status 404 returned error can't find the container with id b03d492b71f1f82f5f444d5caa72876d450d01d7848dd97c42d61002bdf33707 Jan 26 16:47:32 crc kubenswrapper[4754]: W0126 16:47:32.045467 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podef543e1b_8068_4ea3_b32a_61027b32e95d.slice/crio-466b9dd7f9f06ac3258e00b095ebaf63c8e451fa5b48e55d1cfe33e3ef7492c4 WatchSource:0}: Error finding container 466b9dd7f9f06ac3258e00b095ebaf63c8e451fa5b48e55d1cfe33e3ef7492c4: Status 404 returned error can't find the container with id 466b9dd7f9f06ac3258e00b095ebaf63c8e451fa5b48e55d1cfe33e3ef7492c4 Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.051241 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 26 16:47:32 crc kubenswrapper[4754]: W0126 16:47:32.073704 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd75a4c96_2883_4a0b_bab2_0fab2b6c0b49.slice/crio-55075f9e670a9de53be4a05ff63a3af2e36aae767073d5f15262fc71e77f032d WatchSource:0}: Error finding container 55075f9e670a9de53be4a05ff63a3af2e36aae767073d5f15262fc71e77f032d: Status 404 returned error can't find the container with id 55075f9e670a9de53be4a05ff63a3af2e36aae767073d5f15262fc71e77f032d Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.516857 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.516921 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.516940 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.516960 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.516980 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 16:47:32 crc kubenswrapper[4754]: E0126 16:47:32.517075 4754 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 26 16:47:32 crc kubenswrapper[4754]: E0126 16:47:32.517098 4754 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 26 16:47:32 crc kubenswrapper[4754]: E0126 16:47:32.517109 4754 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 26 16:47:32 crc kubenswrapper[4754]: E0126 16:47:32.517150 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-01-26 16:47:33.517138368 +0000 UTC m=+20.041318792 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 26 16:47:32 crc kubenswrapper[4754]: E0126 16:47:32.517518 4754 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 26 16:47:32 crc kubenswrapper[4754]: E0126 16:47:32.517556 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-26 16:47:33.517546309 +0000 UTC m=+20.041726743 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 26 16:47:32 crc kubenswrapper[4754]: E0126 16:47:32.517568 4754 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Jan 26 16:47:32 crc kubenswrapper[4754]: E0126 16:47:32.517651 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-26 16:47:33.517631861 +0000 UTC m=+20.041812295 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Jan 26 16:47:32 crc kubenswrapper[4754]: E0126 16:47:32.517693 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 16:47:33.517683863 +0000 UTC m=+20.041864377 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 16:47:32 crc kubenswrapper[4754]: E0126 16:47:32.517733 4754 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 26 16:47:32 crc kubenswrapper[4754]: E0126 16:47:32.517753 4754 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 26 16:47:32 crc kubenswrapper[4754]: E0126 16:47:32.517768 4754 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 26 16:47:32 crc kubenswrapper[4754]: E0126 16:47:32.517812 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-01-26 16:47:33.517801676 +0000 UTC m=+20.041982180 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.720135 4754 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-07 17:23:18.227252072 +0000 UTC Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.864211 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.867345 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"8c40e018a0979035827963efb9efc48a878a55b9df68ec1a9242d7cff64fe480"} Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.867688 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.868840 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"55075f9e670a9de53be4a05ff63a3af2e36aae767073d5f15262fc71e77f032d"} Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.870825 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"89251fa9c64bc25d6d6f355400d7644c099f22aed1cae812fc991a03595f1e8f"} Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.870861 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"8e90508d8a53be9146f9b388750730fdae9ee200306caf5ad393a356390a69ed"} Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.870889 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"466b9dd7f9f06ac3258e00b095ebaf63c8e451fa5b48e55d1cfe33e3ef7492c4"} Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.873006 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"205ec256e46a8e08b4b0b0805fb6c5f4966fc3ccef0738df4b97cb404e74da17"} Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.873042 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"b03d492b71f1f82f5f444d5caa72876d450d01d7848dd97c42d61002bdf33707"} Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.902312 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:32Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.921278 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b89aa081-0a2e-4757-9210-e2a6c228bee9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9c5f40188fda14b14128514fe84bddfe8dcede3d6c560283ccb2ac0c66895fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f45445cbf1d180d6d4dd9588e85bf678f043bfce8cff0812b78ca3510904f9a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://085fa7b8d4fef8b2722fac8c346107699b5a6bd33fbfc74ed35831cb8fdf9dad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c40e018a0979035827963efb9efc48a878a55b9df68ec1a9242d7cff64fe480\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8356f76fbc3a4e636c12d1da5455d3f6767d9b83ce2f2925a11435d7d7261e9f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0126 16:47:26.205208 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0126 16:47:26.206538 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3645934156/tls.crt::/tmp/serving-cert-3645934156/tls.key\\\\\\\"\\\\nI0126 16:47:31.504518 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0126 16:47:31.509366 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0126 16:47:31.509400 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0126 16:47:31.509428 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0126 16:47:31.509436 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0126 16:47:31.515912 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0126 16:47:31.516036 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 16:47:31.516068 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 16:47:31.516096 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0126 16:47:31.516121 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0126 16:47:31.516145 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0126 16:47:31.516170 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0126 16:47:31.516412 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0126 16:47:31.519294 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27b74e2e65a547e0e28fddd4f15feacb17f4d57596eeb8da82df9d39243c85e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22ba7649a1308815bfad52016000987a7a9a5bc214d7ff61e137ba099b20fc6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://22ba7649a1308815bfad52016000987a7a9a5bc214d7ff61e137ba099b20fc6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:13Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:32Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.938480 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:32Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.954289 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:32Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:32 crc kubenswrapper[4754]: I0126 16:47:32.991426 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:32Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.027690 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:33Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.050919 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:33Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.070016 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:33Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.091480 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89251fa9c64bc25d6d6f355400d7644c099f22aed1cae812fc991a03595f1e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e90508d8a53be9146f9b388750730fdae9ee200306caf5ad393a356390a69ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:33Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.115867 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:33Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.136976 4754 csr.go:261] certificate signing request csr-bjpm7 is approved, waiting to be issued Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.139185 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:33Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.154572 4754 csr.go:257] certificate signing request csr-bjpm7 is issued Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.165162 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:33Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.184651 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b89aa081-0a2e-4757-9210-e2a6c228bee9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9c5f40188fda14b14128514fe84bddfe8dcede3d6c560283ccb2ac0c66895fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f45445cbf1d180d6d4dd9588e85bf678f043bfce8cff0812b78ca3510904f9a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://085fa7b8d4fef8b2722fac8c346107699b5a6bd33fbfc74ed35831cb8fdf9dad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c40e018a0979035827963efb9efc48a878a55b9df68ec1a9242d7cff64fe480\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8356f76fbc3a4e636c12d1da5455d3f6767d9b83ce2f2925a11435d7d7261e9f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0126 16:47:26.205208 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0126 16:47:26.206538 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3645934156/tls.crt::/tmp/serving-cert-3645934156/tls.key\\\\\\\"\\\\nI0126 16:47:31.504518 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0126 16:47:31.509366 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0126 16:47:31.509400 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0126 16:47:31.509428 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0126 16:47:31.509436 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0126 16:47:31.515912 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0126 16:47:31.516036 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 16:47:31.516068 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 16:47:31.516096 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0126 16:47:31.516121 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0126 16:47:31.516145 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0126 16:47:31.516170 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0126 16:47:31.516412 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0126 16:47:31.519294 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27b74e2e65a547e0e28fddd4f15feacb17f4d57596eeb8da82df9d39243c85e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22ba7649a1308815bfad52016000987a7a9a5bc214d7ff61e137ba099b20fc6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://22ba7649a1308815bfad52016000987a7a9a5bc214d7ff61e137ba099b20fc6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:13Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:33Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.206113 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://205ec256e46a8e08b4b0b0805fb6c5f4966fc3ccef0738df4b97cb404e74da17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:33Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.238082 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-2blzw"] Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.238494 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-2blzw" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.240403 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-84lgq"] Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.241057 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-84lgq" Jan 26 16:47:33 crc kubenswrapper[4754]: W0126 16:47:33.242250 4754 reflector.go:561] object-"openshift-image-registry"/"image-registry-certificates": failed to list *v1.ConfigMap: configmaps "image-registry-certificates" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-image-registry": no relationship found between node 'crc' and this object Jan 26 16:47:33 crc kubenswrapper[4754]: E0126 16:47:33.242300 4754 reflector.go:158] "Unhandled Error" err="object-\"openshift-image-registry\"/\"image-registry-certificates\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"image-registry-certificates\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-image-registry\": no relationship found between node 'crc' and this object" logger="UnhandledError" Jan 26 16:47:33 crc kubenswrapper[4754]: W0126 16:47:33.242342 4754 reflector.go:561] object-"openshift-image-registry"/"openshift-service-ca.crt": failed to list *v1.ConfigMap: configmaps "openshift-service-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-image-registry": no relationship found between node 'crc' and this object Jan 26 16:47:33 crc kubenswrapper[4754]: E0126 16:47:33.242395 4754 reflector.go:158] "Unhandled Error" err="object-\"openshift-image-registry\"/\"openshift-service-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"openshift-service-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-image-registry\": no relationship found between node 'crc' and this object" logger="UnhandledError" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.243896 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.244968 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.244970 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.245282 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.245343 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.260888 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:33Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.275215 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2blzw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca65468e-30c6-4666-962e-cc0de05e67fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c25z4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:33Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2blzw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:33Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.291945 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b89aa081-0a2e-4757-9210-e2a6c228bee9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9c5f40188fda14b14128514fe84bddfe8dcede3d6c560283ccb2ac0c66895fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f45445cbf1d180d6d4dd9588e85bf678f043bfce8cff0812b78ca3510904f9a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://085fa7b8d4fef8b2722fac8c346107699b5a6bd33fbfc74ed35831cb8fdf9dad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c40e018a0979035827963efb9efc48a878a55b9df68ec1a9242d7cff64fe480\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8356f76fbc3a4e636c12d1da5455d3f6767d9b83ce2f2925a11435d7d7261e9f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0126 16:47:26.205208 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0126 16:47:26.206538 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3645934156/tls.crt::/tmp/serving-cert-3645934156/tls.key\\\\\\\"\\\\nI0126 16:47:31.504518 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0126 16:47:31.509366 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0126 16:47:31.509400 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0126 16:47:31.509428 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0126 16:47:31.509436 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0126 16:47:31.515912 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0126 16:47:31.516036 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 16:47:31.516068 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 16:47:31.516096 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0126 16:47:31.516121 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0126 16:47:31.516145 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0126 16:47:31.516170 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0126 16:47:31.516412 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0126 16:47:31.519294 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27b74e2e65a547e0e28fddd4f15feacb17f4d57596eeb8da82df9d39243c85e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22ba7649a1308815bfad52016000987a7a9a5bc214d7ff61e137ba099b20fc6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://22ba7649a1308815bfad52016000987a7a9a5bc214d7ff61e137ba099b20fc6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:13Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:33Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.307286 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://205ec256e46a8e08b4b0b0805fb6c5f4966fc3ccef0738df4b97cb404e74da17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:33Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.323182 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/6763d962-0892-4acc-bab9-7a4e93e39a85-hosts-file\") pod \"node-resolver-84lgq\" (UID: \"6763d962-0892-4acc-bab9-7a4e93e39a85\") " pod="openshift-dns/node-resolver-84lgq" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.323233 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ca65468e-30c6-4666-962e-cc0de05e67fe-host\") pod \"node-ca-2blzw\" (UID: \"ca65468e-30c6-4666-962e-cc0de05e67fe\") " pod="openshift-image-registry/node-ca-2blzw" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.323283 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c25z4\" (UniqueName: \"kubernetes.io/projected/ca65468e-30c6-4666-962e-cc0de05e67fe-kube-api-access-c25z4\") pod \"node-ca-2blzw\" (UID: \"ca65468e-30c6-4666-962e-cc0de05e67fe\") " pod="openshift-image-registry/node-ca-2blzw" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.323311 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/ca65468e-30c6-4666-962e-cc0de05e67fe-serviceca\") pod \"node-ca-2blzw\" (UID: \"ca65468e-30c6-4666-962e-cc0de05e67fe\") " pod="openshift-image-registry/node-ca-2blzw" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.323330 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8zlw6\" (UniqueName: \"kubernetes.io/projected/6763d962-0892-4acc-bab9-7a4e93e39a85-kube-api-access-8zlw6\") pod \"node-resolver-84lgq\" (UID: \"6763d962-0892-4acc-bab9-7a4e93e39a85\") " pod="openshift-dns/node-resolver-84lgq" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.324423 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:33Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.360772 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:33Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.418413 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:33Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.423988 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c25z4\" (UniqueName: \"kubernetes.io/projected/ca65468e-30c6-4666-962e-cc0de05e67fe-kube-api-access-c25z4\") pod \"node-ca-2blzw\" (UID: \"ca65468e-30c6-4666-962e-cc0de05e67fe\") " pod="openshift-image-registry/node-ca-2blzw" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.424082 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/ca65468e-30c6-4666-962e-cc0de05e67fe-serviceca\") pod \"node-ca-2blzw\" (UID: \"ca65468e-30c6-4666-962e-cc0de05e67fe\") " pod="openshift-image-registry/node-ca-2blzw" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.424116 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8zlw6\" (UniqueName: \"kubernetes.io/projected/6763d962-0892-4acc-bab9-7a4e93e39a85-kube-api-access-8zlw6\") pod \"node-resolver-84lgq\" (UID: \"6763d962-0892-4acc-bab9-7a4e93e39a85\") " pod="openshift-dns/node-resolver-84lgq" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.424160 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/6763d962-0892-4acc-bab9-7a4e93e39a85-hosts-file\") pod \"node-resolver-84lgq\" (UID: \"6763d962-0892-4acc-bab9-7a4e93e39a85\") " pod="openshift-dns/node-resolver-84lgq" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.424189 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ca65468e-30c6-4666-962e-cc0de05e67fe-host\") pod \"node-ca-2blzw\" (UID: \"ca65468e-30c6-4666-962e-cc0de05e67fe\") " pod="openshift-image-registry/node-ca-2blzw" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.424286 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ca65468e-30c6-4666-962e-cc0de05e67fe-host\") pod \"node-ca-2blzw\" (UID: \"ca65468e-30c6-4666-962e-cc0de05e67fe\") " pod="openshift-image-registry/node-ca-2blzw" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.424343 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/6763d962-0892-4acc-bab9-7a4e93e39a85-hosts-file\") pod \"node-resolver-84lgq\" (UID: \"6763d962-0892-4acc-bab9-7a4e93e39a85\") " pod="openshift-dns/node-resolver-84lgq" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.453418 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8zlw6\" (UniqueName: \"kubernetes.io/projected/6763d962-0892-4acc-bab9-7a4e93e39a85-kube-api-access-8zlw6\") pod \"node-resolver-84lgq\" (UID: \"6763d962-0892-4acc-bab9-7a4e93e39a85\") " pod="openshift-dns/node-resolver-84lgq" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.461963 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89251fa9c64bc25d6d6f355400d7644c099f22aed1cae812fc991a03595f1e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e90508d8a53be9146f9b388750730fdae9ee200306caf5ad393a356390a69ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:33Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.484241 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b89aa081-0a2e-4757-9210-e2a6c228bee9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9c5f40188fda14b14128514fe84bddfe8dcede3d6c560283ccb2ac0c66895fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f45445cbf1d180d6d4dd9588e85bf678f043bfce8cff0812b78ca3510904f9a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://085fa7b8d4fef8b2722fac8c346107699b5a6bd33fbfc74ed35831cb8fdf9dad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c40e018a0979035827963efb9efc48a878a55b9df68ec1a9242d7cff64fe480\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8356f76fbc3a4e636c12d1da5455d3f6767d9b83ce2f2925a11435d7d7261e9f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0126 16:47:26.205208 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0126 16:47:26.206538 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3645934156/tls.crt::/tmp/serving-cert-3645934156/tls.key\\\\\\\"\\\\nI0126 16:47:31.504518 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0126 16:47:31.509366 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0126 16:47:31.509400 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0126 16:47:31.509428 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0126 16:47:31.509436 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0126 16:47:31.515912 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0126 16:47:31.516036 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 16:47:31.516068 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 16:47:31.516096 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0126 16:47:31.516121 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0126 16:47:31.516145 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0126 16:47:31.516170 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0126 16:47:31.516412 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0126 16:47:31.519294 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27b74e2e65a547e0e28fddd4f15feacb17f4d57596eeb8da82df9d39243c85e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22ba7649a1308815bfad52016000987a7a9a5bc214d7ff61e137ba099b20fc6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://22ba7649a1308815bfad52016000987a7a9a5bc214d7ff61e137ba099b20fc6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:13Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:33Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.507502 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://205ec256e46a8e08b4b0b0805fb6c5f4966fc3ccef0738df4b97cb404e74da17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:33Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.525489 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.525560 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.525587 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.525615 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.525636 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 16:47:33 crc kubenswrapper[4754]: E0126 16:47:33.525772 4754 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 26 16:47:33 crc kubenswrapper[4754]: E0126 16:47:33.525767 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 16:47:35.525729346 +0000 UTC m=+22.049909780 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 16:47:33 crc kubenswrapper[4754]: E0126 16:47:33.525789 4754 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 26 16:47:33 crc kubenswrapper[4754]: E0126 16:47:33.525833 4754 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 26 16:47:33 crc kubenswrapper[4754]: E0126 16:47:33.525856 4754 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 26 16:47:33 crc kubenswrapper[4754]: E0126 16:47:33.525884 4754 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 26 16:47:33 crc kubenswrapper[4754]: E0126 16:47:33.525901 4754 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 26 16:47:33 crc kubenswrapper[4754]: E0126 16:47:33.525886 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-01-26 16:47:35.52588061 +0000 UTC m=+22.050061044 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 26 16:47:33 crc kubenswrapper[4754]: E0126 16:47:33.525876 4754 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Jan 26 16:47:33 crc kubenswrapper[4754]: E0126 16:47:33.526067 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-26 16:47:35.526034934 +0000 UTC m=+22.050215518 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 26 16:47:33 crc kubenswrapper[4754]: E0126 16:47:33.526113 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-26 16:47:35.526094666 +0000 UTC m=+22.050275100 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Jan 26 16:47:33 crc kubenswrapper[4754]: E0126 16:47:33.525915 4754 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 26 16:47:33 crc kubenswrapper[4754]: E0126 16:47:33.526183 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-01-26 16:47:35.526170168 +0000 UTC m=+22.050350812 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.533479 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:33Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.549438 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:33Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.562900 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-84lgq" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.563851 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:33Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:33 crc kubenswrapper[4754]: W0126 16:47:33.585779 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6763d962_0892_4acc_bab9_7a4e93e39a85.slice/crio-6f0338d5f1927deb680d619f4137312544a1682aac3ad3dfa917366b5cc5c35c WatchSource:0}: Error finding container 6f0338d5f1927deb680d619f4137312544a1682aac3ad3dfa917366b5cc5c35c: Status 404 returned error can't find the container with id 6f0338d5f1927deb680d619f4137312544a1682aac3ad3dfa917366b5cc5c35c Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.586064 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2blzw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca65468e-30c6-4666-962e-cc0de05e67fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c25z4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:33Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2blzw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:33Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.601316 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:33Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.617004 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89251fa9c64bc25d6d6f355400d7644c099f22aed1cae812fc991a03595f1e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e90508d8a53be9146f9b388750730fdae9ee200306caf5ad393a356390a69ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:33Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.626587 4754 transport.go:147] "Certificate rotation detected, shutting down client connections to start using new credentials" Jan 26 16:47:33 crc kubenswrapper[4754]: W0126 16:47:33.626832 4754 reflector.go:484] object-"openshift-dns"/"openshift-service-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-dns"/"openshift-service-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Jan 26 16:47:33 crc kubenswrapper[4754]: W0126 16:47:33.626896 4754 reflector.go:484] object-"openshift-dns"/"node-resolver-dockercfg-kz9s7": watch of *v1.Secret ended with: very short watch: object-"openshift-dns"/"node-resolver-dockercfg-kz9s7": Unexpected watch close - watch lasted less than a second and no items received Jan 26 16:47:33 crc kubenswrapper[4754]: W0126 16:47:33.627030 4754 reflector.go:484] object-"openshift-image-registry"/"kube-root-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-image-registry"/"kube-root-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Jan 26 16:47:33 crc kubenswrapper[4754]: W0126 16:47:33.627036 4754 reflector.go:484] object-"openshift-dns"/"kube-root-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-dns"/"kube-root-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Jan 26 16:47:33 crc kubenswrapper[4754]: W0126 16:47:33.627401 4754 reflector.go:484] object-"openshift-image-registry"/"node-ca-dockercfg-4777p": watch of *v1.Secret ended with: very short watch: object-"openshift-image-registry"/"node-ca-dockercfg-4777p": Unexpected watch close - watch lasted less than a second and no items received Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.633500 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-84lgq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6763d962-0892-4acc-bab9-7a4e93e39a85\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8zlw6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:33Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-84lgq\": Patch \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-dns/pods/node-resolver-84lgq/status\": read tcp 38.102.83.236:47588->38.102.83.236:6443: use of closed network connection" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.669321 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-skh5n"] Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.669736 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-x65wv"] Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.670013 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.670244 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-skh5n" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.674381 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.676895 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.677121 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.677255 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.677396 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.677525 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.677677 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.677890 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.678547 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.679038 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.692175 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:33Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.708290 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89251fa9c64bc25d6d6f355400d7644c099f22aed1cae812fc991a03595f1e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e90508d8a53be9146f9b388750730fdae9ee200306caf5ad393a356390a69ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:33Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.720914 4754 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-16 05:27:56.345680825 +0000 UTC Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.723892 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-84lgq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6763d962-0892-4acc-bab9-7a4e93e39a85\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8zlw6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:33Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-84lgq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:33Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.728353 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/8c3718a4-f354-4284-92e0-fdfb45a692bd-proxy-tls\") pod \"machine-config-daemon-x65wv\" (UID: \"8c3718a4-f354-4284-92e0-fdfb45a692bd\") " pod="openshift-machine-config-operator/machine-config-daemon-x65wv" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.728432 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b619d233-b592-4b05-a0b7-dc094c88471e-etc-kubernetes\") pod \"multus-skh5n\" (UID: \"b619d233-b592-4b05-a0b7-dc094c88471e\") " pod="openshift-multus/multus-skh5n" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.728459 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/b619d233-b592-4b05-a0b7-dc094c88471e-host-var-lib-cni-bin\") pod \"multus-skh5n\" (UID: \"b619d233-b592-4b05-a0b7-dc094c88471e\") " pod="openshift-multus/multus-skh5n" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.728521 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/b619d233-b592-4b05-a0b7-dc094c88471e-multus-daemon-config\") pod \"multus-skh5n\" (UID: \"b619d233-b592-4b05-a0b7-dc094c88471e\") " pod="openshift-multus/multus-skh5n" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.728571 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/b619d233-b592-4b05-a0b7-dc094c88471e-host-run-multus-certs\") pod \"multus-skh5n\" (UID: \"b619d233-b592-4b05-a0b7-dc094c88471e\") " pod="openshift-multus/multus-skh5n" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.728599 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/b619d233-b592-4b05-a0b7-dc094c88471e-host-run-k8s-cni-cncf-io\") pod \"multus-skh5n\" (UID: \"b619d233-b592-4b05-a0b7-dc094c88471e\") " pod="openshift-multus/multus-skh5n" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.728828 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/b619d233-b592-4b05-a0b7-dc094c88471e-system-cni-dir\") pod \"multus-skh5n\" (UID: \"b619d233-b592-4b05-a0b7-dc094c88471e\") " pod="openshift-multus/multus-skh5n" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.728864 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/b619d233-b592-4b05-a0b7-dc094c88471e-hostroot\") pod \"multus-skh5n\" (UID: \"b619d233-b592-4b05-a0b7-dc094c88471e\") " pod="openshift-multus/multus-skh5n" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.728932 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/b619d233-b592-4b05-a0b7-dc094c88471e-host-var-lib-cni-multus\") pod \"multus-skh5n\" (UID: \"b619d233-b592-4b05-a0b7-dc094c88471e\") " pod="openshift-multus/multus-skh5n" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.728960 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/b619d233-b592-4b05-a0b7-dc094c88471e-host-run-netns\") pod \"multus-skh5n\" (UID: \"b619d233-b592-4b05-a0b7-dc094c88471e\") " pod="openshift-multus/multus-skh5n" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.729009 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/b619d233-b592-4b05-a0b7-dc094c88471e-multus-cni-dir\") pod \"multus-skh5n\" (UID: \"b619d233-b592-4b05-a0b7-dc094c88471e\") " pod="openshift-multus/multus-skh5n" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.729037 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/b619d233-b592-4b05-a0b7-dc094c88471e-os-release\") pod \"multus-skh5n\" (UID: \"b619d233-b592-4b05-a0b7-dc094c88471e\") " pod="openshift-multus/multus-skh5n" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.729088 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/b619d233-b592-4b05-a0b7-dc094c88471e-multus-socket-dir-parent\") pod \"multus-skh5n\" (UID: \"b619d233-b592-4b05-a0b7-dc094c88471e\") " pod="openshift-multus/multus-skh5n" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.729114 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/b619d233-b592-4b05-a0b7-dc094c88471e-multus-conf-dir\") pod \"multus-skh5n\" (UID: \"b619d233-b592-4b05-a0b7-dc094c88471e\") " pod="openshift-multus/multus-skh5n" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.729169 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/b619d233-b592-4b05-a0b7-dc094c88471e-cni-binary-copy\") pod \"multus-skh5n\" (UID: \"b619d233-b592-4b05-a0b7-dc094c88471e\") " pod="openshift-multus/multus-skh5n" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.729193 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/b619d233-b592-4b05-a0b7-dc094c88471e-host-var-lib-kubelet\") pod \"multus-skh5n\" (UID: \"b619d233-b592-4b05-a0b7-dc094c88471e\") " pod="openshift-multus/multus-skh5n" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.729240 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sxjcz\" (UniqueName: \"kubernetes.io/projected/8c3718a4-f354-4284-92e0-fdfb45a692bd-kube-api-access-sxjcz\") pod \"machine-config-daemon-x65wv\" (UID: \"8c3718a4-f354-4284-92e0-fdfb45a692bd\") " pod="openshift-machine-config-operator/machine-config-daemon-x65wv" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.729261 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/b619d233-b592-4b05-a0b7-dc094c88471e-cnibin\") pod \"multus-skh5n\" (UID: \"b619d233-b592-4b05-a0b7-dc094c88471e\") " pod="openshift-multus/multus-skh5n" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.729304 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6tj78\" (UniqueName: \"kubernetes.io/projected/b619d233-b592-4b05-a0b7-dc094c88471e-kube-api-access-6tj78\") pod \"multus-skh5n\" (UID: \"b619d233-b592-4b05-a0b7-dc094c88471e\") " pod="openshift-multus/multus-skh5n" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.729329 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/8c3718a4-f354-4284-92e0-fdfb45a692bd-rootfs\") pod \"machine-config-daemon-x65wv\" (UID: \"8c3718a4-f354-4284-92e0-fdfb45a692bd\") " pod="openshift-machine-config-operator/machine-config-daemon-x65wv" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.729355 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/8c3718a4-f354-4284-92e0-fdfb45a692bd-mcd-auth-proxy-config\") pod \"machine-config-daemon-x65wv\" (UID: \"8c3718a4-f354-4284-92e0-fdfb45a692bd\") " pod="openshift-machine-config-operator/machine-config-daemon-x65wv" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.759527 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-skh5n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b619d233-b592-4b05-a0b7-dc094c88471e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-skh5n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:33Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.767101 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.767199 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 16:47:33 crc kubenswrapper[4754]: E0126 16:47:33.767230 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.767237 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 16:47:33 crc kubenswrapper[4754]: E0126 16:47:33.767414 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 16:47:33 crc kubenswrapper[4754]: E0126 16:47:33.767517 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.771874 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.772604 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.774120 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.775234 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.776578 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.777241 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.778020 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.779257 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.780066 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.782083 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.782763 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.784037 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.784623 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.786540 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.787212 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.788255 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.788900 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.789817 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.790624 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.791448 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.795007 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.795773 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.796259 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.797976 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.799253 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.800956 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.801854 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.802517 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.803933 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.804631 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.806164 4754 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.806334 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.808537 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b89aa081-0a2e-4757-9210-e2a6c228bee9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9c5f40188fda14b14128514fe84bddfe8dcede3d6c560283ccb2ac0c66895fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f45445cbf1d180d6d4dd9588e85bf678f043bfce8cff0812b78ca3510904f9a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://085fa7b8d4fef8b2722fac8c346107699b5a6bd33fbfc74ed35831cb8fdf9dad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c40e018a0979035827963efb9efc48a878a55b9df68ec1a9242d7cff64fe480\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8356f76fbc3a4e636c12d1da5455d3f6767d9b83ce2f2925a11435d7d7261e9f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0126 16:47:26.205208 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0126 16:47:26.206538 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3645934156/tls.crt::/tmp/serving-cert-3645934156/tls.key\\\\\\\"\\\\nI0126 16:47:31.504518 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0126 16:47:31.509366 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0126 16:47:31.509400 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0126 16:47:31.509428 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0126 16:47:31.509436 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0126 16:47:31.515912 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0126 16:47:31.516036 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 16:47:31.516068 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 16:47:31.516096 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0126 16:47:31.516121 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0126 16:47:31.516145 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0126 16:47:31.516170 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0126 16:47:31.516412 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0126 16:47:31.519294 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27b74e2e65a547e0e28fddd4f15feacb17f4d57596eeb8da82df9d39243c85e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22ba7649a1308815bfad52016000987a7a9a5bc214d7ff61e137ba099b20fc6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://22ba7649a1308815bfad52016000987a7a9a5bc214d7ff61e137ba099b20fc6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:13Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:33Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.811812 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.813187 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.813784 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.816261 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.817078 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.818274 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.819251 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.820688 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.821304 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.822861 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.824175 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.825039 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.825722 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.827258 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.828531 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.829742 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/b619d233-b592-4b05-a0b7-dc094c88471e-multus-daemon-config\") pod \"multus-skh5n\" (UID: \"b619d233-b592-4b05-a0b7-dc094c88471e\") " pod="openshift-multus/multus-skh5n" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.829788 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/b619d233-b592-4b05-a0b7-dc094c88471e-host-run-k8s-cni-cncf-io\") pod \"multus-skh5n\" (UID: \"b619d233-b592-4b05-a0b7-dc094c88471e\") " pod="openshift-multus/multus-skh5n" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.829809 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.829816 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/b619d233-b592-4b05-a0b7-dc094c88471e-host-run-multus-certs\") pod \"multus-skh5n\" (UID: \"b619d233-b592-4b05-a0b7-dc094c88471e\") " pod="openshift-multus/multus-skh5n" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.829862 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/b619d233-b592-4b05-a0b7-dc094c88471e-host-run-multus-certs\") pod \"multus-skh5n\" (UID: \"b619d233-b592-4b05-a0b7-dc094c88471e\") " pod="openshift-multus/multus-skh5n" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.829921 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/b619d233-b592-4b05-a0b7-dc094c88471e-system-cni-dir\") pod \"multus-skh5n\" (UID: \"b619d233-b592-4b05-a0b7-dc094c88471e\") " pod="openshift-multus/multus-skh5n" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.829944 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/b619d233-b592-4b05-a0b7-dc094c88471e-hostroot\") pod \"multus-skh5n\" (UID: \"b619d233-b592-4b05-a0b7-dc094c88471e\") " pod="openshift-multus/multus-skh5n" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.829992 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/b619d233-b592-4b05-a0b7-dc094c88471e-host-var-lib-cni-multus\") pod \"multus-skh5n\" (UID: \"b619d233-b592-4b05-a0b7-dc094c88471e\") " pod="openshift-multus/multus-skh5n" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.830026 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/b619d233-b592-4b05-a0b7-dc094c88471e-multus-cni-dir\") pod \"multus-skh5n\" (UID: \"b619d233-b592-4b05-a0b7-dc094c88471e\") " pod="openshift-multus/multus-skh5n" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.830014 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/b619d233-b592-4b05-a0b7-dc094c88471e-host-run-k8s-cni-cncf-io\") pod \"multus-skh5n\" (UID: \"b619d233-b592-4b05-a0b7-dc094c88471e\") " pod="openshift-multus/multus-skh5n" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.830051 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/b619d233-b592-4b05-a0b7-dc094c88471e-host-run-netns\") pod \"multus-skh5n\" (UID: \"b619d233-b592-4b05-a0b7-dc094c88471e\") " pod="openshift-multus/multus-skh5n" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.830087 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/b619d233-b592-4b05-a0b7-dc094c88471e-host-var-lib-cni-multus\") pod \"multus-skh5n\" (UID: \"b619d233-b592-4b05-a0b7-dc094c88471e\") " pod="openshift-multus/multus-skh5n" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.830123 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/b619d233-b592-4b05-a0b7-dc094c88471e-os-release\") pod \"multus-skh5n\" (UID: \"b619d233-b592-4b05-a0b7-dc094c88471e\") " pod="openshift-multus/multus-skh5n" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.830153 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/b619d233-b592-4b05-a0b7-dc094c88471e-multus-socket-dir-parent\") pod \"multus-skh5n\" (UID: \"b619d233-b592-4b05-a0b7-dc094c88471e\") " pod="openshift-multus/multus-skh5n" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.830180 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/b619d233-b592-4b05-a0b7-dc094c88471e-system-cni-dir\") pod \"multus-skh5n\" (UID: \"b619d233-b592-4b05-a0b7-dc094c88471e\") " pod="openshift-multus/multus-skh5n" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.830198 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/b619d233-b592-4b05-a0b7-dc094c88471e-multus-conf-dir\") pod \"multus-skh5n\" (UID: \"b619d233-b592-4b05-a0b7-dc094c88471e\") " pod="openshift-multus/multus-skh5n" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.830209 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/b619d233-b592-4b05-a0b7-dc094c88471e-multus-socket-dir-parent\") pod \"multus-skh5n\" (UID: \"b619d233-b592-4b05-a0b7-dc094c88471e\") " pod="openshift-multus/multus-skh5n" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.830098 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/b619d233-b592-4b05-a0b7-dc094c88471e-hostroot\") pod \"multus-skh5n\" (UID: \"b619d233-b592-4b05-a0b7-dc094c88471e\") " pod="openshift-multus/multus-skh5n" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.830154 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/b619d233-b592-4b05-a0b7-dc094c88471e-host-run-netns\") pod \"multus-skh5n\" (UID: \"b619d233-b592-4b05-a0b7-dc094c88471e\") " pod="openshift-multus/multus-skh5n" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.830240 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/b619d233-b592-4b05-a0b7-dc094c88471e-cni-binary-copy\") pod \"multus-skh5n\" (UID: \"b619d233-b592-4b05-a0b7-dc094c88471e\") " pod="openshift-multus/multus-skh5n" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.830271 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/b619d233-b592-4b05-a0b7-dc094c88471e-host-var-lib-kubelet\") pod \"multus-skh5n\" (UID: \"b619d233-b592-4b05-a0b7-dc094c88471e\") " pod="openshift-multus/multus-skh5n" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.830272 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/b619d233-b592-4b05-a0b7-dc094c88471e-multus-conf-dir\") pod \"multus-skh5n\" (UID: \"b619d233-b592-4b05-a0b7-dc094c88471e\") " pod="openshift-multus/multus-skh5n" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.830301 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sxjcz\" (UniqueName: \"kubernetes.io/projected/8c3718a4-f354-4284-92e0-fdfb45a692bd-kube-api-access-sxjcz\") pod \"machine-config-daemon-x65wv\" (UID: \"8c3718a4-f354-4284-92e0-fdfb45a692bd\") " pod="openshift-machine-config-operator/machine-config-daemon-x65wv" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.830334 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/b619d233-b592-4b05-a0b7-dc094c88471e-cnibin\") pod \"multus-skh5n\" (UID: \"b619d233-b592-4b05-a0b7-dc094c88471e\") " pod="openshift-multus/multus-skh5n" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.830350 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/b619d233-b592-4b05-a0b7-dc094c88471e-os-release\") pod \"multus-skh5n\" (UID: \"b619d233-b592-4b05-a0b7-dc094c88471e\") " pod="openshift-multus/multus-skh5n" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.830385 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/b619d233-b592-4b05-a0b7-dc094c88471e-multus-cni-dir\") pod \"multus-skh5n\" (UID: \"b619d233-b592-4b05-a0b7-dc094c88471e\") " pod="openshift-multus/multus-skh5n" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.830392 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6tj78\" (UniqueName: \"kubernetes.io/projected/b619d233-b592-4b05-a0b7-dc094c88471e-kube-api-access-6tj78\") pod \"multus-skh5n\" (UID: \"b619d233-b592-4b05-a0b7-dc094c88471e\") " pod="openshift-multus/multus-skh5n" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.830427 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/b619d233-b592-4b05-a0b7-dc094c88471e-cnibin\") pod \"multus-skh5n\" (UID: \"b619d233-b592-4b05-a0b7-dc094c88471e\") " pod="openshift-multus/multus-skh5n" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.830357 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/b619d233-b592-4b05-a0b7-dc094c88471e-host-var-lib-kubelet\") pod \"multus-skh5n\" (UID: \"b619d233-b592-4b05-a0b7-dc094c88471e\") " pod="openshift-multus/multus-skh5n" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.830472 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.830478 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/8c3718a4-f354-4284-92e0-fdfb45a692bd-mcd-auth-proxy-config\") pod \"machine-config-daemon-x65wv\" (UID: \"8c3718a4-f354-4284-92e0-fdfb45a692bd\") " pod="openshift-machine-config-operator/machine-config-daemon-x65wv" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.830534 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/8c3718a4-f354-4284-92e0-fdfb45a692bd-rootfs\") pod \"machine-config-daemon-x65wv\" (UID: \"8c3718a4-f354-4284-92e0-fdfb45a692bd\") " pod="openshift-machine-config-operator/machine-config-daemon-x65wv" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.830570 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/8c3718a4-f354-4284-92e0-fdfb45a692bd-proxy-tls\") pod \"machine-config-daemon-x65wv\" (UID: \"8c3718a4-f354-4284-92e0-fdfb45a692bd\") " pod="openshift-machine-config-operator/machine-config-daemon-x65wv" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.830597 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b619d233-b592-4b05-a0b7-dc094c88471e-etc-kubernetes\") pod \"multus-skh5n\" (UID: \"b619d233-b592-4b05-a0b7-dc094c88471e\") " pod="openshift-multus/multus-skh5n" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.830630 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/b619d233-b592-4b05-a0b7-dc094c88471e-host-var-lib-cni-bin\") pod \"multus-skh5n\" (UID: \"b619d233-b592-4b05-a0b7-dc094c88471e\") " pod="openshift-multus/multus-skh5n" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.830725 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/b619d233-b592-4b05-a0b7-dc094c88471e-host-var-lib-cni-bin\") pod \"multus-skh5n\" (UID: \"b619d233-b592-4b05-a0b7-dc094c88471e\") " pod="openshift-multus/multus-skh5n" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.830777 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/8c3718a4-f354-4284-92e0-fdfb45a692bd-rootfs\") pod \"machine-config-daemon-x65wv\" (UID: \"8c3718a4-f354-4284-92e0-fdfb45a692bd\") " pod="openshift-machine-config-operator/machine-config-daemon-x65wv" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.830777 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b619d233-b592-4b05-a0b7-dc094c88471e-etc-kubernetes\") pod \"multus-skh5n\" (UID: \"b619d233-b592-4b05-a0b7-dc094c88471e\") " pod="openshift-multus/multus-skh5n" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.834336 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/b619d233-b592-4b05-a0b7-dc094c88471e-cni-binary-copy\") pod \"multus-skh5n\" (UID: \"b619d233-b592-4b05-a0b7-dc094c88471e\") " pod="openshift-multus/multus-skh5n" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.834521 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/8c3718a4-f354-4284-92e0-fdfb45a692bd-mcd-auth-proxy-config\") pod \"machine-config-daemon-x65wv\" (UID: \"8c3718a4-f354-4284-92e0-fdfb45a692bd\") " pod="openshift-machine-config-operator/machine-config-daemon-x65wv" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.835139 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/b619d233-b592-4b05-a0b7-dc094c88471e-multus-daemon-config\") pod \"multus-skh5n\" (UID: \"b619d233-b592-4b05-a0b7-dc094c88471e\") " pod="openshift-multus/multus-skh5n" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.837604 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.838270 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/8c3718a4-f354-4284-92e0-fdfb45a692bd-proxy-tls\") pod \"machine-config-daemon-x65wv\" (UID: \"8c3718a4-f354-4284-92e0-fdfb45a692bd\") " pod="openshift-machine-config-operator/machine-config-daemon-x65wv" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.838405 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.851071 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.852256 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.853118 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.868822 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://205ec256e46a8e08b4b0b0805fb6c5f4966fc3ccef0738df4b97cb404e74da17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:33Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.872330 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sxjcz\" (UniqueName: \"kubernetes.io/projected/8c3718a4-f354-4284-92e0-fdfb45a692bd-kube-api-access-sxjcz\") pod \"machine-config-daemon-x65wv\" (UID: \"8c3718a4-f354-4284-92e0-fdfb45a692bd\") " pod="openshift-machine-config-operator/machine-config-daemon-x65wv" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.874746 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6tj78\" (UniqueName: \"kubernetes.io/projected/b619d233-b592-4b05-a0b7-dc094c88471e-kube-api-access-6tj78\") pod \"multus-skh5n\" (UID: \"b619d233-b592-4b05-a0b7-dc094c88471e\") " pod="openshift-multus/multus-skh5n" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.881199 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-84lgq" event={"ID":"6763d962-0892-4acc-bab9-7a4e93e39a85","Type":"ContainerStarted","Data":"6f0338d5f1927deb680d619f4137312544a1682aac3ad3dfa917366b5cc5c35c"} Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.911921 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:33Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.932317 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c3718a4-f354-4284-92e0-fdfb45a692bd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxjcz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxjcz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-x65wv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:33Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.954593 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:33Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.976849 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:33Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.991027 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.993706 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2blzw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca65468e-30c6-4666-962e-cc0de05e67fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c25z4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:33Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2blzw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:33Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:33 crc kubenswrapper[4754]: I0126 16:47:33.996929 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-skh5n" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.047838 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-skh5n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b619d233-b592-4b05-a0b7-dc094c88471e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-skh5n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:34Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.075733 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:34Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.088189 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-bwpd6"] Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.089139 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-bwpd6" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.091449 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.093959 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-jsbxt"] Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.094390 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.095041 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-jsbxt" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.096891 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.097441 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.098571 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.098587 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.098933 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.098959 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.098935 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.099911 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89251fa9c64bc25d6d6f355400d7644c099f22aed1cae812fc991a03595f1e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e90508d8a53be9146f9b388750730fdae9ee200306caf5ad393a356390a69ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:34Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.111338 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-84lgq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6763d962-0892-4acc-bab9-7a4e93e39a85\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8zlw6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:33Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-84lgq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:34Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.127863 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b89aa081-0a2e-4757-9210-e2a6c228bee9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9c5f40188fda14b14128514fe84bddfe8dcede3d6c560283ccb2ac0c66895fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f45445cbf1d180d6d4dd9588e85bf678f043bfce8cff0812b78ca3510904f9a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://085fa7b8d4fef8b2722fac8c346107699b5a6bd33fbfc74ed35831cb8fdf9dad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c40e018a0979035827963efb9efc48a878a55b9df68ec1a9242d7cff64fe480\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8356f76fbc3a4e636c12d1da5455d3f6767d9b83ce2f2925a11435d7d7261e9f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0126 16:47:26.205208 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0126 16:47:26.206538 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3645934156/tls.crt::/tmp/serving-cert-3645934156/tls.key\\\\\\\"\\\\nI0126 16:47:31.504518 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0126 16:47:31.509366 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0126 16:47:31.509400 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0126 16:47:31.509428 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0126 16:47:31.509436 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0126 16:47:31.515912 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0126 16:47:31.516036 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 16:47:31.516068 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 16:47:31.516096 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0126 16:47:31.516121 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0126 16:47:31.516145 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0126 16:47:31.516170 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0126 16:47:31.516412 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0126 16:47:31.519294 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27b74e2e65a547e0e28fddd4f15feacb17f4d57596eeb8da82df9d39243c85e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22ba7649a1308815bfad52016000987a7a9a5bc214d7ff61e137ba099b20fc6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://22ba7649a1308815bfad52016000987a7a9a5bc214d7ff61e137ba099b20fc6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:13Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:34Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.134437 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/689569c4-93ee-4b82-93fc-9af0f97e6dc3-host-cni-bin\") pod \"ovnkube-node-jsbxt\" (UID: \"689569c4-93ee-4b82-93fc-9af0f97e6dc3\") " pod="openshift-ovn-kubernetes/ovnkube-node-jsbxt" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.134597 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/689569c4-93ee-4b82-93fc-9af0f97e6dc3-run-systemd\") pod \"ovnkube-node-jsbxt\" (UID: \"689569c4-93ee-4b82-93fc-9af0f97e6dc3\") " pod="openshift-ovn-kubernetes/ovnkube-node-jsbxt" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.134731 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/351cc6c0-b8e5-4589-ae63-f98e8ab1eca3-os-release\") pod \"multus-additional-cni-plugins-bwpd6\" (UID: \"351cc6c0-b8e5-4589-ae63-f98e8ab1eca3\") " pod="openshift-multus/multus-additional-cni-plugins-bwpd6" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.134893 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/689569c4-93ee-4b82-93fc-9af0f97e6dc3-systemd-units\") pod \"ovnkube-node-jsbxt\" (UID: \"689569c4-93ee-4b82-93fc-9af0f97e6dc3\") " pod="openshift-ovn-kubernetes/ovnkube-node-jsbxt" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.135044 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/689569c4-93ee-4b82-93fc-9af0f97e6dc3-env-overrides\") pod \"ovnkube-node-jsbxt\" (UID: \"689569c4-93ee-4b82-93fc-9af0f97e6dc3\") " pod="openshift-ovn-kubernetes/ovnkube-node-jsbxt" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.135153 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/689569c4-93ee-4b82-93fc-9af0f97e6dc3-node-log\") pod \"ovnkube-node-jsbxt\" (UID: \"689569c4-93ee-4b82-93fc-9af0f97e6dc3\") " pod="openshift-ovn-kubernetes/ovnkube-node-jsbxt" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.135269 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wm9cg\" (UniqueName: \"kubernetes.io/projected/351cc6c0-b8e5-4589-ae63-f98e8ab1eca3-kube-api-access-wm9cg\") pod \"multus-additional-cni-plugins-bwpd6\" (UID: \"351cc6c0-b8e5-4589-ae63-f98e8ab1eca3\") " pod="openshift-multus/multus-additional-cni-plugins-bwpd6" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.135437 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/689569c4-93ee-4b82-93fc-9af0f97e6dc3-host-slash\") pod \"ovnkube-node-jsbxt\" (UID: \"689569c4-93ee-4b82-93fc-9af0f97e6dc3\") " pod="openshift-ovn-kubernetes/ovnkube-node-jsbxt" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.135498 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/689569c4-93ee-4b82-93fc-9af0f97e6dc3-log-socket\") pod \"ovnkube-node-jsbxt\" (UID: \"689569c4-93ee-4b82-93fc-9af0f97e6dc3\") " pod="openshift-ovn-kubernetes/ovnkube-node-jsbxt" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.135539 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/689569c4-93ee-4b82-93fc-9af0f97e6dc3-host-run-netns\") pod \"ovnkube-node-jsbxt\" (UID: \"689569c4-93ee-4b82-93fc-9af0f97e6dc3\") " pod="openshift-ovn-kubernetes/ovnkube-node-jsbxt" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.135601 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/689569c4-93ee-4b82-93fc-9af0f97e6dc3-etc-openvswitch\") pod \"ovnkube-node-jsbxt\" (UID: \"689569c4-93ee-4b82-93fc-9af0f97e6dc3\") " pod="openshift-ovn-kubernetes/ovnkube-node-jsbxt" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.135626 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/689569c4-93ee-4b82-93fc-9af0f97e6dc3-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-jsbxt\" (UID: \"689569c4-93ee-4b82-93fc-9af0f97e6dc3\") " pod="openshift-ovn-kubernetes/ovnkube-node-jsbxt" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.135656 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cpxwt\" (UniqueName: \"kubernetes.io/projected/689569c4-93ee-4b82-93fc-9af0f97e6dc3-kube-api-access-cpxwt\") pod \"ovnkube-node-jsbxt\" (UID: \"689569c4-93ee-4b82-93fc-9af0f97e6dc3\") " pod="openshift-ovn-kubernetes/ovnkube-node-jsbxt" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.135838 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/351cc6c0-b8e5-4589-ae63-f98e8ab1eca3-cni-binary-copy\") pod \"multus-additional-cni-plugins-bwpd6\" (UID: \"351cc6c0-b8e5-4589-ae63-f98e8ab1eca3\") " pod="openshift-multus/multus-additional-cni-plugins-bwpd6" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.136183 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/689569c4-93ee-4b82-93fc-9af0f97e6dc3-host-cni-netd\") pod \"ovnkube-node-jsbxt\" (UID: \"689569c4-93ee-4b82-93fc-9af0f97e6dc3\") " pod="openshift-ovn-kubernetes/ovnkube-node-jsbxt" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.136232 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/351cc6c0-b8e5-4589-ae63-f98e8ab1eca3-cnibin\") pod \"multus-additional-cni-plugins-bwpd6\" (UID: \"351cc6c0-b8e5-4589-ae63-f98e8ab1eca3\") " pod="openshift-multus/multus-additional-cni-plugins-bwpd6" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.136433 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/351cc6c0-b8e5-4589-ae63-f98e8ab1eca3-tuning-conf-dir\") pod \"multus-additional-cni-plugins-bwpd6\" (UID: \"351cc6c0-b8e5-4589-ae63-f98e8ab1eca3\") " pod="openshift-multus/multus-additional-cni-plugins-bwpd6" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.136543 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/689569c4-93ee-4b82-93fc-9af0f97e6dc3-ovnkube-config\") pod \"ovnkube-node-jsbxt\" (UID: \"689569c4-93ee-4b82-93fc-9af0f97e6dc3\") " pod="openshift-ovn-kubernetes/ovnkube-node-jsbxt" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.136602 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/689569c4-93ee-4b82-93fc-9af0f97e6dc3-run-ovn\") pod \"ovnkube-node-jsbxt\" (UID: \"689569c4-93ee-4b82-93fc-9af0f97e6dc3\") " pod="openshift-ovn-kubernetes/ovnkube-node-jsbxt" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.136640 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/689569c4-93ee-4b82-93fc-9af0f97e6dc3-ovn-node-metrics-cert\") pod \"ovnkube-node-jsbxt\" (UID: \"689569c4-93ee-4b82-93fc-9af0f97e6dc3\") " pod="openshift-ovn-kubernetes/ovnkube-node-jsbxt" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.136723 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/689569c4-93ee-4b82-93fc-9af0f97e6dc3-ovnkube-script-lib\") pod \"ovnkube-node-jsbxt\" (UID: \"689569c4-93ee-4b82-93fc-9af0f97e6dc3\") " pod="openshift-ovn-kubernetes/ovnkube-node-jsbxt" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.136768 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/351cc6c0-b8e5-4589-ae63-f98e8ab1eca3-system-cni-dir\") pod \"multus-additional-cni-plugins-bwpd6\" (UID: \"351cc6c0-b8e5-4589-ae63-f98e8ab1eca3\") " pod="openshift-multus/multus-additional-cni-plugins-bwpd6" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.136840 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/689569c4-93ee-4b82-93fc-9af0f97e6dc3-run-openvswitch\") pod \"ovnkube-node-jsbxt\" (UID: \"689569c4-93ee-4b82-93fc-9af0f97e6dc3\") " pod="openshift-ovn-kubernetes/ovnkube-node-jsbxt" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.137092 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/689569c4-93ee-4b82-93fc-9af0f97e6dc3-host-kubelet\") pod \"ovnkube-node-jsbxt\" (UID: \"689569c4-93ee-4b82-93fc-9af0f97e6dc3\") " pod="openshift-ovn-kubernetes/ovnkube-node-jsbxt" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.137131 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/689569c4-93ee-4b82-93fc-9af0f97e6dc3-host-run-ovn-kubernetes\") pod \"ovnkube-node-jsbxt\" (UID: \"689569c4-93ee-4b82-93fc-9af0f97e6dc3\") " pod="openshift-ovn-kubernetes/ovnkube-node-jsbxt" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.137204 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/351cc6c0-b8e5-4589-ae63-f98e8ab1eca3-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-bwpd6\" (UID: \"351cc6c0-b8e5-4589-ae63-f98e8ab1eca3\") " pod="openshift-multus/multus-additional-cni-plugins-bwpd6" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.137242 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/689569c4-93ee-4b82-93fc-9af0f97e6dc3-var-lib-openvswitch\") pod \"ovnkube-node-jsbxt\" (UID: \"689569c4-93ee-4b82-93fc-9af0f97e6dc3\") " pod="openshift-ovn-kubernetes/ovnkube-node-jsbxt" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.151226 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://205ec256e46a8e08b4b0b0805fb6c5f4966fc3ccef0738df4b97cb404e74da17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:34Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.156831 4754 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2027-01-26 16:42:33 +0000 UTC, rotation deadline is 2026-11-25 06:19:10.960064141 +0000 UTC Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.156932 4754 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 7261h31m36.803135157s for next certificate rotation Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.169465 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:34Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.182117 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c3718a4-f354-4284-92e0-fdfb45a692bd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxjcz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxjcz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-x65wv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:34Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.197852 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:34Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.209147 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:34Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.220479 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2blzw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca65468e-30c6-4666-962e-cc0de05e67fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c25z4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:33Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2blzw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:34Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.237608 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bwpd6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"351cc6c0-b8e5-4589-ae63-f98e8ab1eca3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bwpd6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:34Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.238072 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/689569c4-93ee-4b82-93fc-9af0f97e6dc3-host-cni-bin\") pod \"ovnkube-node-jsbxt\" (UID: \"689569c4-93ee-4b82-93fc-9af0f97e6dc3\") " pod="openshift-ovn-kubernetes/ovnkube-node-jsbxt" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.238141 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/689569c4-93ee-4b82-93fc-9af0f97e6dc3-systemd-units\") pod \"ovnkube-node-jsbxt\" (UID: \"689569c4-93ee-4b82-93fc-9af0f97e6dc3\") " pod="openshift-ovn-kubernetes/ovnkube-node-jsbxt" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.238169 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/689569c4-93ee-4b82-93fc-9af0f97e6dc3-run-systemd\") pod \"ovnkube-node-jsbxt\" (UID: \"689569c4-93ee-4b82-93fc-9af0f97e6dc3\") " pod="openshift-ovn-kubernetes/ovnkube-node-jsbxt" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.238194 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/689569c4-93ee-4b82-93fc-9af0f97e6dc3-systemd-units\") pod \"ovnkube-node-jsbxt\" (UID: \"689569c4-93ee-4b82-93fc-9af0f97e6dc3\") " pod="openshift-ovn-kubernetes/ovnkube-node-jsbxt" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.238197 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/351cc6c0-b8e5-4589-ae63-f98e8ab1eca3-os-release\") pod \"multus-additional-cni-plugins-bwpd6\" (UID: \"351cc6c0-b8e5-4589-ae63-f98e8ab1eca3\") " pod="openshift-multus/multus-additional-cni-plugins-bwpd6" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.238165 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/689569c4-93ee-4b82-93fc-9af0f97e6dc3-host-cni-bin\") pod \"ovnkube-node-jsbxt\" (UID: \"689569c4-93ee-4b82-93fc-9af0f97e6dc3\") " pod="openshift-ovn-kubernetes/ovnkube-node-jsbxt" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.238322 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/689569c4-93ee-4b82-93fc-9af0f97e6dc3-env-overrides\") pod \"ovnkube-node-jsbxt\" (UID: \"689569c4-93ee-4b82-93fc-9af0f97e6dc3\") " pod="openshift-ovn-kubernetes/ovnkube-node-jsbxt" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.238325 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/689569c4-93ee-4b82-93fc-9af0f97e6dc3-run-systemd\") pod \"ovnkube-node-jsbxt\" (UID: \"689569c4-93ee-4b82-93fc-9af0f97e6dc3\") " pod="openshift-ovn-kubernetes/ovnkube-node-jsbxt" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.238376 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/689569c4-93ee-4b82-93fc-9af0f97e6dc3-node-log\") pod \"ovnkube-node-jsbxt\" (UID: \"689569c4-93ee-4b82-93fc-9af0f97e6dc3\") " pod="openshift-ovn-kubernetes/ovnkube-node-jsbxt" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.238357 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/689569c4-93ee-4b82-93fc-9af0f97e6dc3-node-log\") pod \"ovnkube-node-jsbxt\" (UID: \"689569c4-93ee-4b82-93fc-9af0f97e6dc3\") " pod="openshift-ovn-kubernetes/ovnkube-node-jsbxt" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.238444 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wm9cg\" (UniqueName: \"kubernetes.io/projected/351cc6c0-b8e5-4589-ae63-f98e8ab1eca3-kube-api-access-wm9cg\") pod \"multus-additional-cni-plugins-bwpd6\" (UID: \"351cc6c0-b8e5-4589-ae63-f98e8ab1eca3\") " pod="openshift-multus/multus-additional-cni-plugins-bwpd6" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.238467 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/351cc6c0-b8e5-4589-ae63-f98e8ab1eca3-os-release\") pod \"multus-additional-cni-plugins-bwpd6\" (UID: \"351cc6c0-b8e5-4589-ae63-f98e8ab1eca3\") " pod="openshift-multus/multus-additional-cni-plugins-bwpd6" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.238492 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/689569c4-93ee-4b82-93fc-9af0f97e6dc3-log-socket\") pod \"ovnkube-node-jsbxt\" (UID: \"689569c4-93ee-4b82-93fc-9af0f97e6dc3\") " pod="openshift-ovn-kubernetes/ovnkube-node-jsbxt" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.238529 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/689569c4-93ee-4b82-93fc-9af0f97e6dc3-log-socket\") pod \"ovnkube-node-jsbxt\" (UID: \"689569c4-93ee-4b82-93fc-9af0f97e6dc3\") " pod="openshift-ovn-kubernetes/ovnkube-node-jsbxt" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.238536 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/689569c4-93ee-4b82-93fc-9af0f97e6dc3-host-slash\") pod \"ovnkube-node-jsbxt\" (UID: \"689569c4-93ee-4b82-93fc-9af0f97e6dc3\") " pod="openshift-ovn-kubernetes/ovnkube-node-jsbxt" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.238568 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/689569c4-93ee-4b82-93fc-9af0f97e6dc3-host-run-netns\") pod \"ovnkube-node-jsbxt\" (UID: \"689569c4-93ee-4b82-93fc-9af0f97e6dc3\") " pod="openshift-ovn-kubernetes/ovnkube-node-jsbxt" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.238589 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/689569c4-93ee-4b82-93fc-9af0f97e6dc3-host-slash\") pod \"ovnkube-node-jsbxt\" (UID: \"689569c4-93ee-4b82-93fc-9af0f97e6dc3\") " pod="openshift-ovn-kubernetes/ovnkube-node-jsbxt" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.238598 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/689569c4-93ee-4b82-93fc-9af0f97e6dc3-etc-openvswitch\") pod \"ovnkube-node-jsbxt\" (UID: \"689569c4-93ee-4b82-93fc-9af0f97e6dc3\") " pod="openshift-ovn-kubernetes/ovnkube-node-jsbxt" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.238627 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/689569c4-93ee-4b82-93fc-9af0f97e6dc3-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-jsbxt\" (UID: \"689569c4-93ee-4b82-93fc-9af0f97e6dc3\") " pod="openshift-ovn-kubernetes/ovnkube-node-jsbxt" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.238657 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/689569c4-93ee-4b82-93fc-9af0f97e6dc3-etc-openvswitch\") pod \"ovnkube-node-jsbxt\" (UID: \"689569c4-93ee-4b82-93fc-9af0f97e6dc3\") " pod="openshift-ovn-kubernetes/ovnkube-node-jsbxt" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.238724 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/689569c4-93ee-4b82-93fc-9af0f97e6dc3-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-jsbxt\" (UID: \"689569c4-93ee-4b82-93fc-9af0f97e6dc3\") " pod="openshift-ovn-kubernetes/ovnkube-node-jsbxt" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.238685 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cpxwt\" (UniqueName: \"kubernetes.io/projected/689569c4-93ee-4b82-93fc-9af0f97e6dc3-kube-api-access-cpxwt\") pod \"ovnkube-node-jsbxt\" (UID: \"689569c4-93ee-4b82-93fc-9af0f97e6dc3\") " pod="openshift-ovn-kubernetes/ovnkube-node-jsbxt" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.238629 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/689569c4-93ee-4b82-93fc-9af0f97e6dc3-host-run-netns\") pod \"ovnkube-node-jsbxt\" (UID: \"689569c4-93ee-4b82-93fc-9af0f97e6dc3\") " pod="openshift-ovn-kubernetes/ovnkube-node-jsbxt" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.238775 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/689569c4-93ee-4b82-93fc-9af0f97e6dc3-host-cni-netd\") pod \"ovnkube-node-jsbxt\" (UID: \"689569c4-93ee-4b82-93fc-9af0f97e6dc3\") " pod="openshift-ovn-kubernetes/ovnkube-node-jsbxt" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.238799 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/689569c4-93ee-4b82-93fc-9af0f97e6dc3-host-cni-netd\") pod \"ovnkube-node-jsbxt\" (UID: \"689569c4-93ee-4b82-93fc-9af0f97e6dc3\") " pod="openshift-ovn-kubernetes/ovnkube-node-jsbxt" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.238879 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/351cc6c0-b8e5-4589-ae63-f98e8ab1eca3-cnibin\") pod \"multus-additional-cni-plugins-bwpd6\" (UID: \"351cc6c0-b8e5-4589-ae63-f98e8ab1eca3\") " pod="openshift-multus/multus-additional-cni-plugins-bwpd6" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.238920 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/351cc6c0-b8e5-4589-ae63-f98e8ab1eca3-cni-binary-copy\") pod \"multus-additional-cni-plugins-bwpd6\" (UID: \"351cc6c0-b8e5-4589-ae63-f98e8ab1eca3\") " pod="openshift-multus/multus-additional-cni-plugins-bwpd6" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.238974 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/351cc6c0-b8e5-4589-ae63-f98e8ab1eca3-tuning-conf-dir\") pod \"multus-additional-cni-plugins-bwpd6\" (UID: \"351cc6c0-b8e5-4589-ae63-f98e8ab1eca3\") " pod="openshift-multus/multus-additional-cni-plugins-bwpd6" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.239008 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/689569c4-93ee-4b82-93fc-9af0f97e6dc3-ovnkube-config\") pod \"ovnkube-node-jsbxt\" (UID: \"689569c4-93ee-4b82-93fc-9af0f97e6dc3\") " pod="openshift-ovn-kubernetes/ovnkube-node-jsbxt" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.239026 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/689569c4-93ee-4b82-93fc-9af0f97e6dc3-env-overrides\") pod \"ovnkube-node-jsbxt\" (UID: \"689569c4-93ee-4b82-93fc-9af0f97e6dc3\") " pod="openshift-ovn-kubernetes/ovnkube-node-jsbxt" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.239044 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/689569c4-93ee-4b82-93fc-9af0f97e6dc3-run-ovn\") pod \"ovnkube-node-jsbxt\" (UID: \"689569c4-93ee-4b82-93fc-9af0f97e6dc3\") " pod="openshift-ovn-kubernetes/ovnkube-node-jsbxt" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.239076 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/689569c4-93ee-4b82-93fc-9af0f97e6dc3-ovn-node-metrics-cert\") pod \"ovnkube-node-jsbxt\" (UID: \"689569c4-93ee-4b82-93fc-9af0f97e6dc3\") " pod="openshift-ovn-kubernetes/ovnkube-node-jsbxt" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.239125 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/689569c4-93ee-4b82-93fc-9af0f97e6dc3-run-openvswitch\") pod \"ovnkube-node-jsbxt\" (UID: \"689569c4-93ee-4b82-93fc-9af0f97e6dc3\") " pod="openshift-ovn-kubernetes/ovnkube-node-jsbxt" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.239149 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/689569c4-93ee-4b82-93fc-9af0f97e6dc3-ovnkube-script-lib\") pod \"ovnkube-node-jsbxt\" (UID: \"689569c4-93ee-4b82-93fc-9af0f97e6dc3\") " pod="openshift-ovn-kubernetes/ovnkube-node-jsbxt" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.239171 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/351cc6c0-b8e5-4589-ae63-f98e8ab1eca3-system-cni-dir\") pod \"multus-additional-cni-plugins-bwpd6\" (UID: \"351cc6c0-b8e5-4589-ae63-f98e8ab1eca3\") " pod="openshift-multus/multus-additional-cni-plugins-bwpd6" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.239191 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/351cc6c0-b8e5-4589-ae63-f98e8ab1eca3-tuning-conf-dir\") pod \"multus-additional-cni-plugins-bwpd6\" (UID: \"351cc6c0-b8e5-4589-ae63-f98e8ab1eca3\") " pod="openshift-multus/multus-additional-cni-plugins-bwpd6" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.239078 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/351cc6c0-b8e5-4589-ae63-f98e8ab1eca3-cnibin\") pod \"multus-additional-cni-plugins-bwpd6\" (UID: \"351cc6c0-b8e5-4589-ae63-f98e8ab1eca3\") " pod="openshift-multus/multus-additional-cni-plugins-bwpd6" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.239215 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/689569c4-93ee-4b82-93fc-9af0f97e6dc3-host-run-ovn-kubernetes\") pod \"ovnkube-node-jsbxt\" (UID: \"689569c4-93ee-4b82-93fc-9af0f97e6dc3\") " pod="openshift-ovn-kubernetes/ovnkube-node-jsbxt" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.239244 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/689569c4-93ee-4b82-93fc-9af0f97e6dc3-host-run-ovn-kubernetes\") pod \"ovnkube-node-jsbxt\" (UID: \"689569c4-93ee-4b82-93fc-9af0f97e6dc3\") " pod="openshift-ovn-kubernetes/ovnkube-node-jsbxt" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.239257 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/351cc6c0-b8e5-4589-ae63-f98e8ab1eca3-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-bwpd6\" (UID: \"351cc6c0-b8e5-4589-ae63-f98e8ab1eca3\") " pod="openshift-multus/multus-additional-cni-plugins-bwpd6" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.239290 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/689569c4-93ee-4b82-93fc-9af0f97e6dc3-host-kubelet\") pod \"ovnkube-node-jsbxt\" (UID: \"689569c4-93ee-4b82-93fc-9af0f97e6dc3\") " pod="openshift-ovn-kubernetes/ovnkube-node-jsbxt" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.239321 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/689569c4-93ee-4b82-93fc-9af0f97e6dc3-var-lib-openvswitch\") pod \"ovnkube-node-jsbxt\" (UID: \"689569c4-93ee-4b82-93fc-9af0f97e6dc3\") " pod="openshift-ovn-kubernetes/ovnkube-node-jsbxt" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.239396 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/689569c4-93ee-4b82-93fc-9af0f97e6dc3-var-lib-openvswitch\") pod \"ovnkube-node-jsbxt\" (UID: \"689569c4-93ee-4b82-93fc-9af0f97e6dc3\") " pod="openshift-ovn-kubernetes/ovnkube-node-jsbxt" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.239451 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/689569c4-93ee-4b82-93fc-9af0f97e6dc3-run-ovn\") pod \"ovnkube-node-jsbxt\" (UID: \"689569c4-93ee-4b82-93fc-9af0f97e6dc3\") " pod="openshift-ovn-kubernetes/ovnkube-node-jsbxt" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.239466 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/351cc6c0-b8e5-4589-ae63-f98e8ab1eca3-system-cni-dir\") pod \"multus-additional-cni-plugins-bwpd6\" (UID: \"351cc6c0-b8e5-4589-ae63-f98e8ab1eca3\") " pod="openshift-multus/multus-additional-cni-plugins-bwpd6" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.239523 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/689569c4-93ee-4b82-93fc-9af0f97e6dc3-host-kubelet\") pod \"ovnkube-node-jsbxt\" (UID: \"689569c4-93ee-4b82-93fc-9af0f97e6dc3\") " pod="openshift-ovn-kubernetes/ovnkube-node-jsbxt" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.239618 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/689569c4-93ee-4b82-93fc-9af0f97e6dc3-ovnkube-config\") pod \"ovnkube-node-jsbxt\" (UID: \"689569c4-93ee-4b82-93fc-9af0f97e6dc3\") " pod="openshift-ovn-kubernetes/ovnkube-node-jsbxt" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.239291 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/689569c4-93ee-4b82-93fc-9af0f97e6dc3-run-openvswitch\") pod \"ovnkube-node-jsbxt\" (UID: \"689569c4-93ee-4b82-93fc-9af0f97e6dc3\") " pod="openshift-ovn-kubernetes/ovnkube-node-jsbxt" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.240004 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/689569c4-93ee-4b82-93fc-9af0f97e6dc3-ovnkube-script-lib\") pod \"ovnkube-node-jsbxt\" (UID: \"689569c4-93ee-4b82-93fc-9af0f97e6dc3\") " pod="openshift-ovn-kubernetes/ovnkube-node-jsbxt" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.240064 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/351cc6c0-b8e5-4589-ae63-f98e8ab1eca3-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-bwpd6\" (UID: \"351cc6c0-b8e5-4589-ae63-f98e8ab1eca3\") " pod="openshift-multus/multus-additional-cni-plugins-bwpd6" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.240295 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/351cc6c0-b8e5-4589-ae63-f98e8ab1eca3-cni-binary-copy\") pod \"multus-additional-cni-plugins-bwpd6\" (UID: \"351cc6c0-b8e5-4589-ae63-f98e8ab1eca3\") " pod="openshift-multus/multus-additional-cni-plugins-bwpd6" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.245303 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/689569c4-93ee-4b82-93fc-9af0f97e6dc3-ovn-node-metrics-cert\") pod \"ovnkube-node-jsbxt\" (UID: \"689569c4-93ee-4b82-93fc-9af0f97e6dc3\") " pod="openshift-ovn-kubernetes/ovnkube-node-jsbxt" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.257034 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wm9cg\" (UniqueName: \"kubernetes.io/projected/351cc6c0-b8e5-4589-ae63-f98e8ab1eca3-kube-api-access-wm9cg\") pod \"multus-additional-cni-plugins-bwpd6\" (UID: \"351cc6c0-b8e5-4589-ae63-f98e8ab1eca3\") " pod="openshift-multus/multus-additional-cni-plugins-bwpd6" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.257331 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cpxwt\" (UniqueName: \"kubernetes.io/projected/689569c4-93ee-4b82-93fc-9af0f97e6dc3-kube-api-access-cpxwt\") pod \"ovnkube-node-jsbxt\" (UID: \"689569c4-93ee-4b82-93fc-9af0f97e6dc3\") " pod="openshift-ovn-kubernetes/ovnkube-node-jsbxt" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.269740 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jsbxt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"689569c4-93ee-4b82-93fc-9af0f97e6dc3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jsbxt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:34Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.284800 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:34Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.299246 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:34Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.318287 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2blzw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca65468e-30c6-4666-962e-cc0de05e67fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c25z4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:33Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2blzw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:34Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.337854 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.337861 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-84lgq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6763d962-0892-4acc-bab9-7a4e93e39a85\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8zlw6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:33Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-84lgq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:34Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.347422 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c25z4\" (UniqueName: \"kubernetes.io/projected/ca65468e-30c6-4666-962e-cc0de05e67fe-kube-api-access-c25z4\") pod \"node-ca-2blzw\" (UID: \"ca65468e-30c6-4666-962e-cc0de05e67fe\") " pod="openshift-image-registry/node-ca-2blzw" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.351902 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-skh5n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b619d233-b592-4b05-a0b7-dc094c88471e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-skh5n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:34Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.373547 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:34Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.391352 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89251fa9c64bc25d6d6f355400d7644c099f22aed1cae812fc991a03595f1e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e90508d8a53be9146f9b388750730fdae9ee200306caf5ad393a356390a69ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:34Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.407162 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-bwpd6" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.408461 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b89aa081-0a2e-4757-9210-e2a6c228bee9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9c5f40188fda14b14128514fe84bddfe8dcede3d6c560283ccb2ac0c66895fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f45445cbf1d180d6d4dd9588e85bf678f043bfce8cff0812b78ca3510904f9a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://085fa7b8d4fef8b2722fac8c346107699b5a6bd33fbfc74ed35831cb8fdf9dad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c40e018a0979035827963efb9efc48a878a55b9df68ec1a9242d7cff64fe480\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8356f76fbc3a4e636c12d1da5455d3f6767d9b83ce2f2925a11435d7d7261e9f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0126 16:47:26.205208 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0126 16:47:26.206538 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3645934156/tls.crt::/tmp/serving-cert-3645934156/tls.key\\\\\\\"\\\\nI0126 16:47:31.504518 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0126 16:47:31.509366 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0126 16:47:31.509400 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0126 16:47:31.509428 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0126 16:47:31.509436 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0126 16:47:31.515912 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0126 16:47:31.516036 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 16:47:31.516068 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 16:47:31.516096 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0126 16:47:31.516121 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0126 16:47:31.516145 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0126 16:47:31.516170 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0126 16:47:31.516412 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0126 16:47:31.519294 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27b74e2e65a547e0e28fddd4f15feacb17f4d57596eeb8da82df9d39243c85e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22ba7649a1308815bfad52016000987a7a9a5bc214d7ff61e137ba099b20fc6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://22ba7649a1308815bfad52016000987a7a9a5bc214d7ff61e137ba099b20fc6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:13Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:34Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:34 crc kubenswrapper[4754]: W0126 16:47:34.417334 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod351cc6c0_b8e5_4589_ae63_f98e8ab1eca3.slice/crio-986d52f9473e1e9d7637173b7887afd319cf5794a08ae5282904e99db430f4a4 WatchSource:0}: Error finding container 986d52f9473e1e9d7637173b7887afd319cf5794a08ae5282904e99db430f4a4: Status 404 returned error can't find the container with id 986d52f9473e1e9d7637173b7887afd319cf5794a08ae5282904e99db430f4a4 Jan 26 16:47:34 crc kubenswrapper[4754]: E0126 16:47:34.424970 4754 configmap.go:193] Couldn't get configMap openshift-image-registry/image-registry-certificates: failed to sync configmap cache: timed out waiting for the condition Jan 26 16:47:34 crc kubenswrapper[4754]: E0126 16:47:34.425090 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/ca65468e-30c6-4666-962e-cc0de05e67fe-serviceca podName:ca65468e-30c6-4666-962e-cc0de05e67fe nodeName:}" failed. No retries permitted until 2026-01-26 16:47:34.925062146 +0000 UTC m=+21.449242580 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "serviceca" (UniqueName: "kubernetes.io/configmap/ca65468e-30c6-4666-962e-cc0de05e67fe-serviceca") pod "node-ca-2blzw" (UID: "ca65468e-30c6-4666-962e-cc0de05e67fe") : failed to sync configmap cache: timed out waiting for the condition Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.426050 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://205ec256e46a8e08b4b0b0805fb6c5f4966fc3ccef0738df4b97cb404e74da17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:34Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.435234 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-jsbxt" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.447451 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:34Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:34 crc kubenswrapper[4754]: W0126 16:47:34.457493 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod689569c4_93ee_4b82_93fc_9af0f97e6dc3.slice/crio-25d5c4accfdda0e7dca5e9aac838cceced3c6028b8e097f54e22904a0c839824 WatchSource:0}: Error finding container 25d5c4accfdda0e7dca5e9aac838cceced3c6028b8e097f54e22904a0c839824: Status 404 returned error can't find the container with id 25d5c4accfdda0e7dca5e9aac838cceced3c6028b8e097f54e22904a0c839824 Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.487656 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c3718a4-f354-4284-92e0-fdfb45a692bd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxjcz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxjcz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-x65wv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:34Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.527963 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.530584 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:34Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.557554 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:34Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.563832 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.571753 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.575453 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2blzw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca65468e-30c6-4666-962e-cc0de05e67fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c25z4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:33Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2blzw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:34Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.595057 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bwpd6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"351cc6c0-b8e5-4589-ae63-f98e8ab1eca3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bwpd6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:34Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.614517 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jsbxt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"689569c4-93ee-4b82-93fc-9af0f97e6dc3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jsbxt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:34Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.630204 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:34Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.646610 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89251fa9c64bc25d6d6f355400d7644c099f22aed1cae812fc991a03595f1e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e90508d8a53be9146f9b388750730fdae9ee200306caf5ad393a356390a69ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:34Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.671555 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-84lgq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6763d962-0892-4acc-bab9-7a4e93e39a85\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8zlw6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:33Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-84lgq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:34Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.690916 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.693926 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-skh5n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b619d233-b592-4b05-a0b7-dc094c88471e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-skh5n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:34Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.699942 4754 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.706659 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.706765 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.706779 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.707034 4754 kubelet_node_status.go:76] "Attempting to register node" node="crc" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.720589 4754 kubelet_node_status.go:115] "Node was previously registered" node="crc" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.720996 4754 kubelet_node_status.go:79] "Successfully registered node" node="crc" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.721495 4754 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-23 04:33:44.20017301 +0000 UTC Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.722213 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.722239 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.722248 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.722265 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.722275 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:34Z","lastTransitionTime":"2026-01-26T16:47:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.729430 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b89aa081-0a2e-4757-9210-e2a6c228bee9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9c5f40188fda14b14128514fe84bddfe8dcede3d6c560283ccb2ac0c66895fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f45445cbf1d180d6d4dd9588e85bf678f043bfce8cff0812b78ca3510904f9a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://085fa7b8d4fef8b2722fac8c346107699b5a6bd33fbfc74ed35831cb8fdf9dad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c40e018a0979035827963efb9efc48a878a55b9df68ec1a9242d7cff64fe480\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8356f76fbc3a4e636c12d1da5455d3f6767d9b83ce2f2925a11435d7d7261e9f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0126 16:47:26.205208 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0126 16:47:26.206538 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3645934156/tls.crt::/tmp/serving-cert-3645934156/tls.key\\\\\\\"\\\\nI0126 16:47:31.504518 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0126 16:47:31.509366 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0126 16:47:31.509400 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0126 16:47:31.509428 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0126 16:47:31.509436 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0126 16:47:31.515912 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0126 16:47:31.516036 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 16:47:31.516068 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 16:47:31.516096 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0126 16:47:31.516121 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0126 16:47:31.516145 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0126 16:47:31.516170 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0126 16:47:31.516412 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0126 16:47:31.519294 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27b74e2e65a547e0e28fddd4f15feacb17f4d57596eeb8da82df9d39243c85e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22ba7649a1308815bfad52016000987a7a9a5bc214d7ff61e137ba099b20fc6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://22ba7649a1308815bfad52016000987a7a9a5bc214d7ff61e137ba099b20fc6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:13Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:34Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.734453 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.738157 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 26 16:47:34 crc kubenswrapper[4754]: E0126 16:47:34.742560 4754 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d8506764-e7fa-45cb-a13d-6f527164f548\\\",\\\"systemUUID\\\":\\\"bfbb6c33-ce46-4e81-a8a8-b44409b03821\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:34Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.745422 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://205ec256e46a8e08b4b0b0805fb6c5f4966fc3ccef0738df4b97cb404e74da17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:34Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.746960 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.747029 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.747046 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.747076 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.747094 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:34Z","lastTransitionTime":"2026-01-26T16:47:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.759633 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Jan 26 16:47:34 crc kubenswrapper[4754]: E0126 16:47:34.766115 4754 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d8506764-e7fa-45cb-a13d-6f527164f548\\\",\\\"systemUUID\\\":\\\"bfbb6c33-ce46-4e81-a8a8-b44409b03821\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:34Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.769643 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.769711 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.769724 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.769738 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.769749 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:34Z","lastTransitionTime":"2026-01-26T16:47:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:34 crc kubenswrapper[4754]: E0126 16:47:34.783564 4754 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d8506764-e7fa-45cb-a13d-6f527164f548\\\",\\\"systemUUID\\\":\\\"bfbb6c33-ce46-4e81-a8a8-b44409b03821\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:34Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.788303 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.788356 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.788370 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.788389 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.788401 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:34Z","lastTransitionTime":"2026-01-26T16:47:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.800635 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:34Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:34 crc kubenswrapper[4754]: E0126 16:47:34.804930 4754 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d8506764-e7fa-45cb-a13d-6f527164f548\\\",\\\"systemUUID\\\":\\\"bfbb6c33-ce46-4e81-a8a8-b44409b03821\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:34Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.809320 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.809376 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.809389 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.809407 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.809419 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:34Z","lastTransitionTime":"2026-01-26T16:47:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.809706 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Jan 26 16:47:34 crc kubenswrapper[4754]: E0126 16:47:34.825774 4754 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d8506764-e7fa-45cb-a13d-6f527164f548\\\",\\\"systemUUID\\\":\\\"bfbb6c33-ce46-4e81-a8a8-b44409b03821\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:34Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:34 crc kubenswrapper[4754]: E0126 16:47:34.825893 4754 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.827815 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.827849 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.827872 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.827889 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.827901 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:34Z","lastTransitionTime":"2026-01-26T16:47:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.851251 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.881216 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c3718a4-f354-4284-92e0-fdfb45a692bd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxjcz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxjcz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-x65wv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:34Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.885134 4754 generic.go:334] "Generic (PLEG): container finished" podID="689569c4-93ee-4b82-93fc-9af0f97e6dc3" containerID="8d9a7bafabd44f8b88a6f62cbe45599af8563bc45d9d7ad3a7669f3d8d11eb9c" exitCode=0 Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.885194 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jsbxt" event={"ID":"689569c4-93ee-4b82-93fc-9af0f97e6dc3","Type":"ContainerDied","Data":"8d9a7bafabd44f8b88a6f62cbe45599af8563bc45d9d7ad3a7669f3d8d11eb9c"} Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.885223 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jsbxt" event={"ID":"689569c4-93ee-4b82-93fc-9af0f97e6dc3","Type":"ContainerStarted","Data":"25d5c4accfdda0e7dca5e9aac838cceced3c6028b8e097f54e22904a0c839824"} Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.887873 4754 generic.go:334] "Generic (PLEG): container finished" podID="351cc6c0-b8e5-4589-ae63-f98e8ab1eca3" containerID="d3ce0fe080a094316b9456057e347b90a5ff432daa12b9c56d7aea9bee1f72b6" exitCode=0 Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.887939 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-bwpd6" event={"ID":"351cc6c0-b8e5-4589-ae63-f98e8ab1eca3","Type":"ContainerDied","Data":"d3ce0fe080a094316b9456057e347b90a5ff432daa12b9c56d7aea9bee1f72b6"} Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.887958 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-bwpd6" event={"ID":"351cc6c0-b8e5-4589-ae63-f98e8ab1eca3","Type":"ContainerStarted","Data":"986d52f9473e1e9d7637173b7887afd319cf5794a08ae5282904e99db430f4a4"} Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.896637 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-skh5n" event={"ID":"b619d233-b592-4b05-a0b7-dc094c88471e","Type":"ContainerStarted","Data":"9231a4191565e2e0ef53d8a394bb34615c900b601a0dfec24c2d8e5fa7b0b659"} Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.896708 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-skh5n" event={"ID":"b619d233-b592-4b05-a0b7-dc094c88471e","Type":"ContainerStarted","Data":"11660eb0c0cd5acdaaf5463db90e03ef26ccc8369327bff7ed471ceb07c156cd"} Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.904488 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" event={"ID":"8c3718a4-f354-4284-92e0-fdfb45a692bd","Type":"ContainerStarted","Data":"b86f46f49163bca7eca9458279029af330dbd4890b03325e06fa3b1bdf030de6"} Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.904542 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" event={"ID":"8c3718a4-f354-4284-92e0-fdfb45a692bd","Type":"ContainerStarted","Data":"0fffc03b536bfee3c7bc1bddd500e834e34b391896477a1501e3e06201a19374"} Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.904556 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" event={"ID":"8c3718a4-f354-4284-92e0-fdfb45a692bd","Type":"ContainerStarted","Data":"23f6c2661fe11048a58b45534a6ba3a23b0243406c8425ff57463ec81a1cf0c5"} Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.909453 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"7e459fd8c1093d6f3d0a671751b792b26b7df94c21ced1be1a2250821bcc82d3"} Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.912327 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-84lgq" event={"ID":"6763d962-0892-4acc-bab9-7a4e93e39a85","Type":"ContainerStarted","Data":"953c1917fca896c3def3fc2c69f9e3dbdc9ca26fc2b3d40282be71f71759fd68"} Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.921611 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:34Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.935541 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.935580 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.935590 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.935604 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.935615 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:34Z","lastTransitionTime":"2026-01-26T16:47:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.947188 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/ca65468e-30c6-4666-962e-cc0de05e67fe-serviceca\") pod \"node-ca-2blzw\" (UID: \"ca65468e-30c6-4666-962e-cc0de05e67fe\") " pod="openshift-image-registry/node-ca-2blzw" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.951223 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/ca65468e-30c6-4666-962e-cc0de05e67fe-serviceca\") pod \"node-ca-2blzw\" (UID: \"ca65468e-30c6-4666-962e-cc0de05e67fe\") " pod="openshift-image-registry/node-ca-2blzw" Jan 26 16:47:34 crc kubenswrapper[4754]: I0126 16:47:34.970826 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89251fa9c64bc25d6d6f355400d7644c099f22aed1cae812fc991a03595f1e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e90508d8a53be9146f9b388750730fdae9ee200306caf5ad393a356390a69ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:34Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:35 crc kubenswrapper[4754]: I0126 16:47:35.002696 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-84lgq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6763d962-0892-4acc-bab9-7a4e93e39a85\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8zlw6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:33Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-84lgq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:34Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:35 crc kubenswrapper[4754]: I0126 16:47:35.039624 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:35 crc kubenswrapper[4754]: I0126 16:47:35.040237 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:35 crc kubenswrapper[4754]: I0126 16:47:35.040262 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:35 crc kubenswrapper[4754]: I0126 16:47:35.040288 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:35 crc kubenswrapper[4754]: I0126 16:47:35.040305 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:35Z","lastTransitionTime":"2026-01-26T16:47:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:35 crc kubenswrapper[4754]: I0126 16:47:35.053298 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-2blzw" Jan 26 16:47:35 crc kubenswrapper[4754]: I0126 16:47:35.064003 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-skh5n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b619d233-b592-4b05-a0b7-dc094c88471e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-skh5n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:35Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:35 crc kubenswrapper[4754]: I0126 16:47:35.105974 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"099af702-abf2-4933-b4c9-63ef2326e468\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23017e73f8450839b4c2c02fa61235ffe4604bdce11556def27f25bcadcf89dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a74507c941b9de4d0401c8f1d53bb79ed11359628d3405cdab7f5474a75c862e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4402182b3e2e35a7d5dd4c44ac53e152915b312b7ccb076c4f2b41b925a6e7f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f772a003566198519fe1849fd14f29d4346004b2e655810662507c95b4d5ffa1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:13Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:35Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:35 crc kubenswrapper[4754]: I0126 16:47:35.145734 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:35Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:35 crc kubenswrapper[4754]: I0126 16:47:35.146238 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:35 crc kubenswrapper[4754]: I0126 16:47:35.146261 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:35 crc kubenswrapper[4754]: I0126 16:47:35.146269 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:35 crc kubenswrapper[4754]: I0126 16:47:35.146283 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:35 crc kubenswrapper[4754]: I0126 16:47:35.146293 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:35Z","lastTransitionTime":"2026-01-26T16:47:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:35 crc kubenswrapper[4754]: I0126 16:47:35.164475 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c3718a4-f354-4284-92e0-fdfb45a692bd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxjcz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxjcz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-x65wv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:35Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:35 crc kubenswrapper[4754]: W0126 16:47:35.202300 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podca65468e_30c6_4666_962e_cc0de05e67fe.slice/crio-a9ec824f384dcb81e04d3c21ced811379da15347c09c11f8eb4891498b42c455 WatchSource:0}: Error finding container a9ec824f384dcb81e04d3c21ced811379da15347c09c11f8eb4891498b42c455: Status 404 returned error can't find the container with id a9ec824f384dcb81e04d3c21ced811379da15347c09c11f8eb4891498b42c455 Jan 26 16:47:35 crc kubenswrapper[4754]: I0126 16:47:35.209649 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b89aa081-0a2e-4757-9210-e2a6c228bee9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9c5f40188fda14b14128514fe84bddfe8dcede3d6c560283ccb2ac0c66895fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f45445cbf1d180d6d4dd9588e85bf678f043bfce8cff0812b78ca3510904f9a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://085fa7b8d4fef8b2722fac8c346107699b5a6bd33fbfc74ed35831cb8fdf9dad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c40e018a0979035827963efb9efc48a878a55b9df68ec1a9242d7cff64fe480\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8356f76fbc3a4e636c12d1da5455d3f6767d9b83ce2f2925a11435d7d7261e9f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0126 16:47:26.205208 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0126 16:47:26.206538 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3645934156/tls.crt::/tmp/serving-cert-3645934156/tls.key\\\\\\\"\\\\nI0126 16:47:31.504518 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0126 16:47:31.509366 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0126 16:47:31.509400 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0126 16:47:31.509428 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0126 16:47:31.509436 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0126 16:47:31.515912 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0126 16:47:31.516036 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 16:47:31.516068 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 16:47:31.516096 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0126 16:47:31.516121 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0126 16:47:31.516145 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0126 16:47:31.516170 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0126 16:47:31.516412 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0126 16:47:31.519294 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27b74e2e65a547e0e28fddd4f15feacb17f4d57596eeb8da82df9d39243c85e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22ba7649a1308815bfad52016000987a7a9a5bc214d7ff61e137ba099b20fc6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://22ba7649a1308815bfad52016000987a7a9a5bc214d7ff61e137ba099b20fc6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:13Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:35Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:35 crc kubenswrapper[4754]: I0126 16:47:35.242767 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://205ec256e46a8e08b4b0b0805fb6c5f4966fc3ccef0738df4b97cb404e74da17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:35Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:35 crc kubenswrapper[4754]: I0126 16:47:35.248889 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:35 crc kubenswrapper[4754]: I0126 16:47:35.248911 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:35 crc kubenswrapper[4754]: I0126 16:47:35.248920 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:35 crc kubenswrapper[4754]: I0126 16:47:35.248932 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:35 crc kubenswrapper[4754]: I0126 16:47:35.248941 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:35Z","lastTransitionTime":"2026-01-26T16:47:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:35 crc kubenswrapper[4754]: I0126 16:47:35.284046 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:35Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:35 crc kubenswrapper[4754]: I0126 16:47:35.329544 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2blzw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca65468e-30c6-4666-962e-cc0de05e67fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c25z4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:33Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2blzw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:35Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:35 crc kubenswrapper[4754]: I0126 16:47:35.352301 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:35 crc kubenswrapper[4754]: I0126 16:47:35.352329 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:35 crc kubenswrapper[4754]: I0126 16:47:35.352338 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:35 crc kubenswrapper[4754]: I0126 16:47:35.352351 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:35 crc kubenswrapper[4754]: I0126 16:47:35.352362 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:35Z","lastTransitionTime":"2026-01-26T16:47:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:35 crc kubenswrapper[4754]: I0126 16:47:35.364974 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bwpd6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"351cc6c0-b8e5-4589-ae63-f98e8ab1eca3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bwpd6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:35Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:35 crc kubenswrapper[4754]: I0126 16:47:35.408776 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jsbxt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"689569c4-93ee-4b82-93fc-9af0f97e6dc3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jsbxt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:35Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:35 crc kubenswrapper[4754]: I0126 16:47:35.445888 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:35Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:35 crc kubenswrapper[4754]: I0126 16:47:35.456280 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:35 crc kubenswrapper[4754]: I0126 16:47:35.456325 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:35 crc kubenswrapper[4754]: I0126 16:47:35.456338 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:35 crc kubenswrapper[4754]: I0126 16:47:35.456355 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:35 crc kubenswrapper[4754]: I0126 16:47:35.456366 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:35Z","lastTransitionTime":"2026-01-26T16:47:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:35 crc kubenswrapper[4754]: I0126 16:47:35.491145 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jsbxt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"689569c4-93ee-4b82-93fc-9af0f97e6dc3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d9a7bafabd44f8b88a6f62cbe45599af8563bc45d9d7ad3a7669f3d8d11eb9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d9a7bafabd44f8b88a6f62cbe45599af8563bc45d9d7ad3a7669f3d8d11eb9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jsbxt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:35Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:35 crc kubenswrapper[4754]: I0126 16:47:35.529365 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:35Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:35 crc kubenswrapper[4754]: I0126 16:47:35.553855 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 16:47:35 crc kubenswrapper[4754]: I0126 16:47:35.553983 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 16:47:35 crc kubenswrapper[4754]: I0126 16:47:35.554005 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 16:47:35 crc kubenswrapper[4754]: I0126 16:47:35.554031 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 16:47:35 crc kubenswrapper[4754]: I0126 16:47:35.554060 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 16:47:35 crc kubenswrapper[4754]: E0126 16:47:35.554168 4754 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 26 16:47:35 crc kubenswrapper[4754]: E0126 16:47:35.554229 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-26 16:47:39.554213233 +0000 UTC m=+26.078393667 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 26 16:47:35 crc kubenswrapper[4754]: E0126 16:47:35.554221 4754 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Jan 26 16:47:35 crc kubenswrapper[4754]: E0126 16:47:35.554258 4754 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 26 16:47:35 crc kubenswrapper[4754]: E0126 16:47:35.554292 4754 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 26 16:47:35 crc kubenswrapper[4754]: E0126 16:47:35.554331 4754 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 26 16:47:35 crc kubenswrapper[4754]: E0126 16:47:35.554343 4754 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 26 16:47:35 crc kubenswrapper[4754]: E0126 16:47:35.554364 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-26 16:47:39.554333247 +0000 UTC m=+26.078513681 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Jan 26 16:47:35 crc kubenswrapper[4754]: E0126 16:47:35.554410 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-01-26 16:47:39.554385938 +0000 UTC m=+26.078566552 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 26 16:47:35 crc kubenswrapper[4754]: E0126 16:47:35.554313 4754 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 26 16:47:35 crc kubenswrapper[4754]: E0126 16:47:35.554451 4754 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 26 16:47:35 crc kubenswrapper[4754]: E0126 16:47:35.554501 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 16:47:39.5544416 +0000 UTC m=+26.078622034 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 16:47:35 crc kubenswrapper[4754]: E0126 16:47:35.554534 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-01-26 16:47:39.554522262 +0000 UTC m=+26.078702906 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 26 16:47:35 crc kubenswrapper[4754]: I0126 16:47:35.559001 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:35 crc kubenswrapper[4754]: I0126 16:47:35.559038 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:35 crc kubenswrapper[4754]: I0126 16:47:35.559048 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:35 crc kubenswrapper[4754]: I0126 16:47:35.559064 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:35 crc kubenswrapper[4754]: I0126 16:47:35.559074 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:35Z","lastTransitionTime":"2026-01-26T16:47:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:35 crc kubenswrapper[4754]: I0126 16:47:35.560266 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e459fd8c1093d6f3d0a671751b792b26b7df94c21ced1be1a2250821bcc82d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:35Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:35 crc kubenswrapper[4754]: I0126 16:47:35.596092 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2blzw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca65468e-30c6-4666-962e-cc0de05e67fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c25z4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:33Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2blzw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:35Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:35 crc kubenswrapper[4754]: I0126 16:47:35.643094 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bwpd6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"351cc6c0-b8e5-4589-ae63-f98e8ab1eca3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3ce0fe080a094316b9456057e347b90a5ff432daa12b9c56d7aea9bee1f72b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3ce0fe080a094316b9456057e347b90a5ff432daa12b9c56d7aea9bee1f72b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bwpd6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:35Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:35 crc kubenswrapper[4754]: I0126 16:47:35.661733 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:35 crc kubenswrapper[4754]: I0126 16:47:35.661798 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:35 crc kubenswrapper[4754]: I0126 16:47:35.661812 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:35 crc kubenswrapper[4754]: I0126 16:47:35.661837 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:35 crc kubenswrapper[4754]: I0126 16:47:35.661852 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:35Z","lastTransitionTime":"2026-01-26T16:47:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:35 crc kubenswrapper[4754]: I0126 16:47:35.680355 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-skh5n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b619d233-b592-4b05-a0b7-dc094c88471e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9231a4191565e2e0ef53d8a394bb34615c900b601a0dfec24c2d8e5fa7b0b659\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-skh5n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:35Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:35 crc kubenswrapper[4754]: I0126 16:47:35.719903 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"099af702-abf2-4933-b4c9-63ef2326e468\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23017e73f8450839b4c2c02fa61235ffe4604bdce11556def27f25bcadcf89dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a74507c941b9de4d0401c8f1d53bb79ed11359628d3405cdab7f5474a75c862e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4402182b3e2e35a7d5dd4c44ac53e152915b312b7ccb076c4f2b41b925a6e7f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f772a003566198519fe1849fd14f29d4346004b2e655810662507c95b4d5ffa1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:13Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:35Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:35 crc kubenswrapper[4754]: I0126 16:47:35.721967 4754 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-25 03:17:51.445757947 +0000 UTC Jan 26 16:47:35 crc kubenswrapper[4754]: I0126 16:47:35.760712 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:35Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:35 crc kubenswrapper[4754]: I0126 16:47:35.764170 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:35 crc kubenswrapper[4754]: I0126 16:47:35.764206 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:35 crc kubenswrapper[4754]: I0126 16:47:35.764217 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:35 crc kubenswrapper[4754]: I0126 16:47:35.764233 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:35 crc kubenswrapper[4754]: I0126 16:47:35.764246 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:35Z","lastTransitionTime":"2026-01-26T16:47:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:35 crc kubenswrapper[4754]: I0126 16:47:35.766794 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 16:47:35 crc kubenswrapper[4754]: I0126 16:47:35.766794 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 16:47:35 crc kubenswrapper[4754]: I0126 16:47:35.766935 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 16:47:35 crc kubenswrapper[4754]: E0126 16:47:35.767162 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 16:47:35 crc kubenswrapper[4754]: E0126 16:47:35.767275 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 16:47:35 crc kubenswrapper[4754]: E0126 16:47:35.767454 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 16:47:35 crc kubenswrapper[4754]: I0126 16:47:35.800564 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89251fa9c64bc25d6d6f355400d7644c099f22aed1cae812fc991a03595f1e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e90508d8a53be9146f9b388750730fdae9ee200306caf5ad393a356390a69ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:35Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:35 crc kubenswrapper[4754]: I0126 16:47:35.837571 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-84lgq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6763d962-0892-4acc-bab9-7a4e93e39a85\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://953c1917fca896c3def3fc2c69f9e3dbdc9ca26fc2b3d40282be71f71759fd68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8zlw6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:33Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-84lgq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:35Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:35 crc kubenswrapper[4754]: I0126 16:47:35.868027 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:35 crc kubenswrapper[4754]: I0126 16:47:35.868093 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:35 crc kubenswrapper[4754]: I0126 16:47:35.868110 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:35 crc kubenswrapper[4754]: I0126 16:47:35.868144 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:35 crc kubenswrapper[4754]: I0126 16:47:35.868156 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:35Z","lastTransitionTime":"2026-01-26T16:47:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:35 crc kubenswrapper[4754]: I0126 16:47:35.881084 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b89aa081-0a2e-4757-9210-e2a6c228bee9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9c5f40188fda14b14128514fe84bddfe8dcede3d6c560283ccb2ac0c66895fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f45445cbf1d180d6d4dd9588e85bf678f043bfce8cff0812b78ca3510904f9a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://085fa7b8d4fef8b2722fac8c346107699b5a6bd33fbfc74ed35831cb8fdf9dad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c40e018a0979035827963efb9efc48a878a55b9df68ec1a9242d7cff64fe480\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8356f76fbc3a4e636c12d1da5455d3f6767d9b83ce2f2925a11435d7d7261e9f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0126 16:47:26.205208 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0126 16:47:26.206538 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3645934156/tls.crt::/tmp/serving-cert-3645934156/tls.key\\\\\\\"\\\\nI0126 16:47:31.504518 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0126 16:47:31.509366 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0126 16:47:31.509400 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0126 16:47:31.509428 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0126 16:47:31.509436 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0126 16:47:31.515912 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0126 16:47:31.516036 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 16:47:31.516068 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 16:47:31.516096 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0126 16:47:31.516121 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0126 16:47:31.516145 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0126 16:47:31.516170 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0126 16:47:31.516412 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0126 16:47:31.519294 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27b74e2e65a547e0e28fddd4f15feacb17f4d57596eeb8da82df9d39243c85e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22ba7649a1308815bfad52016000987a7a9a5bc214d7ff61e137ba099b20fc6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://22ba7649a1308815bfad52016000987a7a9a5bc214d7ff61e137ba099b20fc6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:13Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:35Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:35 crc kubenswrapper[4754]: I0126 16:47:35.918848 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-2blzw" event={"ID":"ca65468e-30c6-4666-962e-cc0de05e67fe","Type":"ContainerStarted","Data":"d4603235bb2af495901926b805fe1c0ffc48824a4394ccf928e7fcd53a673ac3"} Jan 26 16:47:35 crc kubenswrapper[4754]: I0126 16:47:35.919116 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-2blzw" event={"ID":"ca65468e-30c6-4666-962e-cc0de05e67fe","Type":"ContainerStarted","Data":"a9ec824f384dcb81e04d3c21ced811379da15347c09c11f8eb4891498b42c455"} Jan 26 16:47:35 crc kubenswrapper[4754]: I0126 16:47:35.920698 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://205ec256e46a8e08b4b0b0805fb6c5f4966fc3ccef0738df4b97cb404e74da17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:35Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:35 crc kubenswrapper[4754]: I0126 16:47:35.921160 4754 generic.go:334] "Generic (PLEG): container finished" podID="351cc6c0-b8e5-4589-ae63-f98e8ab1eca3" containerID="1772bb41eb8a36a863d3c691610bcbee5d16b260480b3e9f9d6b083e09e42799" exitCode=0 Jan 26 16:47:35 crc kubenswrapper[4754]: I0126 16:47:35.921214 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-bwpd6" event={"ID":"351cc6c0-b8e5-4589-ae63-f98e8ab1eca3","Type":"ContainerDied","Data":"1772bb41eb8a36a863d3c691610bcbee5d16b260480b3e9f9d6b083e09e42799"} Jan 26 16:47:35 crc kubenswrapper[4754]: I0126 16:47:35.927544 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jsbxt" event={"ID":"689569c4-93ee-4b82-93fc-9af0f97e6dc3","Type":"ContainerStarted","Data":"b1ac1f49bd564750fcff530a0f76657b0ed41573539acd020a922d19861ef834"} Jan 26 16:47:35 crc kubenswrapper[4754]: I0126 16:47:35.927618 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jsbxt" event={"ID":"689569c4-93ee-4b82-93fc-9af0f97e6dc3","Type":"ContainerStarted","Data":"a300a80ad87f46ba1b4fd7ad3cf6c25767fcfe966b93bdd86d729843f53a5cda"} Jan 26 16:47:35 crc kubenswrapper[4754]: I0126 16:47:35.927631 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jsbxt" event={"ID":"689569c4-93ee-4b82-93fc-9af0f97e6dc3","Type":"ContainerStarted","Data":"919572e731813a98c7d8a94879bba931f75c472c7cd681e86d5197ebab43815d"} Jan 26 16:47:35 crc kubenswrapper[4754]: I0126 16:47:35.927646 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jsbxt" event={"ID":"689569c4-93ee-4b82-93fc-9af0f97e6dc3","Type":"ContainerStarted","Data":"033f8b104d4e9ff6d5f87ca998665f12e0c56696b3d30e01ede05f4da540365d"} Jan 26 16:47:35 crc kubenswrapper[4754]: I0126 16:47:35.927680 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jsbxt" event={"ID":"689569c4-93ee-4b82-93fc-9af0f97e6dc3","Type":"ContainerStarted","Data":"31dcab2f9a2d218b0e49dad1df46377ff4521592824a0fbac254d5b8e6d0632e"} Jan 26 16:47:35 crc kubenswrapper[4754]: I0126 16:47:35.927692 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jsbxt" event={"ID":"689569c4-93ee-4b82-93fc-9af0f97e6dc3","Type":"ContainerStarted","Data":"3fe4d9395447ac570eca1338e749dbefea76295d427a12487b63a142a7d6a2e4"} Jan 26 16:47:35 crc kubenswrapper[4754]: I0126 16:47:35.962963 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:35Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:35 crc kubenswrapper[4754]: I0126 16:47:35.971132 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:35 crc kubenswrapper[4754]: I0126 16:47:35.971193 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:35 crc kubenswrapper[4754]: I0126 16:47:35.971208 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:35 crc kubenswrapper[4754]: I0126 16:47:35.971234 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:35 crc kubenswrapper[4754]: I0126 16:47:35.971250 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:35Z","lastTransitionTime":"2026-01-26T16:47:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:36 crc kubenswrapper[4754]: I0126 16:47:36.000266 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c3718a4-f354-4284-92e0-fdfb45a692bd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b86f46f49163bca7eca9458279029af330dbd4890b03325e06fa3b1bdf030de6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxjcz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fffc03b536bfee3c7bc1bddd500e834e34b391896477a1501e3e06201a19374\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxjcz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-x65wv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:35Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:36 crc kubenswrapper[4754]: I0126 16:47:36.040095 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b89aa081-0a2e-4757-9210-e2a6c228bee9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9c5f40188fda14b14128514fe84bddfe8dcede3d6c560283ccb2ac0c66895fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f45445cbf1d180d6d4dd9588e85bf678f043bfce8cff0812b78ca3510904f9a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://085fa7b8d4fef8b2722fac8c346107699b5a6bd33fbfc74ed35831cb8fdf9dad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c40e018a0979035827963efb9efc48a878a55b9df68ec1a9242d7cff64fe480\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8356f76fbc3a4e636c12d1da5455d3f6767d9b83ce2f2925a11435d7d7261e9f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0126 16:47:26.205208 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0126 16:47:26.206538 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3645934156/tls.crt::/tmp/serving-cert-3645934156/tls.key\\\\\\\"\\\\nI0126 16:47:31.504518 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0126 16:47:31.509366 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0126 16:47:31.509400 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0126 16:47:31.509428 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0126 16:47:31.509436 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0126 16:47:31.515912 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0126 16:47:31.516036 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 16:47:31.516068 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 16:47:31.516096 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0126 16:47:31.516121 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0126 16:47:31.516145 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0126 16:47:31.516170 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0126 16:47:31.516412 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0126 16:47:31.519294 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27b74e2e65a547e0e28fddd4f15feacb17f4d57596eeb8da82df9d39243c85e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22ba7649a1308815bfad52016000987a7a9a5bc214d7ff61e137ba099b20fc6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://22ba7649a1308815bfad52016000987a7a9a5bc214d7ff61e137ba099b20fc6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:13Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:36Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:36 crc kubenswrapper[4754]: I0126 16:47:36.074460 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:36 crc kubenswrapper[4754]: I0126 16:47:36.074513 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:36 crc kubenswrapper[4754]: I0126 16:47:36.074527 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:36 crc kubenswrapper[4754]: I0126 16:47:36.074550 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:36 crc kubenswrapper[4754]: I0126 16:47:36.074580 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:36Z","lastTransitionTime":"2026-01-26T16:47:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:36 crc kubenswrapper[4754]: I0126 16:47:36.081252 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://205ec256e46a8e08b4b0b0805fb6c5f4966fc3ccef0738df4b97cb404e74da17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:36Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:36 crc kubenswrapper[4754]: I0126 16:47:36.121011 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:36Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:36 crc kubenswrapper[4754]: I0126 16:47:36.159974 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c3718a4-f354-4284-92e0-fdfb45a692bd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b86f46f49163bca7eca9458279029af330dbd4890b03325e06fa3b1bdf030de6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxjcz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fffc03b536bfee3c7bc1bddd500e834e34b391896477a1501e3e06201a19374\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxjcz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-x65wv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:36Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:36 crc kubenswrapper[4754]: I0126 16:47:36.177077 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:36 crc kubenswrapper[4754]: I0126 16:47:36.177114 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:36 crc kubenswrapper[4754]: I0126 16:47:36.177122 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:36 crc kubenswrapper[4754]: I0126 16:47:36.177137 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:36 crc kubenswrapper[4754]: I0126 16:47:36.177147 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:36Z","lastTransitionTime":"2026-01-26T16:47:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:36 crc kubenswrapper[4754]: I0126 16:47:36.201647 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:36Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:36 crc kubenswrapper[4754]: I0126 16:47:36.243231 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e459fd8c1093d6f3d0a671751b792b26b7df94c21ced1be1a2250821bcc82d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:36Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:36 crc kubenswrapper[4754]: I0126 16:47:36.277899 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2blzw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca65468e-30c6-4666-962e-cc0de05e67fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4603235bb2af495901926b805fe1c0ffc48824a4394ccf928e7fcd53a673ac3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c25z4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:33Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2blzw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:36Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:36 crc kubenswrapper[4754]: I0126 16:47:36.279239 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:36 crc kubenswrapper[4754]: I0126 16:47:36.279268 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:36 crc kubenswrapper[4754]: I0126 16:47:36.279279 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:36 crc kubenswrapper[4754]: I0126 16:47:36.279296 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:36 crc kubenswrapper[4754]: I0126 16:47:36.279307 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:36Z","lastTransitionTime":"2026-01-26T16:47:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:36 crc kubenswrapper[4754]: I0126 16:47:36.320909 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bwpd6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"351cc6c0-b8e5-4589-ae63-f98e8ab1eca3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3ce0fe080a094316b9456057e347b90a5ff432daa12b9c56d7aea9bee1f72b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3ce0fe080a094316b9456057e347b90a5ff432daa12b9c56d7aea9bee1f72b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1772bb41eb8a36a863d3c691610bcbee5d16b260480b3e9f9d6b083e09e42799\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1772bb41eb8a36a863d3c691610bcbee5d16b260480b3e9f9d6b083e09e42799\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bwpd6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:36Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:36 crc kubenswrapper[4754]: I0126 16:47:36.367965 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jsbxt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"689569c4-93ee-4b82-93fc-9af0f97e6dc3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d9a7bafabd44f8b88a6f62cbe45599af8563bc45d9d7ad3a7669f3d8d11eb9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d9a7bafabd44f8b88a6f62cbe45599af8563bc45d9d7ad3a7669f3d8d11eb9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jsbxt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:36Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:36 crc kubenswrapper[4754]: I0126 16:47:36.381420 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:36 crc kubenswrapper[4754]: I0126 16:47:36.381460 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:36 crc kubenswrapper[4754]: I0126 16:47:36.381470 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:36 crc kubenswrapper[4754]: I0126 16:47:36.381490 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:36 crc kubenswrapper[4754]: I0126 16:47:36.381501 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:36Z","lastTransitionTime":"2026-01-26T16:47:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:36 crc kubenswrapper[4754]: I0126 16:47:36.400757 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"099af702-abf2-4933-b4c9-63ef2326e468\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23017e73f8450839b4c2c02fa61235ffe4604bdce11556def27f25bcadcf89dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a74507c941b9de4d0401c8f1d53bb79ed11359628d3405cdab7f5474a75c862e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4402182b3e2e35a7d5dd4c44ac53e152915b312b7ccb076c4f2b41b925a6e7f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f772a003566198519fe1849fd14f29d4346004b2e655810662507c95b4d5ffa1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:13Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:36Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:36 crc kubenswrapper[4754]: I0126 16:47:36.439346 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:36Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:36 crc kubenswrapper[4754]: I0126 16:47:36.481039 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89251fa9c64bc25d6d6f355400d7644c099f22aed1cae812fc991a03595f1e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e90508d8a53be9146f9b388750730fdae9ee200306caf5ad393a356390a69ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:36Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:36 crc kubenswrapper[4754]: I0126 16:47:36.484363 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:36 crc kubenswrapper[4754]: I0126 16:47:36.484434 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:36 crc kubenswrapper[4754]: I0126 16:47:36.484448 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:36 crc kubenswrapper[4754]: I0126 16:47:36.484468 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:36 crc kubenswrapper[4754]: I0126 16:47:36.484493 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:36Z","lastTransitionTime":"2026-01-26T16:47:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:36 crc kubenswrapper[4754]: I0126 16:47:36.518519 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-84lgq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6763d962-0892-4acc-bab9-7a4e93e39a85\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://953c1917fca896c3def3fc2c69f9e3dbdc9ca26fc2b3d40282be71f71759fd68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8zlw6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:33Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-84lgq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:36Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:36 crc kubenswrapper[4754]: I0126 16:47:36.561105 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-skh5n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b619d233-b592-4b05-a0b7-dc094c88471e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9231a4191565e2e0ef53d8a394bb34615c900b601a0dfec24c2d8e5fa7b0b659\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-skh5n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:36Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:36 crc kubenswrapper[4754]: I0126 16:47:36.587264 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:36 crc kubenswrapper[4754]: I0126 16:47:36.587320 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:36 crc kubenswrapper[4754]: I0126 16:47:36.587330 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:36 crc kubenswrapper[4754]: I0126 16:47:36.587349 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:36 crc kubenswrapper[4754]: I0126 16:47:36.587361 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:36Z","lastTransitionTime":"2026-01-26T16:47:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:36 crc kubenswrapper[4754]: I0126 16:47:36.690504 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:36 crc kubenswrapper[4754]: I0126 16:47:36.690552 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:36 crc kubenswrapper[4754]: I0126 16:47:36.690563 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:36 crc kubenswrapper[4754]: I0126 16:47:36.690581 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:36 crc kubenswrapper[4754]: I0126 16:47:36.690592 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:36Z","lastTransitionTime":"2026-01-26T16:47:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:36 crc kubenswrapper[4754]: I0126 16:47:36.722725 4754 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-04 10:34:58.204921399 +0000 UTC Jan 26 16:47:36 crc kubenswrapper[4754]: I0126 16:47:36.795703 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:36 crc kubenswrapper[4754]: I0126 16:47:36.795764 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:36 crc kubenswrapper[4754]: I0126 16:47:36.795776 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:36 crc kubenswrapper[4754]: I0126 16:47:36.795800 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:36 crc kubenswrapper[4754]: I0126 16:47:36.795811 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:36Z","lastTransitionTime":"2026-01-26T16:47:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:36 crc kubenswrapper[4754]: I0126 16:47:36.899456 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:36 crc kubenswrapper[4754]: I0126 16:47:36.899506 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:36 crc kubenswrapper[4754]: I0126 16:47:36.899519 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:36 crc kubenswrapper[4754]: I0126 16:47:36.899537 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:36 crc kubenswrapper[4754]: I0126 16:47:36.899548 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:36Z","lastTransitionTime":"2026-01-26T16:47:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:36 crc kubenswrapper[4754]: I0126 16:47:36.933687 4754 generic.go:334] "Generic (PLEG): container finished" podID="351cc6c0-b8e5-4589-ae63-f98e8ab1eca3" containerID="24848d69756024cc664affb9492764ace0638ef611f910f4f2bf86b638cc443d" exitCode=0 Jan 26 16:47:36 crc kubenswrapper[4754]: I0126 16:47:36.933908 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-bwpd6" event={"ID":"351cc6c0-b8e5-4589-ae63-f98e8ab1eca3","Type":"ContainerDied","Data":"24848d69756024cc664affb9492764ace0638ef611f910f4f2bf86b638cc443d"} Jan 26 16:47:36 crc kubenswrapper[4754]: I0126 16:47:36.950812 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2blzw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca65468e-30c6-4666-962e-cc0de05e67fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4603235bb2af495901926b805fe1c0ffc48824a4394ccf928e7fcd53a673ac3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c25z4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:33Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2blzw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:36Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:36 crc kubenswrapper[4754]: I0126 16:47:36.972506 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bwpd6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"351cc6c0-b8e5-4589-ae63-f98e8ab1eca3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3ce0fe080a094316b9456057e347b90a5ff432daa12b9c56d7aea9bee1f72b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3ce0fe080a094316b9456057e347b90a5ff432daa12b9c56d7aea9bee1f72b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1772bb41eb8a36a863d3c691610bcbee5d16b260480b3e9f9d6b083e09e42799\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1772bb41eb8a36a863d3c691610bcbee5d16b260480b3e9f9d6b083e09e42799\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24848d69756024cc664affb9492764ace0638ef611f910f4f2bf86b638cc443d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://24848d69756024cc664affb9492764ace0638ef611f910f4f2bf86b638cc443d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bwpd6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:36Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:36 crc kubenswrapper[4754]: I0126 16:47:36.996252 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jsbxt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"689569c4-93ee-4b82-93fc-9af0f97e6dc3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d9a7bafabd44f8b88a6f62cbe45599af8563bc45d9d7ad3a7669f3d8d11eb9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d9a7bafabd44f8b88a6f62cbe45599af8563bc45d9d7ad3a7669f3d8d11eb9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jsbxt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:36Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:37 crc kubenswrapper[4754]: I0126 16:47:37.003244 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:37 crc kubenswrapper[4754]: I0126 16:47:37.003304 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:37 crc kubenswrapper[4754]: I0126 16:47:37.003317 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:37 crc kubenswrapper[4754]: I0126 16:47:37.003338 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:37 crc kubenswrapper[4754]: I0126 16:47:37.003350 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:37Z","lastTransitionTime":"2026-01-26T16:47:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:37 crc kubenswrapper[4754]: I0126 16:47:37.013270 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:37Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:37 crc kubenswrapper[4754]: I0126 16:47:37.027009 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e459fd8c1093d6f3d0a671751b792b26b7df94c21ced1be1a2250821bcc82d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:37Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:37 crc kubenswrapper[4754]: I0126 16:47:37.040248 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89251fa9c64bc25d6d6f355400d7644c099f22aed1cae812fc991a03595f1e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e90508d8a53be9146f9b388750730fdae9ee200306caf5ad393a356390a69ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:37Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:37 crc kubenswrapper[4754]: I0126 16:47:37.053406 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-84lgq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6763d962-0892-4acc-bab9-7a4e93e39a85\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://953c1917fca896c3def3fc2c69f9e3dbdc9ca26fc2b3d40282be71f71759fd68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8zlw6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:33Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-84lgq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:37Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:37 crc kubenswrapper[4754]: I0126 16:47:37.070572 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-skh5n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b619d233-b592-4b05-a0b7-dc094c88471e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9231a4191565e2e0ef53d8a394bb34615c900b601a0dfec24c2d8e5fa7b0b659\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-skh5n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:37Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:37 crc kubenswrapper[4754]: I0126 16:47:37.084058 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"099af702-abf2-4933-b4c9-63ef2326e468\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23017e73f8450839b4c2c02fa61235ffe4604bdce11556def27f25bcadcf89dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a74507c941b9de4d0401c8f1d53bb79ed11359628d3405cdab7f5474a75c862e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4402182b3e2e35a7d5dd4c44ac53e152915b312b7ccb076c4f2b41b925a6e7f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f772a003566198519fe1849fd14f29d4346004b2e655810662507c95b4d5ffa1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:13Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:37Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:37 crc kubenswrapper[4754]: I0126 16:47:37.099025 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:37Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:37 crc kubenswrapper[4754]: I0126 16:47:37.105475 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:37 crc kubenswrapper[4754]: I0126 16:47:37.105510 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:37 crc kubenswrapper[4754]: I0126 16:47:37.105519 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:37 crc kubenswrapper[4754]: I0126 16:47:37.105535 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:37 crc kubenswrapper[4754]: I0126 16:47:37.105545 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:37Z","lastTransitionTime":"2026-01-26T16:47:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:37 crc kubenswrapper[4754]: I0126 16:47:37.113981 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c3718a4-f354-4284-92e0-fdfb45a692bd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b86f46f49163bca7eca9458279029af330dbd4890b03325e06fa3b1bdf030de6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxjcz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fffc03b536bfee3c7bc1bddd500e834e34b391896477a1501e3e06201a19374\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxjcz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-x65wv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:37Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:37 crc kubenswrapper[4754]: I0126 16:47:37.129472 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b89aa081-0a2e-4757-9210-e2a6c228bee9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9c5f40188fda14b14128514fe84bddfe8dcede3d6c560283ccb2ac0c66895fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f45445cbf1d180d6d4dd9588e85bf678f043bfce8cff0812b78ca3510904f9a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://085fa7b8d4fef8b2722fac8c346107699b5a6bd33fbfc74ed35831cb8fdf9dad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c40e018a0979035827963efb9efc48a878a55b9df68ec1a9242d7cff64fe480\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8356f76fbc3a4e636c12d1da5455d3f6767d9b83ce2f2925a11435d7d7261e9f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0126 16:47:26.205208 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0126 16:47:26.206538 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3645934156/tls.crt::/tmp/serving-cert-3645934156/tls.key\\\\\\\"\\\\nI0126 16:47:31.504518 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0126 16:47:31.509366 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0126 16:47:31.509400 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0126 16:47:31.509428 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0126 16:47:31.509436 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0126 16:47:31.515912 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0126 16:47:31.516036 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 16:47:31.516068 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 16:47:31.516096 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0126 16:47:31.516121 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0126 16:47:31.516145 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0126 16:47:31.516170 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0126 16:47:31.516412 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0126 16:47:31.519294 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27b74e2e65a547e0e28fddd4f15feacb17f4d57596eeb8da82df9d39243c85e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22ba7649a1308815bfad52016000987a7a9a5bc214d7ff61e137ba099b20fc6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://22ba7649a1308815bfad52016000987a7a9a5bc214d7ff61e137ba099b20fc6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:13Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:37Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:37 crc kubenswrapper[4754]: I0126 16:47:37.144571 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://205ec256e46a8e08b4b0b0805fb6c5f4966fc3ccef0738df4b97cb404e74da17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:37Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:37 crc kubenswrapper[4754]: I0126 16:47:37.161518 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:37Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:37 crc kubenswrapper[4754]: I0126 16:47:37.208388 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:37 crc kubenswrapper[4754]: I0126 16:47:37.208456 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:37 crc kubenswrapper[4754]: I0126 16:47:37.208470 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:37 crc kubenswrapper[4754]: I0126 16:47:37.208494 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:37 crc kubenswrapper[4754]: I0126 16:47:37.208509 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:37Z","lastTransitionTime":"2026-01-26T16:47:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:37 crc kubenswrapper[4754]: I0126 16:47:37.311309 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:37 crc kubenswrapper[4754]: I0126 16:47:37.311538 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:37 crc kubenswrapper[4754]: I0126 16:47:37.311614 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:37 crc kubenswrapper[4754]: I0126 16:47:37.311733 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:37 crc kubenswrapper[4754]: I0126 16:47:37.311840 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:37Z","lastTransitionTime":"2026-01-26T16:47:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:37 crc kubenswrapper[4754]: I0126 16:47:37.414369 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:37 crc kubenswrapper[4754]: I0126 16:47:37.414397 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:37 crc kubenswrapper[4754]: I0126 16:47:37.414405 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:37 crc kubenswrapper[4754]: I0126 16:47:37.414416 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:37 crc kubenswrapper[4754]: I0126 16:47:37.414424 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:37Z","lastTransitionTime":"2026-01-26T16:47:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:37 crc kubenswrapper[4754]: I0126 16:47:37.517166 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:37 crc kubenswrapper[4754]: I0126 16:47:37.517450 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:37 crc kubenswrapper[4754]: I0126 16:47:37.517589 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:37 crc kubenswrapper[4754]: I0126 16:47:37.517704 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:37 crc kubenswrapper[4754]: I0126 16:47:37.517839 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:37Z","lastTransitionTime":"2026-01-26T16:47:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:37 crc kubenswrapper[4754]: I0126 16:47:37.621412 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:37 crc kubenswrapper[4754]: I0126 16:47:37.621482 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:37 crc kubenswrapper[4754]: I0126 16:47:37.621497 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:37 crc kubenswrapper[4754]: I0126 16:47:37.621522 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:37 crc kubenswrapper[4754]: I0126 16:47:37.621541 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:37Z","lastTransitionTime":"2026-01-26T16:47:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:37 crc kubenswrapper[4754]: I0126 16:47:37.722875 4754 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-10 10:41:13.371224534 +0000 UTC Jan 26 16:47:37 crc kubenswrapper[4754]: I0126 16:47:37.723826 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:37 crc kubenswrapper[4754]: I0126 16:47:37.723856 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:37 crc kubenswrapper[4754]: I0126 16:47:37.723866 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:37 crc kubenswrapper[4754]: I0126 16:47:37.723880 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:37 crc kubenswrapper[4754]: I0126 16:47:37.723890 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:37Z","lastTransitionTime":"2026-01-26T16:47:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:37 crc kubenswrapper[4754]: I0126 16:47:37.767426 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 16:47:37 crc kubenswrapper[4754]: I0126 16:47:37.767528 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 16:47:37 crc kubenswrapper[4754]: E0126 16:47:37.767604 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 16:47:37 crc kubenswrapper[4754]: I0126 16:47:37.767613 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 16:47:37 crc kubenswrapper[4754]: E0126 16:47:37.767764 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 16:47:37 crc kubenswrapper[4754]: E0126 16:47:37.767935 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 16:47:37 crc kubenswrapper[4754]: I0126 16:47:37.826732 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:37 crc kubenswrapper[4754]: I0126 16:47:37.826774 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:37 crc kubenswrapper[4754]: I0126 16:47:37.826786 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:37 crc kubenswrapper[4754]: I0126 16:47:37.826805 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:37 crc kubenswrapper[4754]: I0126 16:47:37.826817 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:37Z","lastTransitionTime":"2026-01-26T16:47:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:37 crc kubenswrapper[4754]: I0126 16:47:37.928717 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:37 crc kubenswrapper[4754]: I0126 16:47:37.928765 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:37 crc kubenswrapper[4754]: I0126 16:47:37.928778 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:37 crc kubenswrapper[4754]: I0126 16:47:37.928797 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:37 crc kubenswrapper[4754]: I0126 16:47:37.928809 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:37Z","lastTransitionTime":"2026-01-26T16:47:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:37 crc kubenswrapper[4754]: I0126 16:47:37.951597 4754 generic.go:334] "Generic (PLEG): container finished" podID="351cc6c0-b8e5-4589-ae63-f98e8ab1eca3" containerID="c264fc7c3300a725566acce4464f01d9314779737624401c164d867831fd93bf" exitCode=0 Jan 26 16:47:37 crc kubenswrapper[4754]: I0126 16:47:37.951710 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-bwpd6" event={"ID":"351cc6c0-b8e5-4589-ae63-f98e8ab1eca3","Type":"ContainerDied","Data":"c264fc7c3300a725566acce4464f01d9314779737624401c164d867831fd93bf"} Jan 26 16:47:37 crc kubenswrapper[4754]: I0126 16:47:37.957508 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jsbxt" event={"ID":"689569c4-93ee-4b82-93fc-9af0f97e6dc3","Type":"ContainerStarted","Data":"07a43ff7860ff65faf51f4a5e8aeb326f06d6d7222677a92560ae2e3a0c81484"} Jan 26 16:47:37 crc kubenswrapper[4754]: I0126 16:47:37.971046 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e459fd8c1093d6f3d0a671751b792b26b7df94c21ced1be1a2250821bcc82d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:37Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:37 crc kubenswrapper[4754]: I0126 16:47:37.985403 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2blzw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca65468e-30c6-4666-962e-cc0de05e67fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4603235bb2af495901926b805fe1c0ffc48824a4394ccf928e7fcd53a673ac3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c25z4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:33Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2blzw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:37Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:38 crc kubenswrapper[4754]: I0126 16:47:38.003416 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bwpd6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"351cc6c0-b8e5-4589-ae63-f98e8ab1eca3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3ce0fe080a094316b9456057e347b90a5ff432daa12b9c56d7aea9bee1f72b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3ce0fe080a094316b9456057e347b90a5ff432daa12b9c56d7aea9bee1f72b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1772bb41eb8a36a863d3c691610bcbee5d16b260480b3e9f9d6b083e09e42799\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1772bb41eb8a36a863d3c691610bcbee5d16b260480b3e9f9d6b083e09e42799\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24848d69756024cc664affb9492764ace0638ef611f910f4f2bf86b638cc443d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://24848d69756024cc664affb9492764ace0638ef611f910f4f2bf86b638cc443d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c264fc7c3300a725566acce4464f01d9314779737624401c164d867831fd93bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c264fc7c3300a725566acce4464f01d9314779737624401c164d867831fd93bf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bwpd6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:37Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:38 crc kubenswrapper[4754]: I0126 16:47:38.024012 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jsbxt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"689569c4-93ee-4b82-93fc-9af0f97e6dc3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d9a7bafabd44f8b88a6f62cbe45599af8563bc45d9d7ad3a7669f3d8d11eb9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d9a7bafabd44f8b88a6f62cbe45599af8563bc45d9d7ad3a7669f3d8d11eb9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jsbxt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:38Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:38 crc kubenswrapper[4754]: I0126 16:47:38.032514 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:38 crc kubenswrapper[4754]: I0126 16:47:38.032567 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:38 crc kubenswrapper[4754]: I0126 16:47:38.032579 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:38 crc kubenswrapper[4754]: I0126 16:47:38.032632 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:38 crc kubenswrapper[4754]: I0126 16:47:38.032646 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:38Z","lastTransitionTime":"2026-01-26T16:47:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:38 crc kubenswrapper[4754]: I0126 16:47:38.038314 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:38Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:38 crc kubenswrapper[4754]: I0126 16:47:38.055998 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:38Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:38 crc kubenswrapper[4754]: I0126 16:47:38.070394 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89251fa9c64bc25d6d6f355400d7644c099f22aed1cae812fc991a03595f1e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e90508d8a53be9146f9b388750730fdae9ee200306caf5ad393a356390a69ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:38Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:38 crc kubenswrapper[4754]: I0126 16:47:38.086077 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-84lgq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6763d962-0892-4acc-bab9-7a4e93e39a85\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://953c1917fca896c3def3fc2c69f9e3dbdc9ca26fc2b3d40282be71f71759fd68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8zlw6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:33Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-84lgq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:38Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:38 crc kubenswrapper[4754]: I0126 16:47:38.102371 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-skh5n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b619d233-b592-4b05-a0b7-dc094c88471e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9231a4191565e2e0ef53d8a394bb34615c900b601a0dfec24c2d8e5fa7b0b659\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-skh5n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:38Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:38 crc kubenswrapper[4754]: I0126 16:47:38.117528 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"099af702-abf2-4933-b4c9-63ef2326e468\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23017e73f8450839b4c2c02fa61235ffe4604bdce11556def27f25bcadcf89dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a74507c941b9de4d0401c8f1d53bb79ed11359628d3405cdab7f5474a75c862e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4402182b3e2e35a7d5dd4c44ac53e152915b312b7ccb076c4f2b41b925a6e7f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f772a003566198519fe1849fd14f29d4346004b2e655810662507c95b4d5ffa1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:13Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:38Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:38 crc kubenswrapper[4754]: I0126 16:47:38.134806 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:38 crc kubenswrapper[4754]: I0126 16:47:38.134865 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:38 crc kubenswrapper[4754]: I0126 16:47:38.134874 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:38 crc kubenswrapper[4754]: I0126 16:47:38.134890 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:38 crc kubenswrapper[4754]: I0126 16:47:38.134902 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:38Z","lastTransitionTime":"2026-01-26T16:47:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:38 crc kubenswrapper[4754]: I0126 16:47:38.134907 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:38Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:38 crc kubenswrapper[4754]: I0126 16:47:38.151597 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c3718a4-f354-4284-92e0-fdfb45a692bd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b86f46f49163bca7eca9458279029af330dbd4890b03325e06fa3b1bdf030de6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxjcz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fffc03b536bfee3c7bc1bddd500e834e34b391896477a1501e3e06201a19374\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxjcz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-x65wv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:38Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:38 crc kubenswrapper[4754]: I0126 16:47:38.168368 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b89aa081-0a2e-4757-9210-e2a6c228bee9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9c5f40188fda14b14128514fe84bddfe8dcede3d6c560283ccb2ac0c66895fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f45445cbf1d180d6d4dd9588e85bf678f043bfce8cff0812b78ca3510904f9a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://085fa7b8d4fef8b2722fac8c346107699b5a6bd33fbfc74ed35831cb8fdf9dad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c40e018a0979035827963efb9efc48a878a55b9df68ec1a9242d7cff64fe480\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8356f76fbc3a4e636c12d1da5455d3f6767d9b83ce2f2925a11435d7d7261e9f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0126 16:47:26.205208 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0126 16:47:26.206538 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3645934156/tls.crt::/tmp/serving-cert-3645934156/tls.key\\\\\\\"\\\\nI0126 16:47:31.504518 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0126 16:47:31.509366 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0126 16:47:31.509400 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0126 16:47:31.509428 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0126 16:47:31.509436 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0126 16:47:31.515912 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0126 16:47:31.516036 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 16:47:31.516068 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 16:47:31.516096 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0126 16:47:31.516121 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0126 16:47:31.516145 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0126 16:47:31.516170 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0126 16:47:31.516412 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0126 16:47:31.519294 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27b74e2e65a547e0e28fddd4f15feacb17f4d57596eeb8da82df9d39243c85e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22ba7649a1308815bfad52016000987a7a9a5bc214d7ff61e137ba099b20fc6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://22ba7649a1308815bfad52016000987a7a9a5bc214d7ff61e137ba099b20fc6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:13Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:38Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:38 crc kubenswrapper[4754]: I0126 16:47:38.183723 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://205ec256e46a8e08b4b0b0805fb6c5f4966fc3ccef0738df4b97cb404e74da17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:38Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:38 crc kubenswrapper[4754]: I0126 16:47:38.237791 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:38 crc kubenswrapper[4754]: I0126 16:47:38.237846 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:38 crc kubenswrapper[4754]: I0126 16:47:38.237858 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:38 crc kubenswrapper[4754]: I0126 16:47:38.237913 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:38 crc kubenswrapper[4754]: I0126 16:47:38.237926 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:38Z","lastTransitionTime":"2026-01-26T16:47:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:38 crc kubenswrapper[4754]: I0126 16:47:38.341445 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:38 crc kubenswrapper[4754]: I0126 16:47:38.341488 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:38 crc kubenswrapper[4754]: I0126 16:47:38.341497 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:38 crc kubenswrapper[4754]: I0126 16:47:38.341513 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:38 crc kubenswrapper[4754]: I0126 16:47:38.341551 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:38Z","lastTransitionTime":"2026-01-26T16:47:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:38 crc kubenswrapper[4754]: I0126 16:47:38.444051 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:38 crc kubenswrapper[4754]: I0126 16:47:38.444105 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:38 crc kubenswrapper[4754]: I0126 16:47:38.444116 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:38 crc kubenswrapper[4754]: I0126 16:47:38.444138 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:38 crc kubenswrapper[4754]: I0126 16:47:38.444149 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:38Z","lastTransitionTime":"2026-01-26T16:47:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:38 crc kubenswrapper[4754]: I0126 16:47:38.549005 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:38 crc kubenswrapper[4754]: I0126 16:47:38.549437 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:38 crc kubenswrapper[4754]: I0126 16:47:38.549492 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:38 crc kubenswrapper[4754]: I0126 16:47:38.549519 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:38 crc kubenswrapper[4754]: I0126 16:47:38.549533 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:38Z","lastTransitionTime":"2026-01-26T16:47:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:38 crc kubenswrapper[4754]: I0126 16:47:38.653047 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:38 crc kubenswrapper[4754]: I0126 16:47:38.653117 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:38 crc kubenswrapper[4754]: I0126 16:47:38.653129 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:38 crc kubenswrapper[4754]: I0126 16:47:38.653152 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:38 crc kubenswrapper[4754]: I0126 16:47:38.653165 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:38Z","lastTransitionTime":"2026-01-26T16:47:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:38 crc kubenswrapper[4754]: I0126 16:47:38.724035 4754 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-29 20:30:26.728536505 +0000 UTC Jan 26 16:47:38 crc kubenswrapper[4754]: I0126 16:47:38.756093 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:38 crc kubenswrapper[4754]: I0126 16:47:38.756128 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:38 crc kubenswrapper[4754]: I0126 16:47:38.756138 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:38 crc kubenswrapper[4754]: I0126 16:47:38.756162 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:38 crc kubenswrapper[4754]: I0126 16:47:38.756171 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:38Z","lastTransitionTime":"2026-01-26T16:47:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:38 crc kubenswrapper[4754]: I0126 16:47:38.858392 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:38 crc kubenswrapper[4754]: I0126 16:47:38.858454 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:38 crc kubenswrapper[4754]: I0126 16:47:38.858469 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:38 crc kubenswrapper[4754]: I0126 16:47:38.858491 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:38 crc kubenswrapper[4754]: I0126 16:47:38.858504 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:38Z","lastTransitionTime":"2026-01-26T16:47:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:38 crc kubenswrapper[4754]: I0126 16:47:38.961512 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:38 crc kubenswrapper[4754]: I0126 16:47:38.961567 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:38 crc kubenswrapper[4754]: I0126 16:47:38.961579 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:38 crc kubenswrapper[4754]: I0126 16:47:38.961598 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:38 crc kubenswrapper[4754]: I0126 16:47:38.961610 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:38Z","lastTransitionTime":"2026-01-26T16:47:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:38 crc kubenswrapper[4754]: I0126 16:47:38.963574 4754 generic.go:334] "Generic (PLEG): container finished" podID="351cc6c0-b8e5-4589-ae63-f98e8ab1eca3" containerID="5acfe327eee4aad927ac3a32d24549eff788b14a1c4bdac359b6b56301d884ee" exitCode=0 Jan 26 16:47:38 crc kubenswrapper[4754]: I0126 16:47:38.963659 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-bwpd6" event={"ID":"351cc6c0-b8e5-4589-ae63-f98e8ab1eca3","Type":"ContainerDied","Data":"5acfe327eee4aad927ac3a32d24549eff788b14a1c4bdac359b6b56301d884ee"} Jan 26 16:47:38 crc kubenswrapper[4754]: I0126 16:47:38.981177 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:38Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:38 crc kubenswrapper[4754]: I0126 16:47:38.992882 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e459fd8c1093d6f3d0a671751b792b26b7df94c21ced1be1a2250821bcc82d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:38Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:39 crc kubenswrapper[4754]: I0126 16:47:39.001910 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2blzw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca65468e-30c6-4666-962e-cc0de05e67fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4603235bb2af495901926b805fe1c0ffc48824a4394ccf928e7fcd53a673ac3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c25z4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:33Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2blzw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:39Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:39 crc kubenswrapper[4754]: I0126 16:47:39.031775 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bwpd6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"351cc6c0-b8e5-4589-ae63-f98e8ab1eca3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3ce0fe080a094316b9456057e347b90a5ff432daa12b9c56d7aea9bee1f72b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3ce0fe080a094316b9456057e347b90a5ff432daa12b9c56d7aea9bee1f72b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1772bb41eb8a36a863d3c691610bcbee5d16b260480b3e9f9d6b083e09e42799\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1772bb41eb8a36a863d3c691610bcbee5d16b260480b3e9f9d6b083e09e42799\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24848d69756024cc664affb9492764ace0638ef611f910f4f2bf86b638cc443d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://24848d69756024cc664affb9492764ace0638ef611f910f4f2bf86b638cc443d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c264fc7c3300a725566acce4464f01d9314779737624401c164d867831fd93bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c264fc7c3300a725566acce4464f01d9314779737624401c164d867831fd93bf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5acfe327eee4aad927ac3a32d24549eff788b14a1c4bdac359b6b56301d884ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5acfe327eee4aad927ac3a32d24549eff788b14a1c4bdac359b6b56301d884ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bwpd6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:39Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:39 crc kubenswrapper[4754]: I0126 16:47:39.052339 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jsbxt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"689569c4-93ee-4b82-93fc-9af0f97e6dc3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d9a7bafabd44f8b88a6f62cbe45599af8563bc45d9d7ad3a7669f3d8d11eb9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d9a7bafabd44f8b88a6f62cbe45599af8563bc45d9d7ad3a7669f3d8d11eb9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jsbxt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:39Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:39 crc kubenswrapper[4754]: I0126 16:47:39.064879 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:39 crc kubenswrapper[4754]: I0126 16:47:39.064988 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:39 crc kubenswrapper[4754]: I0126 16:47:39.065003 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:39 crc kubenswrapper[4754]: I0126 16:47:39.065032 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:39 crc kubenswrapper[4754]: I0126 16:47:39.065052 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:39Z","lastTransitionTime":"2026-01-26T16:47:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:39 crc kubenswrapper[4754]: I0126 16:47:39.069248 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"099af702-abf2-4933-b4c9-63ef2326e468\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23017e73f8450839b4c2c02fa61235ffe4604bdce11556def27f25bcadcf89dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a74507c941b9de4d0401c8f1d53bb79ed11359628d3405cdab7f5474a75c862e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4402182b3e2e35a7d5dd4c44ac53e152915b312b7ccb076c4f2b41b925a6e7f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f772a003566198519fe1849fd14f29d4346004b2e655810662507c95b4d5ffa1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:13Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:39Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:39 crc kubenswrapper[4754]: I0126 16:47:39.086863 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:39Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:39 crc kubenswrapper[4754]: I0126 16:47:39.105455 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89251fa9c64bc25d6d6f355400d7644c099f22aed1cae812fc991a03595f1e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e90508d8a53be9146f9b388750730fdae9ee200306caf5ad393a356390a69ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:39Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:39 crc kubenswrapper[4754]: I0126 16:47:39.118398 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-84lgq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6763d962-0892-4acc-bab9-7a4e93e39a85\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://953c1917fca896c3def3fc2c69f9e3dbdc9ca26fc2b3d40282be71f71759fd68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8zlw6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:33Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-84lgq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:39Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:39 crc kubenswrapper[4754]: I0126 16:47:39.133996 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-skh5n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b619d233-b592-4b05-a0b7-dc094c88471e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9231a4191565e2e0ef53d8a394bb34615c900b601a0dfec24c2d8e5fa7b0b659\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-skh5n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:39Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:39 crc kubenswrapper[4754]: I0126 16:47:39.151136 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b89aa081-0a2e-4757-9210-e2a6c228bee9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9c5f40188fda14b14128514fe84bddfe8dcede3d6c560283ccb2ac0c66895fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f45445cbf1d180d6d4dd9588e85bf678f043bfce8cff0812b78ca3510904f9a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://085fa7b8d4fef8b2722fac8c346107699b5a6bd33fbfc74ed35831cb8fdf9dad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c40e018a0979035827963efb9efc48a878a55b9df68ec1a9242d7cff64fe480\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8356f76fbc3a4e636c12d1da5455d3f6767d9b83ce2f2925a11435d7d7261e9f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0126 16:47:26.205208 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0126 16:47:26.206538 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3645934156/tls.crt::/tmp/serving-cert-3645934156/tls.key\\\\\\\"\\\\nI0126 16:47:31.504518 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0126 16:47:31.509366 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0126 16:47:31.509400 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0126 16:47:31.509428 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0126 16:47:31.509436 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0126 16:47:31.515912 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0126 16:47:31.516036 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 16:47:31.516068 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 16:47:31.516096 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0126 16:47:31.516121 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0126 16:47:31.516145 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0126 16:47:31.516170 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0126 16:47:31.516412 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0126 16:47:31.519294 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27b74e2e65a547e0e28fddd4f15feacb17f4d57596eeb8da82df9d39243c85e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22ba7649a1308815bfad52016000987a7a9a5bc214d7ff61e137ba099b20fc6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://22ba7649a1308815bfad52016000987a7a9a5bc214d7ff61e137ba099b20fc6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:13Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:39Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:39 crc kubenswrapper[4754]: I0126 16:47:39.166833 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://205ec256e46a8e08b4b0b0805fb6c5f4966fc3ccef0738df4b97cb404e74da17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:39Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:39 crc kubenswrapper[4754]: I0126 16:47:39.167785 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:39 crc kubenswrapper[4754]: I0126 16:47:39.167845 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:39 crc kubenswrapper[4754]: I0126 16:47:39.167864 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:39 crc kubenswrapper[4754]: I0126 16:47:39.167892 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:39 crc kubenswrapper[4754]: I0126 16:47:39.167907 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:39Z","lastTransitionTime":"2026-01-26T16:47:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:39 crc kubenswrapper[4754]: I0126 16:47:39.181954 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:39Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:39 crc kubenswrapper[4754]: I0126 16:47:39.195048 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c3718a4-f354-4284-92e0-fdfb45a692bd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b86f46f49163bca7eca9458279029af330dbd4890b03325e06fa3b1bdf030de6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxjcz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fffc03b536bfee3c7bc1bddd500e834e34b391896477a1501e3e06201a19374\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxjcz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-x65wv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:39Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:39 crc kubenswrapper[4754]: I0126 16:47:39.269863 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:39 crc kubenswrapper[4754]: I0126 16:47:39.269902 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:39 crc kubenswrapper[4754]: I0126 16:47:39.269912 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:39 crc kubenswrapper[4754]: I0126 16:47:39.269928 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:39 crc kubenswrapper[4754]: I0126 16:47:39.269939 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:39Z","lastTransitionTime":"2026-01-26T16:47:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:39 crc kubenswrapper[4754]: I0126 16:47:39.372216 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:39 crc kubenswrapper[4754]: I0126 16:47:39.372256 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:39 crc kubenswrapper[4754]: I0126 16:47:39.372269 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:39 crc kubenswrapper[4754]: I0126 16:47:39.372333 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:39 crc kubenswrapper[4754]: I0126 16:47:39.372357 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:39Z","lastTransitionTime":"2026-01-26T16:47:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:39 crc kubenswrapper[4754]: I0126 16:47:39.475817 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:39 crc kubenswrapper[4754]: I0126 16:47:39.475884 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:39 crc kubenswrapper[4754]: I0126 16:47:39.475946 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:39 crc kubenswrapper[4754]: I0126 16:47:39.475982 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:39 crc kubenswrapper[4754]: I0126 16:47:39.475997 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:39Z","lastTransitionTime":"2026-01-26T16:47:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:39 crc kubenswrapper[4754]: I0126 16:47:39.579919 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:39 crc kubenswrapper[4754]: I0126 16:47:39.579967 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:39 crc kubenswrapper[4754]: I0126 16:47:39.579980 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:39 crc kubenswrapper[4754]: I0126 16:47:39.579996 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:39 crc kubenswrapper[4754]: I0126 16:47:39.580007 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:39Z","lastTransitionTime":"2026-01-26T16:47:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:39 crc kubenswrapper[4754]: I0126 16:47:39.604820 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 16:47:39 crc kubenswrapper[4754]: E0126 16:47:39.604954 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 16:47:47.604928891 +0000 UTC m=+34.129109325 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 16:47:39 crc kubenswrapper[4754]: I0126 16:47:39.605008 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 16:47:39 crc kubenswrapper[4754]: I0126 16:47:39.605040 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 16:47:39 crc kubenswrapper[4754]: I0126 16:47:39.605058 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 16:47:39 crc kubenswrapper[4754]: I0126 16:47:39.605080 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 16:47:39 crc kubenswrapper[4754]: E0126 16:47:39.605216 4754 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 26 16:47:39 crc kubenswrapper[4754]: E0126 16:47:39.605225 4754 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 26 16:47:39 crc kubenswrapper[4754]: E0126 16:47:39.605233 4754 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 26 16:47:39 crc kubenswrapper[4754]: E0126 16:47:39.605249 4754 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 26 16:47:39 crc kubenswrapper[4754]: E0126 16:47:39.605287 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-01-26 16:47:47.60527719 +0000 UTC m=+34.129457624 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 26 16:47:39 crc kubenswrapper[4754]: E0126 16:47:39.605338 4754 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 26 16:47:39 crc kubenswrapper[4754]: E0126 16:47:39.605353 4754 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 26 16:47:39 crc kubenswrapper[4754]: E0126 16:47:39.605365 4754 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 26 16:47:39 crc kubenswrapper[4754]: E0126 16:47:39.605365 4754 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Jan 26 16:47:39 crc kubenswrapper[4754]: E0126 16:47:39.605413 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-01-26 16:47:47.605397904 +0000 UTC m=+34.129578338 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 26 16:47:39 crc kubenswrapper[4754]: E0126 16:47:39.605592 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-26 16:47:47.605517797 +0000 UTC m=+34.129698401 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 26 16:47:39 crc kubenswrapper[4754]: E0126 16:47:39.605631 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-26 16:47:47.605616909 +0000 UTC m=+34.129797343 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Jan 26 16:47:39 crc kubenswrapper[4754]: I0126 16:47:39.682005 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:39 crc kubenswrapper[4754]: I0126 16:47:39.682047 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:39 crc kubenswrapper[4754]: I0126 16:47:39.682055 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:39 crc kubenswrapper[4754]: I0126 16:47:39.682070 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:39 crc kubenswrapper[4754]: I0126 16:47:39.682081 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:39Z","lastTransitionTime":"2026-01-26T16:47:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:39 crc kubenswrapper[4754]: I0126 16:47:39.724512 4754 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-05 02:17:04.123067126 +0000 UTC Jan 26 16:47:39 crc kubenswrapper[4754]: I0126 16:47:39.766919 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 16:47:39 crc kubenswrapper[4754]: I0126 16:47:39.767037 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 16:47:39 crc kubenswrapper[4754]: E0126 16:47:39.767110 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 16:47:39 crc kubenswrapper[4754]: I0126 16:47:39.767211 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 16:47:39 crc kubenswrapper[4754]: E0126 16:47:39.767403 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 16:47:39 crc kubenswrapper[4754]: E0126 16:47:39.767532 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 16:47:39 crc kubenswrapper[4754]: I0126 16:47:39.785252 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:39 crc kubenswrapper[4754]: I0126 16:47:39.785310 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:39 crc kubenswrapper[4754]: I0126 16:47:39.785325 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:39 crc kubenswrapper[4754]: I0126 16:47:39.785348 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:39 crc kubenswrapper[4754]: I0126 16:47:39.785368 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:39Z","lastTransitionTime":"2026-01-26T16:47:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:39 crc kubenswrapper[4754]: I0126 16:47:39.888775 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:39 crc kubenswrapper[4754]: I0126 16:47:39.888828 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:39 crc kubenswrapper[4754]: I0126 16:47:39.888839 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:39 crc kubenswrapper[4754]: I0126 16:47:39.888861 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:39 crc kubenswrapper[4754]: I0126 16:47:39.888874 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:39Z","lastTransitionTime":"2026-01-26T16:47:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:39 crc kubenswrapper[4754]: I0126 16:47:39.973368 4754 generic.go:334] "Generic (PLEG): container finished" podID="351cc6c0-b8e5-4589-ae63-f98e8ab1eca3" containerID="63a9b9317494d0e0b33970ef3d461729b10d2a1ad7d3d2bd33f389b184ca56ac" exitCode=0 Jan 26 16:47:39 crc kubenswrapper[4754]: I0126 16:47:39.973445 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-bwpd6" event={"ID":"351cc6c0-b8e5-4589-ae63-f98e8ab1eca3","Type":"ContainerDied","Data":"63a9b9317494d0e0b33970ef3d461729b10d2a1ad7d3d2bd33f389b184ca56ac"} Jan 26 16:47:39 crc kubenswrapper[4754]: I0126 16:47:39.991742 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:39 crc kubenswrapper[4754]: I0126 16:47:39.991781 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:39 crc kubenswrapper[4754]: I0126 16:47:39.991793 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:39 crc kubenswrapper[4754]: I0126 16:47:39.991809 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:39 crc kubenswrapper[4754]: I0126 16:47:39.991818 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:39Z","lastTransitionTime":"2026-01-26T16:47:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:39 crc kubenswrapper[4754]: I0126 16:47:39.997569 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:39Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:40 crc kubenswrapper[4754]: I0126 16:47:40.015020 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e459fd8c1093d6f3d0a671751b792b26b7df94c21ced1be1a2250821bcc82d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:40Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:40 crc kubenswrapper[4754]: I0126 16:47:40.029552 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2blzw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca65468e-30c6-4666-962e-cc0de05e67fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4603235bb2af495901926b805fe1c0ffc48824a4394ccf928e7fcd53a673ac3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c25z4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:33Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2blzw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:40Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:40 crc kubenswrapper[4754]: I0126 16:47:40.046457 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bwpd6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"351cc6c0-b8e5-4589-ae63-f98e8ab1eca3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3ce0fe080a094316b9456057e347b90a5ff432daa12b9c56d7aea9bee1f72b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3ce0fe080a094316b9456057e347b90a5ff432daa12b9c56d7aea9bee1f72b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1772bb41eb8a36a863d3c691610bcbee5d16b260480b3e9f9d6b083e09e42799\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1772bb41eb8a36a863d3c691610bcbee5d16b260480b3e9f9d6b083e09e42799\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24848d69756024cc664affb9492764ace0638ef611f910f4f2bf86b638cc443d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://24848d69756024cc664affb9492764ace0638ef611f910f4f2bf86b638cc443d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c264fc7c3300a725566acce4464f01d9314779737624401c164d867831fd93bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c264fc7c3300a725566acce4464f01d9314779737624401c164d867831fd93bf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5acfe327eee4aad927ac3a32d24549eff788b14a1c4bdac359b6b56301d884ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5acfe327eee4aad927ac3a32d24549eff788b14a1c4bdac359b6b56301d884ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://63a9b9317494d0e0b33970ef3d461729b10d2a1ad7d3d2bd33f389b184ca56ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://63a9b9317494d0e0b33970ef3d461729b10d2a1ad7d3d2bd33f389b184ca56ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bwpd6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:40Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:40 crc kubenswrapper[4754]: I0126 16:47:40.067531 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jsbxt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"689569c4-93ee-4b82-93fc-9af0f97e6dc3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d9a7bafabd44f8b88a6f62cbe45599af8563bc45d9d7ad3a7669f3d8d11eb9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d9a7bafabd44f8b88a6f62cbe45599af8563bc45d9d7ad3a7669f3d8d11eb9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jsbxt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:40Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:40 crc kubenswrapper[4754]: I0126 16:47:40.085106 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:40Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:40 crc kubenswrapper[4754]: I0126 16:47:40.097322 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:40 crc kubenswrapper[4754]: I0126 16:47:40.097384 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:40 crc kubenswrapper[4754]: I0126 16:47:40.097396 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:40 crc kubenswrapper[4754]: I0126 16:47:40.097416 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:40 crc kubenswrapper[4754]: I0126 16:47:40.097428 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:40Z","lastTransitionTime":"2026-01-26T16:47:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:40 crc kubenswrapper[4754]: I0126 16:47:40.101718 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89251fa9c64bc25d6d6f355400d7644c099f22aed1cae812fc991a03595f1e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e90508d8a53be9146f9b388750730fdae9ee200306caf5ad393a356390a69ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:40Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:40 crc kubenswrapper[4754]: I0126 16:47:40.111944 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-84lgq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6763d962-0892-4acc-bab9-7a4e93e39a85\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://953c1917fca896c3def3fc2c69f9e3dbdc9ca26fc2b3d40282be71f71759fd68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8zlw6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:33Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-84lgq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:40Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:40 crc kubenswrapper[4754]: I0126 16:47:40.124442 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-skh5n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b619d233-b592-4b05-a0b7-dc094c88471e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9231a4191565e2e0ef53d8a394bb34615c900b601a0dfec24c2d8e5fa7b0b659\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-skh5n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:40Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:40 crc kubenswrapper[4754]: I0126 16:47:40.136995 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"099af702-abf2-4933-b4c9-63ef2326e468\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23017e73f8450839b4c2c02fa61235ffe4604bdce11556def27f25bcadcf89dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a74507c941b9de4d0401c8f1d53bb79ed11359628d3405cdab7f5474a75c862e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4402182b3e2e35a7d5dd4c44ac53e152915b312b7ccb076c4f2b41b925a6e7f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f772a003566198519fe1849fd14f29d4346004b2e655810662507c95b4d5ffa1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:13Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:40Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:40 crc kubenswrapper[4754]: I0126 16:47:40.149983 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://205ec256e46a8e08b4b0b0805fb6c5f4966fc3ccef0738df4b97cb404e74da17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:40Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:40 crc kubenswrapper[4754]: I0126 16:47:40.164102 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:40Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:40 crc kubenswrapper[4754]: I0126 16:47:40.177252 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c3718a4-f354-4284-92e0-fdfb45a692bd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b86f46f49163bca7eca9458279029af330dbd4890b03325e06fa3b1bdf030de6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxjcz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fffc03b536bfee3c7bc1bddd500e834e34b391896477a1501e3e06201a19374\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxjcz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-x65wv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:40Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:40 crc kubenswrapper[4754]: I0126 16:47:40.191730 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b89aa081-0a2e-4757-9210-e2a6c228bee9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9c5f40188fda14b14128514fe84bddfe8dcede3d6c560283ccb2ac0c66895fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f45445cbf1d180d6d4dd9588e85bf678f043bfce8cff0812b78ca3510904f9a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://085fa7b8d4fef8b2722fac8c346107699b5a6bd33fbfc74ed35831cb8fdf9dad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c40e018a0979035827963efb9efc48a878a55b9df68ec1a9242d7cff64fe480\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8356f76fbc3a4e636c12d1da5455d3f6767d9b83ce2f2925a11435d7d7261e9f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0126 16:47:26.205208 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0126 16:47:26.206538 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3645934156/tls.crt::/tmp/serving-cert-3645934156/tls.key\\\\\\\"\\\\nI0126 16:47:31.504518 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0126 16:47:31.509366 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0126 16:47:31.509400 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0126 16:47:31.509428 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0126 16:47:31.509436 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0126 16:47:31.515912 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0126 16:47:31.516036 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 16:47:31.516068 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 16:47:31.516096 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0126 16:47:31.516121 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0126 16:47:31.516145 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0126 16:47:31.516170 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0126 16:47:31.516412 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0126 16:47:31.519294 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27b74e2e65a547e0e28fddd4f15feacb17f4d57596eeb8da82df9d39243c85e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22ba7649a1308815bfad52016000987a7a9a5bc214d7ff61e137ba099b20fc6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://22ba7649a1308815bfad52016000987a7a9a5bc214d7ff61e137ba099b20fc6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:13Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:40Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:40 crc kubenswrapper[4754]: I0126 16:47:40.200247 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:40 crc kubenswrapper[4754]: I0126 16:47:40.200285 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:40 crc kubenswrapper[4754]: I0126 16:47:40.200302 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:40 crc kubenswrapper[4754]: I0126 16:47:40.200322 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:40 crc kubenswrapper[4754]: I0126 16:47:40.200333 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:40Z","lastTransitionTime":"2026-01-26T16:47:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:40 crc kubenswrapper[4754]: I0126 16:47:40.303870 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:40 crc kubenswrapper[4754]: I0126 16:47:40.303930 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:40 crc kubenswrapper[4754]: I0126 16:47:40.303941 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:40 crc kubenswrapper[4754]: I0126 16:47:40.303964 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:40 crc kubenswrapper[4754]: I0126 16:47:40.303978 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:40Z","lastTransitionTime":"2026-01-26T16:47:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:40 crc kubenswrapper[4754]: I0126 16:47:40.406071 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:40 crc kubenswrapper[4754]: I0126 16:47:40.406127 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:40 crc kubenswrapper[4754]: I0126 16:47:40.406334 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:40 crc kubenswrapper[4754]: I0126 16:47:40.406361 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:40 crc kubenswrapper[4754]: I0126 16:47:40.406375 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:40Z","lastTransitionTime":"2026-01-26T16:47:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:40 crc kubenswrapper[4754]: I0126 16:47:40.508709 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:40 crc kubenswrapper[4754]: I0126 16:47:40.508748 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:40 crc kubenswrapper[4754]: I0126 16:47:40.508760 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:40 crc kubenswrapper[4754]: I0126 16:47:40.508804 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:40 crc kubenswrapper[4754]: I0126 16:47:40.508815 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:40Z","lastTransitionTime":"2026-01-26T16:47:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:40 crc kubenswrapper[4754]: I0126 16:47:40.610997 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:40 crc kubenswrapper[4754]: I0126 16:47:40.611061 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:40 crc kubenswrapper[4754]: I0126 16:47:40.611071 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:40 crc kubenswrapper[4754]: I0126 16:47:40.611090 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:40 crc kubenswrapper[4754]: I0126 16:47:40.611101 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:40Z","lastTransitionTime":"2026-01-26T16:47:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:40 crc kubenswrapper[4754]: I0126 16:47:40.714852 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:40 crc kubenswrapper[4754]: I0126 16:47:40.714928 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:40 crc kubenswrapper[4754]: I0126 16:47:40.714940 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:40 crc kubenswrapper[4754]: I0126 16:47:40.714957 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:40 crc kubenswrapper[4754]: I0126 16:47:40.714969 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:40Z","lastTransitionTime":"2026-01-26T16:47:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:40 crc kubenswrapper[4754]: I0126 16:47:40.725822 4754 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-19 17:30:54.188240103 +0000 UTC Jan 26 16:47:40 crc kubenswrapper[4754]: I0126 16:47:40.817969 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:40 crc kubenswrapper[4754]: I0126 16:47:40.818040 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:40 crc kubenswrapper[4754]: I0126 16:47:40.818053 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:40 crc kubenswrapper[4754]: I0126 16:47:40.818074 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:40 crc kubenswrapper[4754]: I0126 16:47:40.818087 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:40Z","lastTransitionTime":"2026-01-26T16:47:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:40 crc kubenswrapper[4754]: I0126 16:47:40.921314 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:40 crc kubenswrapper[4754]: I0126 16:47:40.921373 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:40 crc kubenswrapper[4754]: I0126 16:47:40.921386 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:40 crc kubenswrapper[4754]: I0126 16:47:40.921408 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:40 crc kubenswrapper[4754]: I0126 16:47:40.921420 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:40Z","lastTransitionTime":"2026-01-26T16:47:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:40 crc kubenswrapper[4754]: I0126 16:47:40.981977 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-bwpd6" event={"ID":"351cc6c0-b8e5-4589-ae63-f98e8ab1eca3","Type":"ContainerStarted","Data":"1022f101d36873b41bbc1ed8a4a2590347b49f7cce6a172aa1d5bc9f529aed65"} Jan 26 16:47:40 crc kubenswrapper[4754]: I0126 16:47:40.993556 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jsbxt" event={"ID":"689569c4-93ee-4b82-93fc-9af0f97e6dc3","Type":"ContainerStarted","Data":"481587b671ea36286ee7fab4da74db0b8fbcab26870e39ab02571e4bdcb46c3b"} Jan 26 16:47:40 crc kubenswrapper[4754]: I0126 16:47:40.994558 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-jsbxt" Jan 26 16:47:40 crc kubenswrapper[4754]: I0126 16:47:40.994587 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-jsbxt" Jan 26 16:47:40 crc kubenswrapper[4754]: I0126 16:47:40.994599 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-jsbxt" Jan 26 16:47:41 crc kubenswrapper[4754]: I0126 16:47:40.998907 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b89aa081-0a2e-4757-9210-e2a6c228bee9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9c5f40188fda14b14128514fe84bddfe8dcede3d6c560283ccb2ac0c66895fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f45445cbf1d180d6d4dd9588e85bf678f043bfce8cff0812b78ca3510904f9a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://085fa7b8d4fef8b2722fac8c346107699b5a6bd33fbfc74ed35831cb8fdf9dad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c40e018a0979035827963efb9efc48a878a55b9df68ec1a9242d7cff64fe480\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8356f76fbc3a4e636c12d1da5455d3f6767d9b83ce2f2925a11435d7d7261e9f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0126 16:47:26.205208 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0126 16:47:26.206538 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3645934156/tls.crt::/tmp/serving-cert-3645934156/tls.key\\\\\\\"\\\\nI0126 16:47:31.504518 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0126 16:47:31.509366 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0126 16:47:31.509400 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0126 16:47:31.509428 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0126 16:47:31.509436 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0126 16:47:31.515912 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0126 16:47:31.516036 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 16:47:31.516068 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 16:47:31.516096 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0126 16:47:31.516121 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0126 16:47:31.516145 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0126 16:47:31.516170 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0126 16:47:31.516412 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0126 16:47:31.519294 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27b74e2e65a547e0e28fddd4f15feacb17f4d57596eeb8da82df9d39243c85e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22ba7649a1308815bfad52016000987a7a9a5bc214d7ff61e137ba099b20fc6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://22ba7649a1308815bfad52016000987a7a9a5bc214d7ff61e137ba099b20fc6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:13Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:40Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:41 crc kubenswrapper[4754]: I0126 16:47:41.015500 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://205ec256e46a8e08b4b0b0805fb6c5f4966fc3ccef0738df4b97cb404e74da17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:41Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:41 crc kubenswrapper[4754]: I0126 16:47:41.024444 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-jsbxt" Jan 26 16:47:41 crc kubenswrapper[4754]: I0126 16:47:41.025335 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-jsbxt" Jan 26 16:47:41 crc kubenswrapper[4754]: I0126 16:47:41.025378 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:41 crc kubenswrapper[4754]: I0126 16:47:41.025550 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:41 crc kubenswrapper[4754]: I0126 16:47:41.025564 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:41 crc kubenswrapper[4754]: I0126 16:47:41.025586 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:41 crc kubenswrapper[4754]: I0126 16:47:41.025600 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:41Z","lastTransitionTime":"2026-01-26T16:47:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:41 crc kubenswrapper[4754]: I0126 16:47:41.033590 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:41Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:41 crc kubenswrapper[4754]: I0126 16:47:41.049120 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c3718a4-f354-4284-92e0-fdfb45a692bd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b86f46f49163bca7eca9458279029af330dbd4890b03325e06fa3b1bdf030de6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxjcz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fffc03b536bfee3c7bc1bddd500e834e34b391896477a1501e3e06201a19374\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxjcz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-x65wv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:41Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:41 crc kubenswrapper[4754]: I0126 16:47:41.067829 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:41Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:41 crc kubenswrapper[4754]: I0126 16:47:41.082802 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e459fd8c1093d6f3d0a671751b792b26b7df94c21ced1be1a2250821bcc82d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:41Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:41 crc kubenswrapper[4754]: I0126 16:47:41.101925 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2blzw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca65468e-30c6-4666-962e-cc0de05e67fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4603235bb2af495901926b805fe1c0ffc48824a4394ccf928e7fcd53a673ac3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c25z4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:33Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2blzw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:41Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:41 crc kubenswrapper[4754]: I0126 16:47:41.124851 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bwpd6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"351cc6c0-b8e5-4589-ae63-f98e8ab1eca3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1022f101d36873b41bbc1ed8a4a2590347b49f7cce6a172aa1d5bc9f529aed65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3ce0fe080a094316b9456057e347b90a5ff432daa12b9c56d7aea9bee1f72b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3ce0fe080a094316b9456057e347b90a5ff432daa12b9c56d7aea9bee1f72b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1772bb41eb8a36a863d3c691610bcbee5d16b260480b3e9f9d6b083e09e42799\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1772bb41eb8a36a863d3c691610bcbee5d16b260480b3e9f9d6b083e09e42799\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24848d69756024cc664affb9492764ace0638ef611f910f4f2bf86b638cc443d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://24848d69756024cc664affb9492764ace0638ef611f910f4f2bf86b638cc443d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c264fc7c3300a725566acce4464f01d9314779737624401c164d867831fd93bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c264fc7c3300a725566acce4464f01d9314779737624401c164d867831fd93bf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5acfe327eee4aad927ac3a32d24549eff788b14a1c4bdac359b6b56301d884ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5acfe327eee4aad927ac3a32d24549eff788b14a1c4bdac359b6b56301d884ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://63a9b9317494d0e0b33970ef3d461729b10d2a1ad7d3d2bd33f389b184ca56ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://63a9b9317494d0e0b33970ef3d461729b10d2a1ad7d3d2bd33f389b184ca56ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bwpd6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:41Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:41 crc kubenswrapper[4754]: I0126 16:47:41.129708 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:41 crc kubenswrapper[4754]: I0126 16:47:41.129753 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:41 crc kubenswrapper[4754]: I0126 16:47:41.129763 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:41 crc kubenswrapper[4754]: I0126 16:47:41.129791 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:41 crc kubenswrapper[4754]: I0126 16:47:41.129803 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:41Z","lastTransitionTime":"2026-01-26T16:47:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:41 crc kubenswrapper[4754]: I0126 16:47:41.150462 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jsbxt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"689569c4-93ee-4b82-93fc-9af0f97e6dc3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d9a7bafabd44f8b88a6f62cbe45599af8563bc45d9d7ad3a7669f3d8d11eb9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d9a7bafabd44f8b88a6f62cbe45599af8563bc45d9d7ad3a7669f3d8d11eb9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jsbxt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:41Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:41 crc kubenswrapper[4754]: I0126 16:47:41.171321 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"099af702-abf2-4933-b4c9-63ef2326e468\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23017e73f8450839b4c2c02fa61235ffe4604bdce11556def27f25bcadcf89dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a74507c941b9de4d0401c8f1d53bb79ed11359628d3405cdab7f5474a75c862e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4402182b3e2e35a7d5dd4c44ac53e152915b312b7ccb076c4f2b41b925a6e7f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f772a003566198519fe1849fd14f29d4346004b2e655810662507c95b4d5ffa1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:13Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:41Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:41 crc kubenswrapper[4754]: I0126 16:47:41.190103 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:41Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:41 crc kubenswrapper[4754]: I0126 16:47:41.209831 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89251fa9c64bc25d6d6f355400d7644c099f22aed1cae812fc991a03595f1e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e90508d8a53be9146f9b388750730fdae9ee200306caf5ad393a356390a69ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:41Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:41 crc kubenswrapper[4754]: I0126 16:47:41.225242 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-84lgq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6763d962-0892-4acc-bab9-7a4e93e39a85\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://953c1917fca896c3def3fc2c69f9e3dbdc9ca26fc2b3d40282be71f71759fd68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8zlw6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:33Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-84lgq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:41Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:41 crc kubenswrapper[4754]: I0126 16:47:41.232435 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:41 crc kubenswrapper[4754]: I0126 16:47:41.232465 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:41 crc kubenswrapper[4754]: I0126 16:47:41.232473 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:41 crc kubenswrapper[4754]: I0126 16:47:41.232485 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:41 crc kubenswrapper[4754]: I0126 16:47:41.232494 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:41Z","lastTransitionTime":"2026-01-26T16:47:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:41 crc kubenswrapper[4754]: I0126 16:47:41.244352 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-skh5n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b619d233-b592-4b05-a0b7-dc094c88471e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9231a4191565e2e0ef53d8a394bb34615c900b601a0dfec24c2d8e5fa7b0b659\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-skh5n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:41Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:41 crc kubenswrapper[4754]: I0126 16:47:41.263442 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:41Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:41 crc kubenswrapper[4754]: I0126 16:47:41.277836 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c3718a4-f354-4284-92e0-fdfb45a692bd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b86f46f49163bca7eca9458279029af330dbd4890b03325e06fa3b1bdf030de6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxjcz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fffc03b536bfee3c7bc1bddd500e834e34b391896477a1501e3e06201a19374\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxjcz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-x65wv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:41Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:41 crc kubenswrapper[4754]: I0126 16:47:41.294010 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b89aa081-0a2e-4757-9210-e2a6c228bee9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9c5f40188fda14b14128514fe84bddfe8dcede3d6c560283ccb2ac0c66895fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f45445cbf1d180d6d4dd9588e85bf678f043bfce8cff0812b78ca3510904f9a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://085fa7b8d4fef8b2722fac8c346107699b5a6bd33fbfc74ed35831cb8fdf9dad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c40e018a0979035827963efb9efc48a878a55b9df68ec1a9242d7cff64fe480\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8356f76fbc3a4e636c12d1da5455d3f6767d9b83ce2f2925a11435d7d7261e9f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0126 16:47:26.205208 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0126 16:47:26.206538 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3645934156/tls.crt::/tmp/serving-cert-3645934156/tls.key\\\\\\\"\\\\nI0126 16:47:31.504518 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0126 16:47:31.509366 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0126 16:47:31.509400 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0126 16:47:31.509428 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0126 16:47:31.509436 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0126 16:47:31.515912 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0126 16:47:31.516036 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 16:47:31.516068 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 16:47:31.516096 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0126 16:47:31.516121 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0126 16:47:31.516145 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0126 16:47:31.516170 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0126 16:47:31.516412 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0126 16:47:31.519294 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27b74e2e65a547e0e28fddd4f15feacb17f4d57596eeb8da82df9d39243c85e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22ba7649a1308815bfad52016000987a7a9a5bc214d7ff61e137ba099b20fc6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://22ba7649a1308815bfad52016000987a7a9a5bc214d7ff61e137ba099b20fc6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:13Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:41Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:41 crc kubenswrapper[4754]: I0126 16:47:41.309269 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://205ec256e46a8e08b4b0b0805fb6c5f4966fc3ccef0738df4b97cb404e74da17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:41Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:41 crc kubenswrapper[4754]: I0126 16:47:41.325846 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e459fd8c1093d6f3d0a671751b792b26b7df94c21ced1be1a2250821bcc82d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:41Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:41 crc kubenswrapper[4754]: I0126 16:47:41.334351 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:41 crc kubenswrapper[4754]: I0126 16:47:41.334400 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:41 crc kubenswrapper[4754]: I0126 16:47:41.334410 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:41 crc kubenswrapper[4754]: I0126 16:47:41.334428 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:41 crc kubenswrapper[4754]: I0126 16:47:41.334439 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:41Z","lastTransitionTime":"2026-01-26T16:47:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:41 crc kubenswrapper[4754]: I0126 16:47:41.337095 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2blzw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca65468e-30c6-4666-962e-cc0de05e67fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4603235bb2af495901926b805fe1c0ffc48824a4394ccf928e7fcd53a673ac3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c25z4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:33Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2blzw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:41Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:41 crc kubenswrapper[4754]: I0126 16:47:41.350097 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bwpd6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"351cc6c0-b8e5-4589-ae63-f98e8ab1eca3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1022f101d36873b41bbc1ed8a4a2590347b49f7cce6a172aa1d5bc9f529aed65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3ce0fe080a094316b9456057e347b90a5ff432daa12b9c56d7aea9bee1f72b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3ce0fe080a094316b9456057e347b90a5ff432daa12b9c56d7aea9bee1f72b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1772bb41eb8a36a863d3c691610bcbee5d16b260480b3e9f9d6b083e09e42799\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1772bb41eb8a36a863d3c691610bcbee5d16b260480b3e9f9d6b083e09e42799\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24848d69756024cc664affb9492764ace0638ef611f910f4f2bf86b638cc443d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://24848d69756024cc664affb9492764ace0638ef611f910f4f2bf86b638cc443d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c264fc7c3300a725566acce4464f01d9314779737624401c164d867831fd93bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c264fc7c3300a725566acce4464f01d9314779737624401c164d867831fd93bf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5acfe327eee4aad927ac3a32d24549eff788b14a1c4bdac359b6b56301d884ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5acfe327eee4aad927ac3a32d24549eff788b14a1c4bdac359b6b56301d884ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://63a9b9317494d0e0b33970ef3d461729b10d2a1ad7d3d2bd33f389b184ca56ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://63a9b9317494d0e0b33970ef3d461729b10d2a1ad7d3d2bd33f389b184ca56ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bwpd6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:41Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:41 crc kubenswrapper[4754]: I0126 16:47:41.369536 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jsbxt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"689569c4-93ee-4b82-93fc-9af0f97e6dc3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://033f8b104d4e9ff6d5f87ca998665f12e0c56696b3d30e01ede05f4da540365d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://919572e731813a98c7d8a94879bba931f75c472c7cd681e86d5197ebab43815d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1ac1f49bd564750fcff530a0f76657b0ed41573539acd020a922d19861ef834\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a300a80ad87f46ba1b4fd7ad3cf6c25767fcfe966b93bdd86d729843f53a5cda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31dcab2f9a2d218b0e49dad1df46377ff4521592824a0fbac254d5b8e6d0632e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fe4d9395447ac570eca1338e749dbefea76295d427a12487b63a142a7d6a2e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://481587b671ea36286ee7fab4da74db0b8fbcab26870e39ab02571e4bdcb46c3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://07a43ff7860ff65faf51f4a5e8aeb326f06d6d7222677a92560ae2e3a0c81484\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d9a7bafabd44f8b88a6f62cbe45599af8563bc45d9d7ad3a7669f3d8d11eb9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d9a7bafabd44f8b88a6f62cbe45599af8563bc45d9d7ad3a7669f3d8d11eb9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jsbxt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:41Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:41 crc kubenswrapper[4754]: I0126 16:47:41.385874 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:41Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:41 crc kubenswrapper[4754]: I0126 16:47:41.398560 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:41Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:41 crc kubenswrapper[4754]: I0126 16:47:41.411716 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89251fa9c64bc25d6d6f355400d7644c099f22aed1cae812fc991a03595f1e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e90508d8a53be9146f9b388750730fdae9ee200306caf5ad393a356390a69ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:41Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:41 crc kubenswrapper[4754]: I0126 16:47:41.431706 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-84lgq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6763d962-0892-4acc-bab9-7a4e93e39a85\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://953c1917fca896c3def3fc2c69f9e3dbdc9ca26fc2b3d40282be71f71759fd68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8zlw6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:33Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-84lgq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:41Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:41 crc kubenswrapper[4754]: I0126 16:47:41.437848 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:41 crc kubenswrapper[4754]: I0126 16:47:41.437904 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:41 crc kubenswrapper[4754]: I0126 16:47:41.437915 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:41 crc kubenswrapper[4754]: I0126 16:47:41.437934 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:41 crc kubenswrapper[4754]: I0126 16:47:41.437946 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:41Z","lastTransitionTime":"2026-01-26T16:47:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:41 crc kubenswrapper[4754]: I0126 16:47:41.447929 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-skh5n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b619d233-b592-4b05-a0b7-dc094c88471e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9231a4191565e2e0ef53d8a394bb34615c900b601a0dfec24c2d8e5fa7b0b659\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-skh5n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:41Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:41 crc kubenswrapper[4754]: I0126 16:47:41.464231 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"099af702-abf2-4933-b4c9-63ef2326e468\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23017e73f8450839b4c2c02fa61235ffe4604bdce11556def27f25bcadcf89dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a74507c941b9de4d0401c8f1d53bb79ed11359628d3405cdab7f5474a75c862e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4402182b3e2e35a7d5dd4c44ac53e152915b312b7ccb076c4f2b41b925a6e7f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f772a003566198519fe1849fd14f29d4346004b2e655810662507c95b4d5ffa1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:13Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:41Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:41 crc kubenswrapper[4754]: I0126 16:47:41.540770 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:41 crc kubenswrapper[4754]: I0126 16:47:41.540813 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:41 crc kubenswrapper[4754]: I0126 16:47:41.540822 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:41 crc kubenswrapper[4754]: I0126 16:47:41.540836 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:41 crc kubenswrapper[4754]: I0126 16:47:41.540850 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:41Z","lastTransitionTime":"2026-01-26T16:47:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:41 crc kubenswrapper[4754]: I0126 16:47:41.642894 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:41 crc kubenswrapper[4754]: I0126 16:47:41.642924 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:41 crc kubenswrapper[4754]: I0126 16:47:41.642933 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:41 crc kubenswrapper[4754]: I0126 16:47:41.642945 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:41 crc kubenswrapper[4754]: I0126 16:47:41.642954 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:41Z","lastTransitionTime":"2026-01-26T16:47:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:41 crc kubenswrapper[4754]: I0126 16:47:41.726825 4754 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-04 01:23:37.939272501 +0000 UTC Jan 26 16:47:41 crc kubenswrapper[4754]: I0126 16:47:41.746260 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:41 crc kubenswrapper[4754]: I0126 16:47:41.746596 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:41 crc kubenswrapper[4754]: I0126 16:47:41.746617 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:41 crc kubenswrapper[4754]: I0126 16:47:41.746647 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:41 crc kubenswrapper[4754]: I0126 16:47:41.746690 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:41Z","lastTransitionTime":"2026-01-26T16:47:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:41 crc kubenswrapper[4754]: I0126 16:47:41.767205 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 16:47:41 crc kubenswrapper[4754]: I0126 16:47:41.767336 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 16:47:41 crc kubenswrapper[4754]: E0126 16:47:41.767402 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 16:47:41 crc kubenswrapper[4754]: I0126 16:47:41.767498 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 16:47:41 crc kubenswrapper[4754]: E0126 16:47:41.767539 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 16:47:41 crc kubenswrapper[4754]: E0126 16:47:41.767643 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 16:47:41 crc kubenswrapper[4754]: I0126 16:47:41.849199 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:41 crc kubenswrapper[4754]: I0126 16:47:41.849239 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:41 crc kubenswrapper[4754]: I0126 16:47:41.849249 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:41 crc kubenswrapper[4754]: I0126 16:47:41.849264 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:41 crc kubenswrapper[4754]: I0126 16:47:41.849274 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:41Z","lastTransitionTime":"2026-01-26T16:47:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:41 crc kubenswrapper[4754]: I0126 16:47:41.951978 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:41 crc kubenswrapper[4754]: I0126 16:47:41.952029 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:41 crc kubenswrapper[4754]: I0126 16:47:41.952046 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:41 crc kubenswrapper[4754]: I0126 16:47:41.952069 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:41 crc kubenswrapper[4754]: I0126 16:47:41.952081 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:41Z","lastTransitionTime":"2026-01-26T16:47:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:42 crc kubenswrapper[4754]: I0126 16:47:42.054374 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:42 crc kubenswrapper[4754]: I0126 16:47:42.054435 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:42 crc kubenswrapper[4754]: I0126 16:47:42.054446 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:42 crc kubenswrapper[4754]: I0126 16:47:42.054475 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:42 crc kubenswrapper[4754]: I0126 16:47:42.054493 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:42Z","lastTransitionTime":"2026-01-26T16:47:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:42 crc kubenswrapper[4754]: I0126 16:47:42.157714 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:42 crc kubenswrapper[4754]: I0126 16:47:42.157774 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:42 crc kubenswrapper[4754]: I0126 16:47:42.157784 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:42 crc kubenswrapper[4754]: I0126 16:47:42.157800 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:42 crc kubenswrapper[4754]: I0126 16:47:42.157812 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:42Z","lastTransitionTime":"2026-01-26T16:47:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:42 crc kubenswrapper[4754]: I0126 16:47:42.260649 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:42 crc kubenswrapper[4754]: I0126 16:47:42.260765 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:42 crc kubenswrapper[4754]: I0126 16:47:42.260782 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:42 crc kubenswrapper[4754]: I0126 16:47:42.260809 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:42 crc kubenswrapper[4754]: I0126 16:47:42.260828 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:42Z","lastTransitionTime":"2026-01-26T16:47:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:42 crc kubenswrapper[4754]: I0126 16:47:42.363410 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:42 crc kubenswrapper[4754]: I0126 16:47:42.363474 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:42 crc kubenswrapper[4754]: I0126 16:47:42.363486 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:42 crc kubenswrapper[4754]: I0126 16:47:42.363505 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:42 crc kubenswrapper[4754]: I0126 16:47:42.363520 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:42Z","lastTransitionTime":"2026-01-26T16:47:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:42 crc kubenswrapper[4754]: I0126 16:47:42.466571 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:42 crc kubenswrapper[4754]: I0126 16:47:42.466629 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:42 crc kubenswrapper[4754]: I0126 16:47:42.466641 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:42 crc kubenswrapper[4754]: I0126 16:47:42.466660 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:42 crc kubenswrapper[4754]: I0126 16:47:42.466690 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:42Z","lastTransitionTime":"2026-01-26T16:47:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:42 crc kubenswrapper[4754]: I0126 16:47:42.569563 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:42 crc kubenswrapper[4754]: I0126 16:47:42.569634 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:42 crc kubenswrapper[4754]: I0126 16:47:42.569646 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:42 crc kubenswrapper[4754]: I0126 16:47:42.569682 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:42 crc kubenswrapper[4754]: I0126 16:47:42.569694 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:42Z","lastTransitionTime":"2026-01-26T16:47:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:42 crc kubenswrapper[4754]: I0126 16:47:42.672703 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:42 crc kubenswrapper[4754]: I0126 16:47:42.672751 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:42 crc kubenswrapper[4754]: I0126 16:47:42.672760 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:42 crc kubenswrapper[4754]: I0126 16:47:42.672777 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:42 crc kubenswrapper[4754]: I0126 16:47:42.672789 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:42Z","lastTransitionTime":"2026-01-26T16:47:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:42 crc kubenswrapper[4754]: I0126 16:47:42.727759 4754 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-29 11:43:05.535594831 +0000 UTC Jan 26 16:47:42 crc kubenswrapper[4754]: I0126 16:47:42.776630 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:42 crc kubenswrapper[4754]: I0126 16:47:42.776708 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:42 crc kubenswrapper[4754]: I0126 16:47:42.776721 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:42 crc kubenswrapper[4754]: I0126 16:47:42.776735 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:42 crc kubenswrapper[4754]: I0126 16:47:42.776746 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:42Z","lastTransitionTime":"2026-01-26T16:47:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:42 crc kubenswrapper[4754]: I0126 16:47:42.879637 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:42 crc kubenswrapper[4754]: I0126 16:47:42.879682 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:42 crc kubenswrapper[4754]: I0126 16:47:42.879692 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:42 crc kubenswrapper[4754]: I0126 16:47:42.879704 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:42 crc kubenswrapper[4754]: I0126 16:47:42.879713 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:42Z","lastTransitionTime":"2026-01-26T16:47:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:42 crc kubenswrapper[4754]: I0126 16:47:42.982251 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:42 crc kubenswrapper[4754]: I0126 16:47:42.982282 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:42 crc kubenswrapper[4754]: I0126 16:47:42.982290 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:42 crc kubenswrapper[4754]: I0126 16:47:42.982305 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:42 crc kubenswrapper[4754]: I0126 16:47:42.982314 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:42Z","lastTransitionTime":"2026-01-26T16:47:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:43 crc kubenswrapper[4754]: I0126 16:47:43.085395 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:43 crc kubenswrapper[4754]: I0126 16:47:43.085439 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:43 crc kubenswrapper[4754]: I0126 16:47:43.085450 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:43 crc kubenswrapper[4754]: I0126 16:47:43.085464 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:43 crc kubenswrapper[4754]: I0126 16:47:43.085473 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:43Z","lastTransitionTime":"2026-01-26T16:47:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:43 crc kubenswrapper[4754]: I0126 16:47:43.188113 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:43 crc kubenswrapper[4754]: I0126 16:47:43.188202 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:43 crc kubenswrapper[4754]: I0126 16:47:43.188215 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:43 crc kubenswrapper[4754]: I0126 16:47:43.188244 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:43 crc kubenswrapper[4754]: I0126 16:47:43.188262 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:43Z","lastTransitionTime":"2026-01-26T16:47:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:43 crc kubenswrapper[4754]: I0126 16:47:43.292059 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:43 crc kubenswrapper[4754]: I0126 16:47:43.292107 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:43 crc kubenswrapper[4754]: I0126 16:47:43.292118 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:43 crc kubenswrapper[4754]: I0126 16:47:43.292141 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:43 crc kubenswrapper[4754]: I0126 16:47:43.292153 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:43Z","lastTransitionTime":"2026-01-26T16:47:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:43 crc kubenswrapper[4754]: I0126 16:47:43.394630 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:43 crc kubenswrapper[4754]: I0126 16:47:43.394972 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:43 crc kubenswrapper[4754]: I0126 16:47:43.394987 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:43 crc kubenswrapper[4754]: I0126 16:47:43.395003 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:43 crc kubenswrapper[4754]: I0126 16:47:43.395014 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:43Z","lastTransitionTime":"2026-01-26T16:47:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:43 crc kubenswrapper[4754]: I0126 16:47:43.497630 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:43 crc kubenswrapper[4754]: I0126 16:47:43.497695 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:43 crc kubenswrapper[4754]: I0126 16:47:43.497710 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:43 crc kubenswrapper[4754]: I0126 16:47:43.497726 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:43 crc kubenswrapper[4754]: I0126 16:47:43.497738 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:43Z","lastTransitionTime":"2026-01-26T16:47:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:43 crc kubenswrapper[4754]: I0126 16:47:43.600151 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:43 crc kubenswrapper[4754]: I0126 16:47:43.600202 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:43 crc kubenswrapper[4754]: I0126 16:47:43.600212 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:43 crc kubenswrapper[4754]: I0126 16:47:43.600228 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:43 crc kubenswrapper[4754]: I0126 16:47:43.600238 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:43Z","lastTransitionTime":"2026-01-26T16:47:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:43 crc kubenswrapper[4754]: I0126 16:47:43.703263 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:43 crc kubenswrapper[4754]: I0126 16:47:43.703324 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:43 crc kubenswrapper[4754]: I0126 16:47:43.703334 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:43 crc kubenswrapper[4754]: I0126 16:47:43.703347 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:43 crc kubenswrapper[4754]: I0126 16:47:43.703355 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:43Z","lastTransitionTime":"2026-01-26T16:47:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:43 crc kubenswrapper[4754]: I0126 16:47:43.728772 4754 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-01 15:43:21.96647088 +0000 UTC Jan 26 16:47:43 crc kubenswrapper[4754]: I0126 16:47:43.766362 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 16:47:43 crc kubenswrapper[4754]: I0126 16:47:43.766394 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 16:47:43 crc kubenswrapper[4754]: E0126 16:47:43.766942 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 16:47:43 crc kubenswrapper[4754]: I0126 16:47:43.767000 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 16:47:43 crc kubenswrapper[4754]: E0126 16:47:43.767100 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 16:47:43 crc kubenswrapper[4754]: E0126 16:47:43.767179 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 16:47:43 crc kubenswrapper[4754]: I0126 16:47:43.783981 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b89aa081-0a2e-4757-9210-e2a6c228bee9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9c5f40188fda14b14128514fe84bddfe8dcede3d6c560283ccb2ac0c66895fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f45445cbf1d180d6d4dd9588e85bf678f043bfce8cff0812b78ca3510904f9a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://085fa7b8d4fef8b2722fac8c346107699b5a6bd33fbfc74ed35831cb8fdf9dad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c40e018a0979035827963efb9efc48a878a55b9df68ec1a9242d7cff64fe480\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8356f76fbc3a4e636c12d1da5455d3f6767d9b83ce2f2925a11435d7d7261e9f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0126 16:47:26.205208 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0126 16:47:26.206538 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3645934156/tls.crt::/tmp/serving-cert-3645934156/tls.key\\\\\\\"\\\\nI0126 16:47:31.504518 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0126 16:47:31.509366 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0126 16:47:31.509400 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0126 16:47:31.509428 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0126 16:47:31.509436 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0126 16:47:31.515912 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0126 16:47:31.516036 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 16:47:31.516068 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 16:47:31.516096 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0126 16:47:31.516121 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0126 16:47:31.516145 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0126 16:47:31.516170 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0126 16:47:31.516412 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0126 16:47:31.519294 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27b74e2e65a547e0e28fddd4f15feacb17f4d57596eeb8da82df9d39243c85e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22ba7649a1308815bfad52016000987a7a9a5bc214d7ff61e137ba099b20fc6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://22ba7649a1308815bfad52016000987a7a9a5bc214d7ff61e137ba099b20fc6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:13Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:43Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:43 crc kubenswrapper[4754]: I0126 16:47:43.798817 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://205ec256e46a8e08b4b0b0805fb6c5f4966fc3ccef0738df4b97cb404e74da17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:43Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:43 crc kubenswrapper[4754]: I0126 16:47:43.806144 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:43 crc kubenswrapper[4754]: I0126 16:47:43.806193 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:43 crc kubenswrapper[4754]: I0126 16:47:43.806204 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:43 crc kubenswrapper[4754]: I0126 16:47:43.806219 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:43 crc kubenswrapper[4754]: I0126 16:47:43.806230 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:43Z","lastTransitionTime":"2026-01-26T16:47:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:43 crc kubenswrapper[4754]: I0126 16:47:43.812464 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:43Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:43 crc kubenswrapper[4754]: I0126 16:47:43.825580 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c3718a4-f354-4284-92e0-fdfb45a692bd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b86f46f49163bca7eca9458279029af330dbd4890b03325e06fa3b1bdf030de6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxjcz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fffc03b536bfee3c7bc1bddd500e834e34b391896477a1501e3e06201a19374\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxjcz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-x65wv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:43Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:43 crc kubenswrapper[4754]: I0126 16:47:43.843165 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bwpd6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"351cc6c0-b8e5-4589-ae63-f98e8ab1eca3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1022f101d36873b41bbc1ed8a4a2590347b49f7cce6a172aa1d5bc9f529aed65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3ce0fe080a094316b9456057e347b90a5ff432daa12b9c56d7aea9bee1f72b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3ce0fe080a094316b9456057e347b90a5ff432daa12b9c56d7aea9bee1f72b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1772bb41eb8a36a863d3c691610bcbee5d16b260480b3e9f9d6b083e09e42799\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1772bb41eb8a36a863d3c691610bcbee5d16b260480b3e9f9d6b083e09e42799\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24848d69756024cc664affb9492764ace0638ef611f910f4f2bf86b638cc443d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://24848d69756024cc664affb9492764ace0638ef611f910f4f2bf86b638cc443d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c264fc7c3300a725566acce4464f01d9314779737624401c164d867831fd93bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c264fc7c3300a725566acce4464f01d9314779737624401c164d867831fd93bf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5acfe327eee4aad927ac3a32d24549eff788b14a1c4bdac359b6b56301d884ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5acfe327eee4aad927ac3a32d24549eff788b14a1c4bdac359b6b56301d884ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://63a9b9317494d0e0b33970ef3d461729b10d2a1ad7d3d2bd33f389b184ca56ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://63a9b9317494d0e0b33970ef3d461729b10d2a1ad7d3d2bd33f389b184ca56ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bwpd6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:43Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:43 crc kubenswrapper[4754]: I0126 16:47:43.862439 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jsbxt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"689569c4-93ee-4b82-93fc-9af0f97e6dc3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://033f8b104d4e9ff6d5f87ca998665f12e0c56696b3d30e01ede05f4da540365d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://919572e731813a98c7d8a94879bba931f75c472c7cd681e86d5197ebab43815d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1ac1f49bd564750fcff530a0f76657b0ed41573539acd020a922d19861ef834\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a300a80ad87f46ba1b4fd7ad3cf6c25767fcfe966b93bdd86d729843f53a5cda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31dcab2f9a2d218b0e49dad1df46377ff4521592824a0fbac254d5b8e6d0632e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fe4d9395447ac570eca1338e749dbefea76295d427a12487b63a142a7d6a2e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://481587b671ea36286ee7fab4da74db0b8fbcab26870e39ab02571e4bdcb46c3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://07a43ff7860ff65faf51f4a5e8aeb326f06d6d7222677a92560ae2e3a0c81484\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d9a7bafabd44f8b88a6f62cbe45599af8563bc45d9d7ad3a7669f3d8d11eb9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d9a7bafabd44f8b88a6f62cbe45599af8563bc45d9d7ad3a7669f3d8d11eb9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jsbxt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:43Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:43 crc kubenswrapper[4754]: I0126 16:47:43.877121 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:43Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:43 crc kubenswrapper[4754]: I0126 16:47:43.890713 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e459fd8c1093d6f3d0a671751b792b26b7df94c21ced1be1a2250821bcc82d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:43Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:43 crc kubenswrapper[4754]: I0126 16:47:43.908791 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2blzw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca65468e-30c6-4666-962e-cc0de05e67fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4603235bb2af495901926b805fe1c0ffc48824a4394ccf928e7fcd53a673ac3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c25z4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:33Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2blzw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:43Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:43 crc kubenswrapper[4754]: I0126 16:47:43.908988 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:43 crc kubenswrapper[4754]: I0126 16:47:43.909133 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:43 crc kubenswrapper[4754]: I0126 16:47:43.909145 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:43 crc kubenswrapper[4754]: I0126 16:47:43.909161 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:43 crc kubenswrapper[4754]: I0126 16:47:43.909173 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:43Z","lastTransitionTime":"2026-01-26T16:47:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:43 crc kubenswrapper[4754]: I0126 16:47:43.921469 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-84lgq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6763d962-0892-4acc-bab9-7a4e93e39a85\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://953c1917fca896c3def3fc2c69f9e3dbdc9ca26fc2b3d40282be71f71759fd68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8zlw6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:33Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-84lgq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:43Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:43 crc kubenswrapper[4754]: I0126 16:47:43.933349 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-skh5n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b619d233-b592-4b05-a0b7-dc094c88471e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9231a4191565e2e0ef53d8a394bb34615c900b601a0dfec24c2d8e5fa7b0b659\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-skh5n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:43Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:43 crc kubenswrapper[4754]: I0126 16:47:43.946894 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"099af702-abf2-4933-b4c9-63ef2326e468\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23017e73f8450839b4c2c02fa61235ffe4604bdce11556def27f25bcadcf89dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a74507c941b9de4d0401c8f1d53bb79ed11359628d3405cdab7f5474a75c862e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4402182b3e2e35a7d5dd4c44ac53e152915b312b7ccb076c4f2b41b925a6e7f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f772a003566198519fe1849fd14f29d4346004b2e655810662507c95b4d5ffa1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:13Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:43Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:43 crc kubenswrapper[4754]: I0126 16:47:43.961797 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:43Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:43 crc kubenswrapper[4754]: I0126 16:47:43.974200 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89251fa9c64bc25d6d6f355400d7644c099f22aed1cae812fc991a03595f1e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e90508d8a53be9146f9b388750730fdae9ee200306caf5ad393a356390a69ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:43Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:44 crc kubenswrapper[4754]: I0126 16:47:44.005610 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-jsbxt_689569c4-93ee-4b82-93fc-9af0f97e6dc3/ovnkube-controller/0.log" Jan 26 16:47:44 crc kubenswrapper[4754]: I0126 16:47:44.008959 4754 generic.go:334] "Generic (PLEG): container finished" podID="689569c4-93ee-4b82-93fc-9af0f97e6dc3" containerID="481587b671ea36286ee7fab4da74db0b8fbcab26870e39ab02571e4bdcb46c3b" exitCode=1 Jan 26 16:47:44 crc kubenswrapper[4754]: I0126 16:47:44.009023 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jsbxt" event={"ID":"689569c4-93ee-4b82-93fc-9af0f97e6dc3","Type":"ContainerDied","Data":"481587b671ea36286ee7fab4da74db0b8fbcab26870e39ab02571e4bdcb46c3b"} Jan 26 16:47:44 crc kubenswrapper[4754]: I0126 16:47:44.009766 4754 scope.go:117] "RemoveContainer" containerID="481587b671ea36286ee7fab4da74db0b8fbcab26870e39ab02571e4bdcb46c3b" Jan 26 16:47:44 crc kubenswrapper[4754]: I0126 16:47:44.011079 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:44 crc kubenswrapper[4754]: I0126 16:47:44.011108 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:44 crc kubenswrapper[4754]: I0126 16:47:44.011120 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:44 crc kubenswrapper[4754]: I0126 16:47:44.011140 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:44 crc kubenswrapper[4754]: I0126 16:47:44.011157 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:44Z","lastTransitionTime":"2026-01-26T16:47:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:44 crc kubenswrapper[4754]: I0126 16:47:44.026386 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b89aa081-0a2e-4757-9210-e2a6c228bee9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9c5f40188fda14b14128514fe84bddfe8dcede3d6c560283ccb2ac0c66895fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f45445cbf1d180d6d4dd9588e85bf678f043bfce8cff0812b78ca3510904f9a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://085fa7b8d4fef8b2722fac8c346107699b5a6bd33fbfc74ed35831cb8fdf9dad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c40e018a0979035827963efb9efc48a878a55b9df68ec1a9242d7cff64fe480\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8356f76fbc3a4e636c12d1da5455d3f6767d9b83ce2f2925a11435d7d7261e9f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0126 16:47:26.205208 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0126 16:47:26.206538 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3645934156/tls.crt::/tmp/serving-cert-3645934156/tls.key\\\\\\\"\\\\nI0126 16:47:31.504518 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0126 16:47:31.509366 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0126 16:47:31.509400 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0126 16:47:31.509428 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0126 16:47:31.509436 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0126 16:47:31.515912 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0126 16:47:31.516036 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 16:47:31.516068 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 16:47:31.516096 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0126 16:47:31.516121 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0126 16:47:31.516145 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0126 16:47:31.516170 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0126 16:47:31.516412 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0126 16:47:31.519294 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27b74e2e65a547e0e28fddd4f15feacb17f4d57596eeb8da82df9d39243c85e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22ba7649a1308815bfad52016000987a7a9a5bc214d7ff61e137ba099b20fc6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://22ba7649a1308815bfad52016000987a7a9a5bc214d7ff61e137ba099b20fc6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:13Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:44Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:44 crc kubenswrapper[4754]: I0126 16:47:44.040970 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://205ec256e46a8e08b4b0b0805fb6c5f4966fc3ccef0738df4b97cb404e74da17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:44Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:44 crc kubenswrapper[4754]: I0126 16:47:44.056073 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:44Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:44 crc kubenswrapper[4754]: I0126 16:47:44.074439 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c3718a4-f354-4284-92e0-fdfb45a692bd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b86f46f49163bca7eca9458279029af330dbd4890b03325e06fa3b1bdf030de6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxjcz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fffc03b536bfee3c7bc1bddd500e834e34b391896477a1501e3e06201a19374\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxjcz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-x65wv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:44Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:44 crc kubenswrapper[4754]: I0126 16:47:44.110003 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jsbxt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"689569c4-93ee-4b82-93fc-9af0f97e6dc3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://033f8b104d4e9ff6d5f87ca998665f12e0c56696b3d30e01ede05f4da540365d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://919572e731813a98c7d8a94879bba931f75c472c7cd681e86d5197ebab43815d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1ac1f49bd564750fcff530a0f76657b0ed41573539acd020a922d19861ef834\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a300a80ad87f46ba1b4fd7ad3cf6c25767fcfe966b93bdd86d729843f53a5cda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31dcab2f9a2d218b0e49dad1df46377ff4521592824a0fbac254d5b8e6d0632e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fe4d9395447ac570eca1338e749dbefea76295d427a12487b63a142a7d6a2e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://481587b671ea36286ee7fab4da74db0b8fbcab26870e39ab02571e4bdcb46c3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://481587b671ea36286ee7fab4da74db0b8fbcab26870e39ab02571e4bdcb46c3b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-26T16:47:43Z\\\",\\\"message\\\":\\\"vn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI0126 16:47:43.017554 6107 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0126 16:47:43.017588 6107 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0126 16:47:43.017622 6107 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0126 16:47:43.017605 6107 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0126 16:47:43.018017 6107 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0126 16:47:43.018060 6107 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0126 16:47:43.018099 6107 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0126 16:47:43.018111 6107 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0126 16:47:43.018122 6107 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0126 16:47:43.018120 6107 handler.go:208] Removed *v1.Node event handler 2\\\\nI0126 16:47:43.018181 6107 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0126 16:47:43.018199 6107 factory.go:656] Stopping watch factory\\\\nI0126 16:47:43.018227 6107 handler.go:208] Removed *v1.Node event handler 7\\\\nI0126 16:47:43.018234 6107 handler.go:208] Removed *v1.Namespace ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://07a43ff7860ff65faf51f4a5e8aeb326f06d6d7222677a92560ae2e3a0c81484\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d9a7bafabd44f8b88a6f62cbe45599af8563bc45d9d7ad3a7669f3d8d11eb9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d9a7bafabd44f8b88a6f62cbe45599af8563bc45d9d7ad3a7669f3d8d11eb9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jsbxt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:44Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:44 crc kubenswrapper[4754]: I0126 16:47:44.114163 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:44 crc kubenswrapper[4754]: I0126 16:47:44.114217 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:44 crc kubenswrapper[4754]: I0126 16:47:44.114228 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:44 crc kubenswrapper[4754]: I0126 16:47:44.114247 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:44 crc kubenswrapper[4754]: I0126 16:47:44.114259 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:44Z","lastTransitionTime":"2026-01-26T16:47:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:44 crc kubenswrapper[4754]: I0126 16:47:44.128553 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:44Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:44 crc kubenswrapper[4754]: I0126 16:47:44.246389 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:44 crc kubenswrapper[4754]: I0126 16:47:44.246455 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:44 crc kubenswrapper[4754]: I0126 16:47:44.246469 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:44 crc kubenswrapper[4754]: I0126 16:47:44.246496 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:44 crc kubenswrapper[4754]: I0126 16:47:44.246510 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:44Z","lastTransitionTime":"2026-01-26T16:47:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:44 crc kubenswrapper[4754]: I0126 16:47:44.253615 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e459fd8c1093d6f3d0a671751b792b26b7df94c21ced1be1a2250821bcc82d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:44Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:44 crc kubenswrapper[4754]: I0126 16:47:44.265285 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2blzw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca65468e-30c6-4666-962e-cc0de05e67fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4603235bb2af495901926b805fe1c0ffc48824a4394ccf928e7fcd53a673ac3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c25z4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:33Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2blzw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:44Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:44 crc kubenswrapper[4754]: I0126 16:47:44.280919 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bwpd6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"351cc6c0-b8e5-4589-ae63-f98e8ab1eca3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1022f101d36873b41bbc1ed8a4a2590347b49f7cce6a172aa1d5bc9f529aed65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3ce0fe080a094316b9456057e347b90a5ff432daa12b9c56d7aea9bee1f72b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3ce0fe080a094316b9456057e347b90a5ff432daa12b9c56d7aea9bee1f72b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1772bb41eb8a36a863d3c691610bcbee5d16b260480b3e9f9d6b083e09e42799\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1772bb41eb8a36a863d3c691610bcbee5d16b260480b3e9f9d6b083e09e42799\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24848d69756024cc664affb9492764ace0638ef611f910f4f2bf86b638cc443d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://24848d69756024cc664affb9492764ace0638ef611f910f4f2bf86b638cc443d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c264fc7c3300a725566acce4464f01d9314779737624401c164d867831fd93bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c264fc7c3300a725566acce4464f01d9314779737624401c164d867831fd93bf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5acfe327eee4aad927ac3a32d24549eff788b14a1c4bdac359b6b56301d884ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5acfe327eee4aad927ac3a32d24549eff788b14a1c4bdac359b6b56301d884ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://63a9b9317494d0e0b33970ef3d461729b10d2a1ad7d3d2bd33f389b184ca56ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://63a9b9317494d0e0b33970ef3d461729b10d2a1ad7d3d2bd33f389b184ca56ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bwpd6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:44Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:44 crc kubenswrapper[4754]: I0126 16:47:44.294978 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-skh5n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b619d233-b592-4b05-a0b7-dc094c88471e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9231a4191565e2e0ef53d8a394bb34615c900b601a0dfec24c2d8e5fa7b0b659\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-skh5n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:44Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:44 crc kubenswrapper[4754]: I0126 16:47:44.310077 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"099af702-abf2-4933-b4c9-63ef2326e468\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23017e73f8450839b4c2c02fa61235ffe4604bdce11556def27f25bcadcf89dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a74507c941b9de4d0401c8f1d53bb79ed11359628d3405cdab7f5474a75c862e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4402182b3e2e35a7d5dd4c44ac53e152915b312b7ccb076c4f2b41b925a6e7f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f772a003566198519fe1849fd14f29d4346004b2e655810662507c95b4d5ffa1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:13Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:44Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:44 crc kubenswrapper[4754]: I0126 16:47:44.318477 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 26 16:47:44 crc kubenswrapper[4754]: I0126 16:47:44.326226 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:44Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:44 crc kubenswrapper[4754]: I0126 16:47:44.338620 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89251fa9c64bc25d6d6f355400d7644c099f22aed1cae812fc991a03595f1e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e90508d8a53be9146f9b388750730fdae9ee200306caf5ad393a356390a69ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:44Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:44 crc kubenswrapper[4754]: I0126 16:47:44.349131 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:44 crc kubenswrapper[4754]: I0126 16:47:44.349185 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:44 crc kubenswrapper[4754]: I0126 16:47:44.349198 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:44 crc kubenswrapper[4754]: I0126 16:47:44.349219 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:44 crc kubenswrapper[4754]: I0126 16:47:44.349231 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:44Z","lastTransitionTime":"2026-01-26T16:47:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:44 crc kubenswrapper[4754]: I0126 16:47:44.351896 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-84lgq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6763d962-0892-4acc-bab9-7a4e93e39a85\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://953c1917fca896c3def3fc2c69f9e3dbdc9ca26fc2b3d40282be71f71759fd68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8zlw6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:33Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-84lgq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:44Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:44 crc kubenswrapper[4754]: I0126 16:47:44.366838 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:44Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:44 crc kubenswrapper[4754]: I0126 16:47:44.380766 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e459fd8c1093d6f3d0a671751b792b26b7df94c21ced1be1a2250821bcc82d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:44Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:44 crc kubenswrapper[4754]: I0126 16:47:44.391790 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2blzw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca65468e-30c6-4666-962e-cc0de05e67fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4603235bb2af495901926b805fe1c0ffc48824a4394ccf928e7fcd53a673ac3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c25z4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:33Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2blzw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:44Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:44 crc kubenswrapper[4754]: I0126 16:47:44.410835 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bwpd6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"351cc6c0-b8e5-4589-ae63-f98e8ab1eca3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1022f101d36873b41bbc1ed8a4a2590347b49f7cce6a172aa1d5bc9f529aed65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3ce0fe080a094316b9456057e347b90a5ff432daa12b9c56d7aea9bee1f72b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3ce0fe080a094316b9456057e347b90a5ff432daa12b9c56d7aea9bee1f72b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1772bb41eb8a36a863d3c691610bcbee5d16b260480b3e9f9d6b083e09e42799\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1772bb41eb8a36a863d3c691610bcbee5d16b260480b3e9f9d6b083e09e42799\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24848d69756024cc664affb9492764ace0638ef611f910f4f2bf86b638cc443d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://24848d69756024cc664affb9492764ace0638ef611f910f4f2bf86b638cc443d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c264fc7c3300a725566acce4464f01d9314779737624401c164d867831fd93bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c264fc7c3300a725566acce4464f01d9314779737624401c164d867831fd93bf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5acfe327eee4aad927ac3a32d24549eff788b14a1c4bdac359b6b56301d884ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5acfe327eee4aad927ac3a32d24549eff788b14a1c4bdac359b6b56301d884ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://63a9b9317494d0e0b33970ef3d461729b10d2a1ad7d3d2bd33f389b184ca56ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://63a9b9317494d0e0b33970ef3d461729b10d2a1ad7d3d2bd33f389b184ca56ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bwpd6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:44Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:44 crc kubenswrapper[4754]: I0126 16:47:44.429101 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jsbxt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"689569c4-93ee-4b82-93fc-9af0f97e6dc3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://033f8b104d4e9ff6d5f87ca998665f12e0c56696b3d30e01ede05f4da540365d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://919572e731813a98c7d8a94879bba931f75c472c7cd681e86d5197ebab43815d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1ac1f49bd564750fcff530a0f76657b0ed41573539acd020a922d19861ef834\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a300a80ad87f46ba1b4fd7ad3cf6c25767fcfe966b93bdd86d729843f53a5cda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31dcab2f9a2d218b0e49dad1df46377ff4521592824a0fbac254d5b8e6d0632e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fe4d9395447ac570eca1338e749dbefea76295d427a12487b63a142a7d6a2e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://481587b671ea36286ee7fab4da74db0b8fbcab26870e39ab02571e4bdcb46c3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://481587b671ea36286ee7fab4da74db0b8fbcab26870e39ab02571e4bdcb46c3b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-26T16:47:43Z\\\",\\\"message\\\":\\\"vn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI0126 16:47:43.017554 6107 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0126 16:47:43.017588 6107 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0126 16:47:43.017622 6107 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0126 16:47:43.017605 6107 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0126 16:47:43.018017 6107 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0126 16:47:43.018060 6107 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0126 16:47:43.018099 6107 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0126 16:47:43.018111 6107 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0126 16:47:43.018122 6107 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0126 16:47:43.018120 6107 handler.go:208] Removed *v1.Node event handler 2\\\\nI0126 16:47:43.018181 6107 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0126 16:47:43.018199 6107 factory.go:656] Stopping watch factory\\\\nI0126 16:47:43.018227 6107 handler.go:208] Removed *v1.Node event handler 7\\\\nI0126 16:47:43.018234 6107 handler.go:208] Removed *v1.Namespace ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://07a43ff7860ff65faf51f4a5e8aeb326f06d6d7222677a92560ae2e3a0c81484\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d9a7bafabd44f8b88a6f62cbe45599af8563bc45d9d7ad3a7669f3d8d11eb9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d9a7bafabd44f8b88a6f62cbe45599af8563bc45d9d7ad3a7669f3d8d11eb9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jsbxt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:44Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:44 crc kubenswrapper[4754]: I0126 16:47:44.446307 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"099af702-abf2-4933-b4c9-63ef2326e468\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23017e73f8450839b4c2c02fa61235ffe4604bdce11556def27f25bcadcf89dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a74507c941b9de4d0401c8f1d53bb79ed11359628d3405cdab7f5474a75c862e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4402182b3e2e35a7d5dd4c44ac53e152915b312b7ccb076c4f2b41b925a6e7f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f772a003566198519fe1849fd14f29d4346004b2e655810662507c95b4d5ffa1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:13Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:44Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:44 crc kubenswrapper[4754]: I0126 16:47:44.451230 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:44 crc kubenswrapper[4754]: I0126 16:47:44.451269 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:44 crc kubenswrapper[4754]: I0126 16:47:44.451278 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:44 crc kubenswrapper[4754]: I0126 16:47:44.451292 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:44 crc kubenswrapper[4754]: I0126 16:47:44.451301 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:44Z","lastTransitionTime":"2026-01-26T16:47:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:44 crc kubenswrapper[4754]: I0126 16:47:44.467342 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:44Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:44 crc kubenswrapper[4754]: I0126 16:47:44.482327 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89251fa9c64bc25d6d6f355400d7644c099f22aed1cae812fc991a03595f1e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e90508d8a53be9146f9b388750730fdae9ee200306caf5ad393a356390a69ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:44Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:44 crc kubenswrapper[4754]: I0126 16:47:44.494621 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-84lgq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6763d962-0892-4acc-bab9-7a4e93e39a85\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://953c1917fca896c3def3fc2c69f9e3dbdc9ca26fc2b3d40282be71f71759fd68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8zlw6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:33Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-84lgq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:44Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:44 crc kubenswrapper[4754]: I0126 16:47:44.509097 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-skh5n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b619d233-b592-4b05-a0b7-dc094c88471e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9231a4191565e2e0ef53d8a394bb34615c900b601a0dfec24c2d8e5fa7b0b659\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-skh5n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:44Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:44 crc kubenswrapper[4754]: I0126 16:47:44.523387 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b89aa081-0a2e-4757-9210-e2a6c228bee9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9c5f40188fda14b14128514fe84bddfe8dcede3d6c560283ccb2ac0c66895fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f45445cbf1d180d6d4dd9588e85bf678f043bfce8cff0812b78ca3510904f9a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://085fa7b8d4fef8b2722fac8c346107699b5a6bd33fbfc74ed35831cb8fdf9dad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c40e018a0979035827963efb9efc48a878a55b9df68ec1a9242d7cff64fe480\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8356f76fbc3a4e636c12d1da5455d3f6767d9b83ce2f2925a11435d7d7261e9f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0126 16:47:26.205208 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0126 16:47:26.206538 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3645934156/tls.crt::/tmp/serving-cert-3645934156/tls.key\\\\\\\"\\\\nI0126 16:47:31.504518 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0126 16:47:31.509366 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0126 16:47:31.509400 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0126 16:47:31.509428 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0126 16:47:31.509436 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0126 16:47:31.515912 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0126 16:47:31.516036 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 16:47:31.516068 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 16:47:31.516096 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0126 16:47:31.516121 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0126 16:47:31.516145 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0126 16:47:31.516170 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0126 16:47:31.516412 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0126 16:47:31.519294 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27b74e2e65a547e0e28fddd4f15feacb17f4d57596eeb8da82df9d39243c85e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22ba7649a1308815bfad52016000987a7a9a5bc214d7ff61e137ba099b20fc6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://22ba7649a1308815bfad52016000987a7a9a5bc214d7ff61e137ba099b20fc6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:13Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:44Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:44 crc kubenswrapper[4754]: I0126 16:47:44.537339 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://205ec256e46a8e08b4b0b0805fb6c5f4966fc3ccef0738df4b97cb404e74da17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:44Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:44 crc kubenswrapper[4754]: I0126 16:47:44.551176 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:44Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:44 crc kubenswrapper[4754]: I0126 16:47:44.553593 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:44 crc kubenswrapper[4754]: I0126 16:47:44.553621 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:44 crc kubenswrapper[4754]: I0126 16:47:44.553634 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:44 crc kubenswrapper[4754]: I0126 16:47:44.553653 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:44 crc kubenswrapper[4754]: I0126 16:47:44.553680 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:44Z","lastTransitionTime":"2026-01-26T16:47:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:44 crc kubenswrapper[4754]: I0126 16:47:44.566619 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c3718a4-f354-4284-92e0-fdfb45a692bd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b86f46f49163bca7eca9458279029af330dbd4890b03325e06fa3b1bdf030de6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxjcz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fffc03b536bfee3c7bc1bddd500e834e34b391896477a1501e3e06201a19374\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxjcz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-x65wv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:44Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:44 crc kubenswrapper[4754]: I0126 16:47:44.656463 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:44 crc kubenswrapper[4754]: I0126 16:47:44.656508 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:44 crc kubenswrapper[4754]: I0126 16:47:44.656519 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:44 crc kubenswrapper[4754]: I0126 16:47:44.656533 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:44 crc kubenswrapper[4754]: I0126 16:47:44.656545 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:44Z","lastTransitionTime":"2026-01-26T16:47:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:44 crc kubenswrapper[4754]: I0126 16:47:44.729186 4754 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-01 17:00:44.511970273 +0000 UTC Jan 26 16:47:44 crc kubenswrapper[4754]: I0126 16:47:44.759301 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:44 crc kubenswrapper[4754]: I0126 16:47:44.759370 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:44 crc kubenswrapper[4754]: I0126 16:47:44.759379 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:44 crc kubenswrapper[4754]: I0126 16:47:44.759393 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:44 crc kubenswrapper[4754]: I0126 16:47:44.759404 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:44Z","lastTransitionTime":"2026-01-26T16:47:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:44 crc kubenswrapper[4754]: I0126 16:47:44.843822 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:44 crc kubenswrapper[4754]: I0126 16:47:44.843882 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:44 crc kubenswrapper[4754]: I0126 16:47:44.843892 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:44 crc kubenswrapper[4754]: I0126 16:47:44.843910 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:44 crc kubenswrapper[4754]: I0126 16:47:44.843920 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:44Z","lastTransitionTime":"2026-01-26T16:47:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:44 crc kubenswrapper[4754]: E0126 16:47:44.857616 4754 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:47:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:47:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:44Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:47:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:47:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:44Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d8506764-e7fa-45cb-a13d-6f527164f548\\\",\\\"systemUUID\\\":\\\"bfbb6c33-ce46-4e81-a8a8-b44409b03821\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:44Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:44 crc kubenswrapper[4754]: I0126 16:47:44.861298 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:44 crc kubenswrapper[4754]: I0126 16:47:44.861341 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:44 crc kubenswrapper[4754]: I0126 16:47:44.861351 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:44 crc kubenswrapper[4754]: I0126 16:47:44.861371 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:44 crc kubenswrapper[4754]: I0126 16:47:44.861385 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:44Z","lastTransitionTime":"2026-01-26T16:47:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:44 crc kubenswrapper[4754]: E0126 16:47:44.874204 4754 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:47:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:47:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:44Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:47:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:47:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:44Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d8506764-e7fa-45cb-a13d-6f527164f548\\\",\\\"systemUUID\\\":\\\"bfbb6c33-ce46-4e81-a8a8-b44409b03821\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:44Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:44 crc kubenswrapper[4754]: I0126 16:47:44.877861 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:44 crc kubenswrapper[4754]: I0126 16:47:44.877892 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:44 crc kubenswrapper[4754]: I0126 16:47:44.877900 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:44 crc kubenswrapper[4754]: I0126 16:47:44.877912 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:44 crc kubenswrapper[4754]: I0126 16:47:44.877920 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:44Z","lastTransitionTime":"2026-01-26T16:47:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:44 crc kubenswrapper[4754]: E0126 16:47:44.889618 4754 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:47:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:47:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:44Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:47:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:47:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:44Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d8506764-e7fa-45cb-a13d-6f527164f548\\\",\\\"systemUUID\\\":\\\"bfbb6c33-ce46-4e81-a8a8-b44409b03821\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:44Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:44 crc kubenswrapper[4754]: I0126 16:47:44.892773 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:44 crc kubenswrapper[4754]: I0126 16:47:44.892807 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:44 crc kubenswrapper[4754]: I0126 16:47:44.892816 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:44 crc kubenswrapper[4754]: I0126 16:47:44.892829 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:44 crc kubenswrapper[4754]: I0126 16:47:44.892841 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:44Z","lastTransitionTime":"2026-01-26T16:47:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:44 crc kubenswrapper[4754]: E0126 16:47:44.904140 4754 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:47:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:47:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:44Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:47:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:47:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:44Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d8506764-e7fa-45cb-a13d-6f527164f548\\\",\\\"systemUUID\\\":\\\"bfbb6c33-ce46-4e81-a8a8-b44409b03821\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:44Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:44 crc kubenswrapper[4754]: I0126 16:47:44.907197 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:44 crc kubenswrapper[4754]: I0126 16:47:44.907228 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:44 crc kubenswrapper[4754]: I0126 16:47:44.907237 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:44 crc kubenswrapper[4754]: I0126 16:47:44.907250 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:44 crc kubenswrapper[4754]: I0126 16:47:44.907259 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:44Z","lastTransitionTime":"2026-01-26T16:47:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:44 crc kubenswrapper[4754]: E0126 16:47:44.922454 4754 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:47:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:47:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:44Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:47:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:47:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:44Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d8506764-e7fa-45cb-a13d-6f527164f548\\\",\\\"systemUUID\\\":\\\"bfbb6c33-ce46-4e81-a8a8-b44409b03821\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:44Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:44 crc kubenswrapper[4754]: E0126 16:47:44.922573 4754 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Jan 26 16:47:44 crc kubenswrapper[4754]: I0126 16:47:44.924062 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:44 crc kubenswrapper[4754]: I0126 16:47:44.924093 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:44 crc kubenswrapper[4754]: I0126 16:47:44.924101 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:44 crc kubenswrapper[4754]: I0126 16:47:44.924116 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:44 crc kubenswrapper[4754]: I0126 16:47:44.924126 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:44Z","lastTransitionTime":"2026-01-26T16:47:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:45 crc kubenswrapper[4754]: I0126 16:47:45.015068 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-jsbxt_689569c4-93ee-4b82-93fc-9af0f97e6dc3/ovnkube-controller/0.log" Jan 26 16:47:45 crc kubenswrapper[4754]: I0126 16:47:45.017830 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jsbxt" event={"ID":"689569c4-93ee-4b82-93fc-9af0f97e6dc3","Type":"ContainerStarted","Data":"eaed406e767c0eedc024810541a22ed6ac3b1dcc2b5d41aa007b9ad282eb57fd"} Jan 26 16:47:45 crc kubenswrapper[4754]: I0126 16:47:45.018282 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-jsbxt" Jan 26 16:47:45 crc kubenswrapper[4754]: I0126 16:47:45.026626 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:45 crc kubenswrapper[4754]: I0126 16:47:45.027350 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:45 crc kubenswrapper[4754]: I0126 16:47:45.027368 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:45 crc kubenswrapper[4754]: I0126 16:47:45.027385 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:45 crc kubenswrapper[4754]: I0126 16:47:45.027397 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:45Z","lastTransitionTime":"2026-01-26T16:47:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:45 crc kubenswrapper[4754]: I0126 16:47:45.031311 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:45Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:45 crc kubenswrapper[4754]: I0126 16:47:45.043861 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e459fd8c1093d6f3d0a671751b792b26b7df94c21ced1be1a2250821bcc82d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:45Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:45 crc kubenswrapper[4754]: I0126 16:47:45.056523 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2blzw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca65468e-30c6-4666-962e-cc0de05e67fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4603235bb2af495901926b805fe1c0ffc48824a4394ccf928e7fcd53a673ac3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c25z4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:33Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2blzw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:45Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:45 crc kubenswrapper[4754]: I0126 16:47:45.072414 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bwpd6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"351cc6c0-b8e5-4589-ae63-f98e8ab1eca3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1022f101d36873b41bbc1ed8a4a2590347b49f7cce6a172aa1d5bc9f529aed65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3ce0fe080a094316b9456057e347b90a5ff432daa12b9c56d7aea9bee1f72b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3ce0fe080a094316b9456057e347b90a5ff432daa12b9c56d7aea9bee1f72b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1772bb41eb8a36a863d3c691610bcbee5d16b260480b3e9f9d6b083e09e42799\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1772bb41eb8a36a863d3c691610bcbee5d16b260480b3e9f9d6b083e09e42799\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24848d69756024cc664affb9492764ace0638ef611f910f4f2bf86b638cc443d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://24848d69756024cc664affb9492764ace0638ef611f910f4f2bf86b638cc443d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c264fc7c3300a725566acce4464f01d9314779737624401c164d867831fd93bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c264fc7c3300a725566acce4464f01d9314779737624401c164d867831fd93bf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5acfe327eee4aad927ac3a32d24549eff788b14a1c4bdac359b6b56301d884ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5acfe327eee4aad927ac3a32d24549eff788b14a1c4bdac359b6b56301d884ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://63a9b9317494d0e0b33970ef3d461729b10d2a1ad7d3d2bd33f389b184ca56ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://63a9b9317494d0e0b33970ef3d461729b10d2a1ad7d3d2bd33f389b184ca56ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bwpd6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:45Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:45 crc kubenswrapper[4754]: I0126 16:47:45.092976 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jsbxt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"689569c4-93ee-4b82-93fc-9af0f97e6dc3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://033f8b104d4e9ff6d5f87ca998665f12e0c56696b3d30e01ede05f4da540365d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://919572e731813a98c7d8a94879bba931f75c472c7cd681e86d5197ebab43815d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1ac1f49bd564750fcff530a0f76657b0ed41573539acd020a922d19861ef834\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a300a80ad87f46ba1b4fd7ad3cf6c25767fcfe966b93bdd86d729843f53a5cda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31dcab2f9a2d218b0e49dad1df46377ff4521592824a0fbac254d5b8e6d0632e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fe4d9395447ac570eca1338e749dbefea76295d427a12487b63a142a7d6a2e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eaed406e767c0eedc024810541a22ed6ac3b1dcc2b5d41aa007b9ad282eb57fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://481587b671ea36286ee7fab4da74db0b8fbcab26870e39ab02571e4bdcb46c3b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-26T16:47:43Z\\\",\\\"message\\\":\\\"vn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI0126 16:47:43.017554 6107 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0126 16:47:43.017588 6107 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0126 16:47:43.017622 6107 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0126 16:47:43.017605 6107 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0126 16:47:43.018017 6107 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0126 16:47:43.018060 6107 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0126 16:47:43.018099 6107 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0126 16:47:43.018111 6107 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0126 16:47:43.018122 6107 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0126 16:47:43.018120 6107 handler.go:208] Removed *v1.Node event handler 2\\\\nI0126 16:47:43.018181 6107 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0126 16:47:43.018199 6107 factory.go:656] Stopping watch factory\\\\nI0126 16:47:43.018227 6107 handler.go:208] Removed *v1.Node event handler 7\\\\nI0126 16:47:43.018234 6107 handler.go:208] Removed *v1.Namespace ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:40Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://07a43ff7860ff65faf51f4a5e8aeb326f06d6d7222677a92560ae2e3a0c81484\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d9a7bafabd44f8b88a6f62cbe45599af8563bc45d9d7ad3a7669f3d8d11eb9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d9a7bafabd44f8b88a6f62cbe45599af8563bc45d9d7ad3a7669f3d8d11eb9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jsbxt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:45Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:45 crc kubenswrapper[4754]: I0126 16:47:45.106837 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:45Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:45 crc kubenswrapper[4754]: I0126 16:47:45.123100 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89251fa9c64bc25d6d6f355400d7644c099f22aed1cae812fc991a03595f1e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e90508d8a53be9146f9b388750730fdae9ee200306caf5ad393a356390a69ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:45Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:45 crc kubenswrapper[4754]: I0126 16:47:45.129762 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:45 crc kubenswrapper[4754]: I0126 16:47:45.129805 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:45 crc kubenswrapper[4754]: I0126 16:47:45.129815 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:45 crc kubenswrapper[4754]: I0126 16:47:45.129831 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:45 crc kubenswrapper[4754]: I0126 16:47:45.129842 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:45Z","lastTransitionTime":"2026-01-26T16:47:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:45 crc kubenswrapper[4754]: I0126 16:47:45.139173 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-84lgq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6763d962-0892-4acc-bab9-7a4e93e39a85\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://953c1917fca896c3def3fc2c69f9e3dbdc9ca26fc2b3d40282be71f71759fd68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8zlw6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:33Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-84lgq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:45Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:45 crc kubenswrapper[4754]: I0126 16:47:45.161496 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-skh5n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b619d233-b592-4b05-a0b7-dc094c88471e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9231a4191565e2e0ef53d8a394bb34615c900b601a0dfec24c2d8e5fa7b0b659\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-skh5n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:45Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:45 crc kubenswrapper[4754]: I0126 16:47:45.179736 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"099af702-abf2-4933-b4c9-63ef2326e468\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23017e73f8450839b4c2c02fa61235ffe4604bdce11556def27f25bcadcf89dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a74507c941b9de4d0401c8f1d53bb79ed11359628d3405cdab7f5474a75c862e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4402182b3e2e35a7d5dd4c44ac53e152915b312b7ccb076c4f2b41b925a6e7f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f772a003566198519fe1849fd14f29d4346004b2e655810662507c95b4d5ffa1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:13Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:45Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:45 crc kubenswrapper[4754]: I0126 16:47:45.204404 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://205ec256e46a8e08b4b0b0805fb6c5f4966fc3ccef0738df4b97cb404e74da17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:45Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:45 crc kubenswrapper[4754]: I0126 16:47:45.220504 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:45Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:45 crc kubenswrapper[4754]: I0126 16:47:45.232052 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:45 crc kubenswrapper[4754]: I0126 16:47:45.232092 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:45 crc kubenswrapper[4754]: I0126 16:47:45.232103 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:45 crc kubenswrapper[4754]: I0126 16:47:45.232119 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:45 crc kubenswrapper[4754]: I0126 16:47:45.232130 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:45Z","lastTransitionTime":"2026-01-26T16:47:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:45 crc kubenswrapper[4754]: I0126 16:47:45.233106 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c3718a4-f354-4284-92e0-fdfb45a692bd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b86f46f49163bca7eca9458279029af330dbd4890b03325e06fa3b1bdf030de6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxjcz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fffc03b536bfee3c7bc1bddd500e834e34b391896477a1501e3e06201a19374\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxjcz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-x65wv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:45Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:45 crc kubenswrapper[4754]: I0126 16:47:45.247010 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b89aa081-0a2e-4757-9210-e2a6c228bee9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9c5f40188fda14b14128514fe84bddfe8dcede3d6c560283ccb2ac0c66895fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f45445cbf1d180d6d4dd9588e85bf678f043bfce8cff0812b78ca3510904f9a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://085fa7b8d4fef8b2722fac8c346107699b5a6bd33fbfc74ed35831cb8fdf9dad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c40e018a0979035827963efb9efc48a878a55b9df68ec1a9242d7cff64fe480\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8356f76fbc3a4e636c12d1da5455d3f6767d9b83ce2f2925a11435d7d7261e9f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0126 16:47:26.205208 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0126 16:47:26.206538 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3645934156/tls.crt::/tmp/serving-cert-3645934156/tls.key\\\\\\\"\\\\nI0126 16:47:31.504518 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0126 16:47:31.509366 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0126 16:47:31.509400 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0126 16:47:31.509428 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0126 16:47:31.509436 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0126 16:47:31.515912 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0126 16:47:31.516036 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 16:47:31.516068 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 16:47:31.516096 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0126 16:47:31.516121 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0126 16:47:31.516145 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0126 16:47:31.516170 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0126 16:47:31.516412 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0126 16:47:31.519294 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27b74e2e65a547e0e28fddd4f15feacb17f4d57596eeb8da82df9d39243c85e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22ba7649a1308815bfad52016000987a7a9a5bc214d7ff61e137ba099b20fc6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://22ba7649a1308815bfad52016000987a7a9a5bc214d7ff61e137ba099b20fc6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:13Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:45Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:45 crc kubenswrapper[4754]: I0126 16:47:45.335049 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:45 crc kubenswrapper[4754]: I0126 16:47:45.335094 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:45 crc kubenswrapper[4754]: I0126 16:47:45.335104 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:45 crc kubenswrapper[4754]: I0126 16:47:45.335120 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:45 crc kubenswrapper[4754]: I0126 16:47:45.335130 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:45Z","lastTransitionTime":"2026-01-26T16:47:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:45 crc kubenswrapper[4754]: I0126 16:47:45.437460 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:45 crc kubenswrapper[4754]: I0126 16:47:45.437739 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:45 crc kubenswrapper[4754]: I0126 16:47:45.437811 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:45 crc kubenswrapper[4754]: I0126 16:47:45.437881 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:45 crc kubenswrapper[4754]: I0126 16:47:45.437977 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:45Z","lastTransitionTime":"2026-01-26T16:47:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:45 crc kubenswrapper[4754]: I0126 16:47:45.540349 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:45 crc kubenswrapper[4754]: I0126 16:47:45.540433 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:45 crc kubenswrapper[4754]: I0126 16:47:45.540449 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:45 crc kubenswrapper[4754]: I0126 16:47:45.540475 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:45 crc kubenswrapper[4754]: I0126 16:47:45.540499 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:45Z","lastTransitionTime":"2026-01-26T16:47:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:45 crc kubenswrapper[4754]: I0126 16:47:45.643528 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:45 crc kubenswrapper[4754]: I0126 16:47:45.643875 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:45 crc kubenswrapper[4754]: I0126 16:47:45.644047 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:45 crc kubenswrapper[4754]: I0126 16:47:45.644184 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:45 crc kubenswrapper[4754]: I0126 16:47:45.644307 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:45Z","lastTransitionTime":"2026-01-26T16:47:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:45 crc kubenswrapper[4754]: I0126 16:47:45.729574 4754 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-16 02:58:40.145788858 +0000 UTC Jan 26 16:47:45 crc kubenswrapper[4754]: I0126 16:47:45.746438 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:45 crc kubenswrapper[4754]: I0126 16:47:45.746684 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:45 crc kubenswrapper[4754]: I0126 16:47:45.746754 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:45 crc kubenswrapper[4754]: I0126 16:47:45.746894 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:45 crc kubenswrapper[4754]: I0126 16:47:45.746976 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:45Z","lastTransitionTime":"2026-01-26T16:47:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:45 crc kubenswrapper[4754]: I0126 16:47:45.767464 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 16:47:45 crc kubenswrapper[4754]: E0126 16:47:45.767750 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 16:47:45 crc kubenswrapper[4754]: I0126 16:47:45.767613 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 16:47:45 crc kubenswrapper[4754]: E0126 16:47:45.767852 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 16:47:45 crc kubenswrapper[4754]: I0126 16:47:45.767487 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 16:47:45 crc kubenswrapper[4754]: E0126 16:47:45.767916 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 16:47:45 crc kubenswrapper[4754]: I0126 16:47:45.848849 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:45 crc kubenswrapper[4754]: I0126 16:47:45.849091 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:45 crc kubenswrapper[4754]: I0126 16:47:45.849185 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:45 crc kubenswrapper[4754]: I0126 16:47:45.849275 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:45 crc kubenswrapper[4754]: I0126 16:47:45.849360 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:45Z","lastTransitionTime":"2026-01-26T16:47:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:45 crc kubenswrapper[4754]: I0126 16:47:45.951743 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:45 crc kubenswrapper[4754]: I0126 16:47:45.951787 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:45 crc kubenswrapper[4754]: I0126 16:47:45.951795 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:45 crc kubenswrapper[4754]: I0126 16:47:45.951810 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:45 crc kubenswrapper[4754]: I0126 16:47:45.951819 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:45Z","lastTransitionTime":"2026-01-26T16:47:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:46 crc kubenswrapper[4754]: I0126 16:47:46.022890 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-jsbxt_689569c4-93ee-4b82-93fc-9af0f97e6dc3/ovnkube-controller/1.log" Jan 26 16:47:46 crc kubenswrapper[4754]: I0126 16:47:46.023342 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-jsbxt_689569c4-93ee-4b82-93fc-9af0f97e6dc3/ovnkube-controller/0.log" Jan 26 16:47:46 crc kubenswrapper[4754]: I0126 16:47:46.031327 4754 generic.go:334] "Generic (PLEG): container finished" podID="689569c4-93ee-4b82-93fc-9af0f97e6dc3" containerID="eaed406e767c0eedc024810541a22ed6ac3b1dcc2b5d41aa007b9ad282eb57fd" exitCode=1 Jan 26 16:47:46 crc kubenswrapper[4754]: I0126 16:47:46.031388 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jsbxt" event={"ID":"689569c4-93ee-4b82-93fc-9af0f97e6dc3","Type":"ContainerDied","Data":"eaed406e767c0eedc024810541a22ed6ac3b1dcc2b5d41aa007b9ad282eb57fd"} Jan 26 16:47:46 crc kubenswrapper[4754]: I0126 16:47:46.031464 4754 scope.go:117] "RemoveContainer" containerID="481587b671ea36286ee7fab4da74db0b8fbcab26870e39ab02571e4bdcb46c3b" Jan 26 16:47:46 crc kubenswrapper[4754]: I0126 16:47:46.032820 4754 scope.go:117] "RemoveContainer" containerID="eaed406e767c0eedc024810541a22ed6ac3b1dcc2b5d41aa007b9ad282eb57fd" Jan 26 16:47:46 crc kubenswrapper[4754]: E0126 16:47:46.034057 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-jsbxt_openshift-ovn-kubernetes(689569c4-93ee-4b82-93fc-9af0f97e6dc3)\"" pod="openshift-ovn-kubernetes/ovnkube-node-jsbxt" podUID="689569c4-93ee-4b82-93fc-9af0f97e6dc3" Jan 26 16:47:46 crc kubenswrapper[4754]: I0126 16:47:46.049477 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"099af702-abf2-4933-b4c9-63ef2326e468\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23017e73f8450839b4c2c02fa61235ffe4604bdce11556def27f25bcadcf89dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a74507c941b9de4d0401c8f1d53bb79ed11359628d3405cdab7f5474a75c862e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4402182b3e2e35a7d5dd4c44ac53e152915b312b7ccb076c4f2b41b925a6e7f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f772a003566198519fe1849fd14f29d4346004b2e655810662507c95b4d5ffa1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:13Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:46Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:46 crc kubenswrapper[4754]: I0126 16:47:46.053509 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:46 crc kubenswrapper[4754]: I0126 16:47:46.053555 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:46 crc kubenswrapper[4754]: I0126 16:47:46.053569 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:46 crc kubenswrapper[4754]: I0126 16:47:46.053586 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:46 crc kubenswrapper[4754]: I0126 16:47:46.053599 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:46Z","lastTransitionTime":"2026-01-26T16:47:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:46 crc kubenswrapper[4754]: I0126 16:47:46.070143 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:46Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:46 crc kubenswrapper[4754]: I0126 16:47:46.086637 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89251fa9c64bc25d6d6f355400d7644c099f22aed1cae812fc991a03595f1e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e90508d8a53be9146f9b388750730fdae9ee200306caf5ad393a356390a69ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:46Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:46 crc kubenswrapper[4754]: I0126 16:47:46.100808 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-84lgq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6763d962-0892-4acc-bab9-7a4e93e39a85\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://953c1917fca896c3def3fc2c69f9e3dbdc9ca26fc2b3d40282be71f71759fd68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8zlw6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:33Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-84lgq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:46Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:46 crc kubenswrapper[4754]: I0126 16:47:46.116288 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-skh5n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b619d233-b592-4b05-a0b7-dc094c88471e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9231a4191565e2e0ef53d8a394bb34615c900b601a0dfec24c2d8e5fa7b0b659\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-skh5n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:46Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:46 crc kubenswrapper[4754]: I0126 16:47:46.132796 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b89aa081-0a2e-4757-9210-e2a6c228bee9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9c5f40188fda14b14128514fe84bddfe8dcede3d6c560283ccb2ac0c66895fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f45445cbf1d180d6d4dd9588e85bf678f043bfce8cff0812b78ca3510904f9a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://085fa7b8d4fef8b2722fac8c346107699b5a6bd33fbfc74ed35831cb8fdf9dad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c40e018a0979035827963efb9efc48a878a55b9df68ec1a9242d7cff64fe480\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8356f76fbc3a4e636c12d1da5455d3f6767d9b83ce2f2925a11435d7d7261e9f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0126 16:47:26.205208 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0126 16:47:26.206538 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3645934156/tls.crt::/tmp/serving-cert-3645934156/tls.key\\\\\\\"\\\\nI0126 16:47:31.504518 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0126 16:47:31.509366 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0126 16:47:31.509400 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0126 16:47:31.509428 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0126 16:47:31.509436 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0126 16:47:31.515912 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0126 16:47:31.516036 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 16:47:31.516068 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 16:47:31.516096 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0126 16:47:31.516121 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0126 16:47:31.516145 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0126 16:47:31.516170 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0126 16:47:31.516412 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0126 16:47:31.519294 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27b74e2e65a547e0e28fddd4f15feacb17f4d57596eeb8da82df9d39243c85e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22ba7649a1308815bfad52016000987a7a9a5bc214d7ff61e137ba099b20fc6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://22ba7649a1308815bfad52016000987a7a9a5bc214d7ff61e137ba099b20fc6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:13Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:46Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:46 crc kubenswrapper[4754]: I0126 16:47:46.149782 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://205ec256e46a8e08b4b0b0805fb6c5f4966fc3ccef0738df4b97cb404e74da17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:46Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:46 crc kubenswrapper[4754]: I0126 16:47:46.156570 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:46 crc kubenswrapper[4754]: I0126 16:47:46.156615 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:46 crc kubenswrapper[4754]: I0126 16:47:46.156627 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:46 crc kubenswrapper[4754]: I0126 16:47:46.156645 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:46 crc kubenswrapper[4754]: I0126 16:47:46.156658 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:46Z","lastTransitionTime":"2026-01-26T16:47:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:46 crc kubenswrapper[4754]: I0126 16:47:46.166383 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:46Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:46 crc kubenswrapper[4754]: I0126 16:47:46.179104 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c3718a4-f354-4284-92e0-fdfb45a692bd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b86f46f49163bca7eca9458279029af330dbd4890b03325e06fa3b1bdf030de6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxjcz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fffc03b536bfee3c7bc1bddd500e834e34b391896477a1501e3e06201a19374\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxjcz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-x65wv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:46Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:46 crc kubenswrapper[4754]: I0126 16:47:46.192805 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:46Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:46 crc kubenswrapper[4754]: I0126 16:47:46.204568 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e459fd8c1093d6f3d0a671751b792b26b7df94c21ced1be1a2250821bcc82d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:46Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:46 crc kubenswrapper[4754]: I0126 16:47:46.215467 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2blzw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca65468e-30c6-4666-962e-cc0de05e67fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4603235bb2af495901926b805fe1c0ffc48824a4394ccf928e7fcd53a673ac3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c25z4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:33Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2blzw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:46Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:46 crc kubenswrapper[4754]: I0126 16:47:46.230264 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bwpd6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"351cc6c0-b8e5-4589-ae63-f98e8ab1eca3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1022f101d36873b41bbc1ed8a4a2590347b49f7cce6a172aa1d5bc9f529aed65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3ce0fe080a094316b9456057e347b90a5ff432daa12b9c56d7aea9bee1f72b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3ce0fe080a094316b9456057e347b90a5ff432daa12b9c56d7aea9bee1f72b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1772bb41eb8a36a863d3c691610bcbee5d16b260480b3e9f9d6b083e09e42799\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1772bb41eb8a36a863d3c691610bcbee5d16b260480b3e9f9d6b083e09e42799\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24848d69756024cc664affb9492764ace0638ef611f910f4f2bf86b638cc443d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://24848d69756024cc664affb9492764ace0638ef611f910f4f2bf86b638cc443d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c264fc7c3300a725566acce4464f01d9314779737624401c164d867831fd93bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c264fc7c3300a725566acce4464f01d9314779737624401c164d867831fd93bf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5acfe327eee4aad927ac3a32d24549eff788b14a1c4bdac359b6b56301d884ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5acfe327eee4aad927ac3a32d24549eff788b14a1c4bdac359b6b56301d884ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://63a9b9317494d0e0b33970ef3d461729b10d2a1ad7d3d2bd33f389b184ca56ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://63a9b9317494d0e0b33970ef3d461729b10d2a1ad7d3d2bd33f389b184ca56ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bwpd6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:46Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:46 crc kubenswrapper[4754]: I0126 16:47:46.249588 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jsbxt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"689569c4-93ee-4b82-93fc-9af0f97e6dc3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://033f8b104d4e9ff6d5f87ca998665f12e0c56696b3d30e01ede05f4da540365d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://919572e731813a98c7d8a94879bba931f75c472c7cd681e86d5197ebab43815d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1ac1f49bd564750fcff530a0f76657b0ed41573539acd020a922d19861ef834\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a300a80ad87f46ba1b4fd7ad3cf6c25767fcfe966b93bdd86d729843f53a5cda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31dcab2f9a2d218b0e49dad1df46377ff4521592824a0fbac254d5b8e6d0632e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fe4d9395447ac570eca1338e749dbefea76295d427a12487b63a142a7d6a2e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eaed406e767c0eedc024810541a22ed6ac3b1dcc2b5d41aa007b9ad282eb57fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://481587b671ea36286ee7fab4da74db0b8fbcab26870e39ab02571e4bdcb46c3b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-26T16:47:43Z\\\",\\\"message\\\":\\\"vn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI0126 16:47:43.017554 6107 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0126 16:47:43.017588 6107 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0126 16:47:43.017622 6107 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0126 16:47:43.017605 6107 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0126 16:47:43.018017 6107 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0126 16:47:43.018060 6107 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0126 16:47:43.018099 6107 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0126 16:47:43.018111 6107 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0126 16:47:43.018122 6107 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0126 16:47:43.018120 6107 handler.go:208] Removed *v1.Node event handler 2\\\\nI0126 16:47:43.018181 6107 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0126 16:47:43.018199 6107 factory.go:656] Stopping watch factory\\\\nI0126 16:47:43.018227 6107 handler.go:208] Removed *v1.Node event handler 7\\\\nI0126 16:47:43.018234 6107 handler.go:208] Removed *v1.Namespace ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:40Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eaed406e767c0eedc024810541a22ed6ac3b1dcc2b5d41aa007b9ad282eb57fd\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-26T16:47:45Z\\\",\\\"message\\\":\\\".EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0126 16:47:45.693395 6233 handler.go:208] Removed *v1.Node event handler 7\\\\nI0126 16:47:45.694648 6233 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI0126 16:47:45.694956 6233 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI0126 16:47:45.694961 6233 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0126 16:47:45.695763 6233 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0126 16:47:45.695841 6233 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0126 16:47:45.695852 6233 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0126 16:47:45.695893 6233 factory.go:656] Stopping watch factory\\\\nI0126 16:47:45.695913 6233 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0126 16:47:45.695920 6233 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0126 16:47:45.695935 6233 handler.go:208] Removed *v1.Namespace ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://07a43ff7860ff65faf51f4a5e8aeb326f06d6d7222677a92560ae2e3a0c81484\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d9a7bafabd44f8b88a6f62cbe45599af8563bc45d9d7ad3a7669f3d8d11eb9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d9a7bafabd44f8b88a6f62cbe45599af8563bc45d9d7ad3a7669f3d8d11eb9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jsbxt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:46Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:46 crc kubenswrapper[4754]: I0126 16:47:46.259906 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:46 crc kubenswrapper[4754]: I0126 16:47:46.259964 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:46 crc kubenswrapper[4754]: I0126 16:47:46.259982 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:46 crc kubenswrapper[4754]: I0126 16:47:46.260004 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:46 crc kubenswrapper[4754]: I0126 16:47:46.260020 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:46Z","lastTransitionTime":"2026-01-26T16:47:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:46 crc kubenswrapper[4754]: I0126 16:47:46.362465 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:46 crc kubenswrapper[4754]: I0126 16:47:46.362499 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:46 crc kubenswrapper[4754]: I0126 16:47:46.362507 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:46 crc kubenswrapper[4754]: I0126 16:47:46.362521 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:46 crc kubenswrapper[4754]: I0126 16:47:46.362534 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:46Z","lastTransitionTime":"2026-01-26T16:47:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:46 crc kubenswrapper[4754]: I0126 16:47:46.465212 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:46 crc kubenswrapper[4754]: I0126 16:47:46.465283 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:46 crc kubenswrapper[4754]: I0126 16:47:46.465297 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:46 crc kubenswrapper[4754]: I0126 16:47:46.465318 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:46 crc kubenswrapper[4754]: I0126 16:47:46.465332 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:46Z","lastTransitionTime":"2026-01-26T16:47:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:46 crc kubenswrapper[4754]: I0126 16:47:46.568718 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:46 crc kubenswrapper[4754]: I0126 16:47:46.568781 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:46 crc kubenswrapper[4754]: I0126 16:47:46.568793 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:46 crc kubenswrapper[4754]: I0126 16:47:46.568817 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:46 crc kubenswrapper[4754]: I0126 16:47:46.568833 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:46Z","lastTransitionTime":"2026-01-26T16:47:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:46 crc kubenswrapper[4754]: I0126 16:47:46.672315 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:46 crc kubenswrapper[4754]: I0126 16:47:46.672350 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:46 crc kubenswrapper[4754]: I0126 16:47:46.672360 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:46 crc kubenswrapper[4754]: I0126 16:47:46.672373 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:46 crc kubenswrapper[4754]: I0126 16:47:46.672382 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:46Z","lastTransitionTime":"2026-01-26T16:47:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:46 crc kubenswrapper[4754]: I0126 16:47:46.730502 4754 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-30 07:24:28.205541652 +0000 UTC Jan 26 16:47:46 crc kubenswrapper[4754]: I0126 16:47:46.774797 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:46 crc kubenswrapper[4754]: I0126 16:47:46.774830 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:46 crc kubenswrapper[4754]: I0126 16:47:46.774838 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:46 crc kubenswrapper[4754]: I0126 16:47:46.774852 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:46 crc kubenswrapper[4754]: I0126 16:47:46.774862 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:46Z","lastTransitionTime":"2026-01-26T16:47:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:46 crc kubenswrapper[4754]: I0126 16:47:46.876801 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:46 crc kubenswrapper[4754]: I0126 16:47:46.876846 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:46 crc kubenswrapper[4754]: I0126 16:47:46.876858 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:46 crc kubenswrapper[4754]: I0126 16:47:46.876874 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:46 crc kubenswrapper[4754]: I0126 16:47:46.876885 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:46Z","lastTransitionTime":"2026-01-26T16:47:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:46 crc kubenswrapper[4754]: I0126 16:47:46.979379 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:46 crc kubenswrapper[4754]: I0126 16:47:46.979428 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:46 crc kubenswrapper[4754]: I0126 16:47:46.979444 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:46 crc kubenswrapper[4754]: I0126 16:47:46.979460 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:46 crc kubenswrapper[4754]: I0126 16:47:46.979471 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:46Z","lastTransitionTime":"2026-01-26T16:47:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:47 crc kubenswrapper[4754]: I0126 16:47:47.035233 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-jsbxt_689569c4-93ee-4b82-93fc-9af0f97e6dc3/ovnkube-controller/1.log" Jan 26 16:47:47 crc kubenswrapper[4754]: I0126 16:47:47.038380 4754 scope.go:117] "RemoveContainer" containerID="eaed406e767c0eedc024810541a22ed6ac3b1dcc2b5d41aa007b9ad282eb57fd" Jan 26 16:47:47 crc kubenswrapper[4754]: E0126 16:47:47.038561 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-jsbxt_openshift-ovn-kubernetes(689569c4-93ee-4b82-93fc-9af0f97e6dc3)\"" pod="openshift-ovn-kubernetes/ovnkube-node-jsbxt" podUID="689569c4-93ee-4b82-93fc-9af0f97e6dc3" Jan 26 16:47:47 crc kubenswrapper[4754]: I0126 16:47:47.051154 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:47Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:47 crc kubenswrapper[4754]: I0126 16:47:47.062593 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e459fd8c1093d6f3d0a671751b792b26b7df94c21ced1be1a2250821bcc82d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:47Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:47 crc kubenswrapper[4754]: I0126 16:47:47.073656 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2blzw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca65468e-30c6-4666-962e-cc0de05e67fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4603235bb2af495901926b805fe1c0ffc48824a4394ccf928e7fcd53a673ac3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c25z4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:33Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2blzw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:47Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:47 crc kubenswrapper[4754]: I0126 16:47:47.081757 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:47 crc kubenswrapper[4754]: I0126 16:47:47.081804 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:47 crc kubenswrapper[4754]: I0126 16:47:47.081822 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:47 crc kubenswrapper[4754]: I0126 16:47:47.081843 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:47 crc kubenswrapper[4754]: I0126 16:47:47.081860 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:47Z","lastTransitionTime":"2026-01-26T16:47:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:47 crc kubenswrapper[4754]: I0126 16:47:47.087848 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bwpd6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"351cc6c0-b8e5-4589-ae63-f98e8ab1eca3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1022f101d36873b41bbc1ed8a4a2590347b49f7cce6a172aa1d5bc9f529aed65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3ce0fe080a094316b9456057e347b90a5ff432daa12b9c56d7aea9bee1f72b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3ce0fe080a094316b9456057e347b90a5ff432daa12b9c56d7aea9bee1f72b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1772bb41eb8a36a863d3c691610bcbee5d16b260480b3e9f9d6b083e09e42799\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1772bb41eb8a36a863d3c691610bcbee5d16b260480b3e9f9d6b083e09e42799\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24848d69756024cc664affb9492764ace0638ef611f910f4f2bf86b638cc443d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://24848d69756024cc664affb9492764ace0638ef611f910f4f2bf86b638cc443d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c264fc7c3300a725566acce4464f01d9314779737624401c164d867831fd93bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c264fc7c3300a725566acce4464f01d9314779737624401c164d867831fd93bf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5acfe327eee4aad927ac3a32d24549eff788b14a1c4bdac359b6b56301d884ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5acfe327eee4aad927ac3a32d24549eff788b14a1c4bdac359b6b56301d884ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://63a9b9317494d0e0b33970ef3d461729b10d2a1ad7d3d2bd33f389b184ca56ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://63a9b9317494d0e0b33970ef3d461729b10d2a1ad7d3d2bd33f389b184ca56ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bwpd6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:47Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:47 crc kubenswrapper[4754]: I0126 16:47:47.108606 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jsbxt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"689569c4-93ee-4b82-93fc-9af0f97e6dc3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://033f8b104d4e9ff6d5f87ca998665f12e0c56696b3d30e01ede05f4da540365d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://919572e731813a98c7d8a94879bba931f75c472c7cd681e86d5197ebab43815d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1ac1f49bd564750fcff530a0f76657b0ed41573539acd020a922d19861ef834\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a300a80ad87f46ba1b4fd7ad3cf6c25767fcfe966b93bdd86d729843f53a5cda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31dcab2f9a2d218b0e49dad1df46377ff4521592824a0fbac254d5b8e6d0632e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fe4d9395447ac570eca1338e749dbefea76295d427a12487b63a142a7d6a2e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eaed406e767c0eedc024810541a22ed6ac3b1dcc2b5d41aa007b9ad282eb57fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eaed406e767c0eedc024810541a22ed6ac3b1dcc2b5d41aa007b9ad282eb57fd\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-26T16:47:45Z\\\",\\\"message\\\":\\\".EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0126 16:47:45.693395 6233 handler.go:208] Removed *v1.Node event handler 7\\\\nI0126 16:47:45.694648 6233 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI0126 16:47:45.694956 6233 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI0126 16:47:45.694961 6233 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0126 16:47:45.695763 6233 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0126 16:47:45.695841 6233 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0126 16:47:45.695852 6233 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0126 16:47:45.695893 6233 factory.go:656] Stopping watch factory\\\\nI0126 16:47:45.695913 6233 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0126 16:47:45.695920 6233 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0126 16:47:45.695935 6233 handler.go:208] Removed *v1.Namespace ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:44Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-jsbxt_openshift-ovn-kubernetes(689569c4-93ee-4b82-93fc-9af0f97e6dc3)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://07a43ff7860ff65faf51f4a5e8aeb326f06d6d7222677a92560ae2e3a0c81484\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d9a7bafabd44f8b88a6f62cbe45599af8563bc45d9d7ad3a7669f3d8d11eb9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d9a7bafabd44f8b88a6f62cbe45599af8563bc45d9d7ad3a7669f3d8d11eb9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jsbxt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:47Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:47 crc kubenswrapper[4754]: I0126 16:47:47.123787 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"099af702-abf2-4933-b4c9-63ef2326e468\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23017e73f8450839b4c2c02fa61235ffe4604bdce11556def27f25bcadcf89dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a74507c941b9de4d0401c8f1d53bb79ed11359628d3405cdab7f5474a75c862e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4402182b3e2e35a7d5dd4c44ac53e152915b312b7ccb076c4f2b41b925a6e7f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f772a003566198519fe1849fd14f29d4346004b2e655810662507c95b4d5ffa1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:13Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:47Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:47 crc kubenswrapper[4754]: I0126 16:47:47.138092 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:47Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:47 crc kubenswrapper[4754]: I0126 16:47:47.151427 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89251fa9c64bc25d6d6f355400d7644c099f22aed1cae812fc991a03595f1e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e90508d8a53be9146f9b388750730fdae9ee200306caf5ad393a356390a69ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:47Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:47 crc kubenswrapper[4754]: I0126 16:47:47.162176 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-84lgq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6763d962-0892-4acc-bab9-7a4e93e39a85\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://953c1917fca896c3def3fc2c69f9e3dbdc9ca26fc2b3d40282be71f71759fd68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8zlw6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:33Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-84lgq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:47Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:47 crc kubenswrapper[4754]: I0126 16:47:47.175220 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-skh5n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b619d233-b592-4b05-a0b7-dc094c88471e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9231a4191565e2e0ef53d8a394bb34615c900b601a0dfec24c2d8e5fa7b0b659\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-skh5n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:47Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:47 crc kubenswrapper[4754]: I0126 16:47:47.183543 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:47 crc kubenswrapper[4754]: I0126 16:47:47.183575 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:47 crc kubenswrapper[4754]: I0126 16:47:47.183584 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:47 crc kubenswrapper[4754]: I0126 16:47:47.183597 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:47 crc kubenswrapper[4754]: I0126 16:47:47.183606 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:47Z","lastTransitionTime":"2026-01-26T16:47:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:47 crc kubenswrapper[4754]: I0126 16:47:47.189035 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b89aa081-0a2e-4757-9210-e2a6c228bee9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9c5f40188fda14b14128514fe84bddfe8dcede3d6c560283ccb2ac0c66895fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f45445cbf1d180d6d4dd9588e85bf678f043bfce8cff0812b78ca3510904f9a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://085fa7b8d4fef8b2722fac8c346107699b5a6bd33fbfc74ed35831cb8fdf9dad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c40e018a0979035827963efb9efc48a878a55b9df68ec1a9242d7cff64fe480\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8356f76fbc3a4e636c12d1da5455d3f6767d9b83ce2f2925a11435d7d7261e9f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0126 16:47:26.205208 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0126 16:47:26.206538 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3645934156/tls.crt::/tmp/serving-cert-3645934156/tls.key\\\\\\\"\\\\nI0126 16:47:31.504518 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0126 16:47:31.509366 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0126 16:47:31.509400 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0126 16:47:31.509428 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0126 16:47:31.509436 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0126 16:47:31.515912 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0126 16:47:31.516036 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 16:47:31.516068 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 16:47:31.516096 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0126 16:47:31.516121 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0126 16:47:31.516145 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0126 16:47:31.516170 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0126 16:47:31.516412 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0126 16:47:31.519294 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27b74e2e65a547e0e28fddd4f15feacb17f4d57596eeb8da82df9d39243c85e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22ba7649a1308815bfad52016000987a7a9a5bc214d7ff61e137ba099b20fc6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://22ba7649a1308815bfad52016000987a7a9a5bc214d7ff61e137ba099b20fc6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:13Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:47Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:47 crc kubenswrapper[4754]: I0126 16:47:47.205580 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://205ec256e46a8e08b4b0b0805fb6c5f4966fc3ccef0738df4b97cb404e74da17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:47Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:47 crc kubenswrapper[4754]: I0126 16:47:47.217378 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:47Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:47 crc kubenswrapper[4754]: I0126 16:47:47.226999 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c3718a4-f354-4284-92e0-fdfb45a692bd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b86f46f49163bca7eca9458279029af330dbd4890b03325e06fa3b1bdf030de6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxjcz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fffc03b536bfee3c7bc1bddd500e834e34b391896477a1501e3e06201a19374\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxjcz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-x65wv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:47Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:47 crc kubenswrapper[4754]: I0126 16:47:47.285890 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:47 crc kubenswrapper[4754]: I0126 16:47:47.286156 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:47 crc kubenswrapper[4754]: I0126 16:47:47.286301 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:47 crc kubenswrapper[4754]: I0126 16:47:47.286406 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:47 crc kubenswrapper[4754]: I0126 16:47:47.286504 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:47Z","lastTransitionTime":"2026-01-26T16:47:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:47 crc kubenswrapper[4754]: I0126 16:47:47.388830 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:47 crc kubenswrapper[4754]: I0126 16:47:47.388878 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:47 crc kubenswrapper[4754]: I0126 16:47:47.388889 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:47 crc kubenswrapper[4754]: I0126 16:47:47.388905 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:47 crc kubenswrapper[4754]: I0126 16:47:47.388916 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:47Z","lastTransitionTime":"2026-01-26T16:47:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:47 crc kubenswrapper[4754]: I0126 16:47:47.491474 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:47 crc kubenswrapper[4754]: I0126 16:47:47.491528 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:47 crc kubenswrapper[4754]: I0126 16:47:47.491539 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:47 crc kubenswrapper[4754]: I0126 16:47:47.491555 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:47 crc kubenswrapper[4754]: I0126 16:47:47.491564 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:47Z","lastTransitionTime":"2026-01-26T16:47:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:47 crc kubenswrapper[4754]: I0126 16:47:47.594189 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:47 crc kubenswrapper[4754]: I0126 16:47:47.594247 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:47 crc kubenswrapper[4754]: I0126 16:47:47.594258 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:47 crc kubenswrapper[4754]: I0126 16:47:47.594275 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:47 crc kubenswrapper[4754]: I0126 16:47:47.594285 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:47Z","lastTransitionTime":"2026-01-26T16:47:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:47 crc kubenswrapper[4754]: I0126 16:47:47.679889 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 16:47:47 crc kubenswrapper[4754]: I0126 16:47:47.679972 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 16:47:47 crc kubenswrapper[4754]: I0126 16:47:47.679992 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 16:47:47 crc kubenswrapper[4754]: I0126 16:47:47.680013 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 16:47:47 crc kubenswrapper[4754]: I0126 16:47:47.680039 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 16:47:47 crc kubenswrapper[4754]: E0126 16:47:47.680133 4754 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 26 16:47:47 crc kubenswrapper[4754]: E0126 16:47:47.680178 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-26 16:48:03.680164968 +0000 UTC m=+50.204345402 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 26 16:47:47 crc kubenswrapper[4754]: E0126 16:47:47.680320 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 16:48:03.680304772 +0000 UTC m=+50.204485206 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 16:47:47 crc kubenswrapper[4754]: E0126 16:47:47.680324 4754 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Jan 26 16:47:47 crc kubenswrapper[4754]: E0126 16:47:47.680393 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-26 16:48:03.680382684 +0000 UTC m=+50.204563118 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Jan 26 16:47:47 crc kubenswrapper[4754]: E0126 16:47:47.680430 4754 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 26 16:47:47 crc kubenswrapper[4754]: E0126 16:47:47.680458 4754 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 26 16:47:47 crc kubenswrapper[4754]: E0126 16:47:47.680473 4754 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 26 16:47:47 crc kubenswrapper[4754]: E0126 16:47:47.680554 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-01-26 16:48:03.680533968 +0000 UTC m=+50.204714472 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 26 16:47:47 crc kubenswrapper[4754]: E0126 16:47:47.680760 4754 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 26 16:47:47 crc kubenswrapper[4754]: E0126 16:47:47.680872 4754 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 26 16:47:47 crc kubenswrapper[4754]: E0126 16:47:47.680957 4754 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 26 16:47:47 crc kubenswrapper[4754]: E0126 16:47:47.681091 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-01-26 16:48:03.681074432 +0000 UTC m=+50.205254866 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 26 16:47:47 crc kubenswrapper[4754]: I0126 16:47:47.702166 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:47 crc kubenswrapper[4754]: I0126 16:47:47.702217 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:47 crc kubenswrapper[4754]: I0126 16:47:47.702227 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:47 crc kubenswrapper[4754]: I0126 16:47:47.702243 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:47 crc kubenswrapper[4754]: I0126 16:47:47.702259 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:47Z","lastTransitionTime":"2026-01-26T16:47:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:47 crc kubenswrapper[4754]: I0126 16:47:47.725888 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jspvh"] Jan 26 16:47:47 crc kubenswrapper[4754]: I0126 16:47:47.726596 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jspvh" Jan 26 16:47:47 crc kubenswrapper[4754]: I0126 16:47:47.728798 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Jan 26 16:47:47 crc kubenswrapper[4754]: I0126 16:47:47.728900 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Jan 26 16:47:47 crc kubenswrapper[4754]: I0126 16:47:47.731213 4754 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-24 05:26:28.242367365 +0000 UTC Jan 26 16:47:47 crc kubenswrapper[4754]: I0126 16:47:47.748112 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"099af702-abf2-4933-b4c9-63ef2326e468\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23017e73f8450839b4c2c02fa61235ffe4604bdce11556def27f25bcadcf89dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a74507c941b9de4d0401c8f1d53bb79ed11359628d3405cdab7f5474a75c862e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4402182b3e2e35a7d5dd4c44ac53e152915b312b7ccb076c4f2b41b925a6e7f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f772a003566198519fe1849fd14f29d4346004b2e655810662507c95b4d5ffa1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:13Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:47Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:47 crc kubenswrapper[4754]: I0126 16:47:47.760709 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:47Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:47 crc kubenswrapper[4754]: I0126 16:47:47.767083 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 16:47:47 crc kubenswrapper[4754]: I0126 16:47:47.767110 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 16:47:47 crc kubenswrapper[4754]: I0126 16:47:47.767083 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 16:47:47 crc kubenswrapper[4754]: E0126 16:47:47.767210 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 16:47:47 crc kubenswrapper[4754]: E0126 16:47:47.767323 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 16:47:47 crc kubenswrapper[4754]: E0126 16:47:47.767384 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 16:47:47 crc kubenswrapper[4754]: I0126 16:47:47.773847 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89251fa9c64bc25d6d6f355400d7644c099f22aed1cae812fc991a03595f1e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e90508d8a53be9146f9b388750730fdae9ee200306caf5ad393a356390a69ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:47Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:47 crc kubenswrapper[4754]: I0126 16:47:47.780532 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/5c81aab0-939e-45ce-b19a-89d4818980c1-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-jspvh\" (UID: \"5c81aab0-939e-45ce-b19a-89d4818980c1\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jspvh" Jan 26 16:47:47 crc kubenswrapper[4754]: I0126 16:47:47.780584 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z8bps\" (UniqueName: \"kubernetes.io/projected/5c81aab0-939e-45ce-b19a-89d4818980c1-kube-api-access-z8bps\") pod \"ovnkube-control-plane-749d76644c-jspvh\" (UID: \"5c81aab0-939e-45ce-b19a-89d4818980c1\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jspvh" Jan 26 16:47:47 crc kubenswrapper[4754]: I0126 16:47:47.780614 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/5c81aab0-939e-45ce-b19a-89d4818980c1-env-overrides\") pod \"ovnkube-control-plane-749d76644c-jspvh\" (UID: \"5c81aab0-939e-45ce-b19a-89d4818980c1\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jspvh" Jan 26 16:47:47 crc kubenswrapper[4754]: I0126 16:47:47.780682 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/5c81aab0-939e-45ce-b19a-89d4818980c1-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-jspvh\" (UID: \"5c81aab0-939e-45ce-b19a-89d4818980c1\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jspvh" Jan 26 16:47:47 crc kubenswrapper[4754]: I0126 16:47:47.784823 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-84lgq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6763d962-0892-4acc-bab9-7a4e93e39a85\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://953c1917fca896c3def3fc2c69f9e3dbdc9ca26fc2b3d40282be71f71759fd68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8zlw6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:33Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-84lgq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:47Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:47 crc kubenswrapper[4754]: I0126 16:47:47.799888 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-skh5n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b619d233-b592-4b05-a0b7-dc094c88471e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9231a4191565e2e0ef53d8a394bb34615c900b601a0dfec24c2d8e5fa7b0b659\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-skh5n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:47Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:47 crc kubenswrapper[4754]: I0126 16:47:47.804298 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:47 crc kubenswrapper[4754]: I0126 16:47:47.804324 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:47 crc kubenswrapper[4754]: I0126 16:47:47.804334 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:47 crc kubenswrapper[4754]: I0126 16:47:47.804349 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:47 crc kubenswrapper[4754]: I0126 16:47:47.804357 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:47Z","lastTransitionTime":"2026-01-26T16:47:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:47 crc kubenswrapper[4754]: I0126 16:47:47.811924 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jspvh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5c81aab0-939e-45ce-b19a-89d4818980c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8bps\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8bps\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-jspvh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:47Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:47 crc kubenswrapper[4754]: I0126 16:47:47.826794 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b89aa081-0a2e-4757-9210-e2a6c228bee9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9c5f40188fda14b14128514fe84bddfe8dcede3d6c560283ccb2ac0c66895fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f45445cbf1d180d6d4dd9588e85bf678f043bfce8cff0812b78ca3510904f9a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://085fa7b8d4fef8b2722fac8c346107699b5a6bd33fbfc74ed35831cb8fdf9dad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c40e018a0979035827963efb9efc48a878a55b9df68ec1a9242d7cff64fe480\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8356f76fbc3a4e636c12d1da5455d3f6767d9b83ce2f2925a11435d7d7261e9f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0126 16:47:26.205208 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0126 16:47:26.206538 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3645934156/tls.crt::/tmp/serving-cert-3645934156/tls.key\\\\\\\"\\\\nI0126 16:47:31.504518 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0126 16:47:31.509366 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0126 16:47:31.509400 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0126 16:47:31.509428 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0126 16:47:31.509436 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0126 16:47:31.515912 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0126 16:47:31.516036 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 16:47:31.516068 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 16:47:31.516096 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0126 16:47:31.516121 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0126 16:47:31.516145 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0126 16:47:31.516170 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0126 16:47:31.516412 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0126 16:47:31.519294 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27b74e2e65a547e0e28fddd4f15feacb17f4d57596eeb8da82df9d39243c85e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22ba7649a1308815bfad52016000987a7a9a5bc214d7ff61e137ba099b20fc6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://22ba7649a1308815bfad52016000987a7a9a5bc214d7ff61e137ba099b20fc6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:13Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:47Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:47 crc kubenswrapper[4754]: I0126 16:47:47.840490 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://205ec256e46a8e08b4b0b0805fb6c5f4966fc3ccef0738df4b97cb404e74da17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:47Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:47 crc kubenswrapper[4754]: I0126 16:47:47.854342 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:47Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:47 crc kubenswrapper[4754]: I0126 16:47:47.865933 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c3718a4-f354-4284-92e0-fdfb45a692bd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b86f46f49163bca7eca9458279029af330dbd4890b03325e06fa3b1bdf030de6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxjcz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fffc03b536bfee3c7bc1bddd500e834e34b391896477a1501e3e06201a19374\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxjcz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-x65wv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:47Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:47 crc kubenswrapper[4754]: I0126 16:47:47.879158 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:47Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:47 crc kubenswrapper[4754]: I0126 16:47:47.881524 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z8bps\" (UniqueName: \"kubernetes.io/projected/5c81aab0-939e-45ce-b19a-89d4818980c1-kube-api-access-z8bps\") pod \"ovnkube-control-plane-749d76644c-jspvh\" (UID: \"5c81aab0-939e-45ce-b19a-89d4818980c1\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jspvh" Jan 26 16:47:47 crc kubenswrapper[4754]: I0126 16:47:47.881564 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/5c81aab0-939e-45ce-b19a-89d4818980c1-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-jspvh\" (UID: \"5c81aab0-939e-45ce-b19a-89d4818980c1\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jspvh" Jan 26 16:47:47 crc kubenswrapper[4754]: I0126 16:47:47.881586 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/5c81aab0-939e-45ce-b19a-89d4818980c1-env-overrides\") pod \"ovnkube-control-plane-749d76644c-jspvh\" (UID: \"5c81aab0-939e-45ce-b19a-89d4818980c1\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jspvh" Jan 26 16:47:47 crc kubenswrapper[4754]: I0126 16:47:47.881628 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/5c81aab0-939e-45ce-b19a-89d4818980c1-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-jspvh\" (UID: \"5c81aab0-939e-45ce-b19a-89d4818980c1\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jspvh" Jan 26 16:47:47 crc kubenswrapper[4754]: I0126 16:47:47.882494 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/5c81aab0-939e-45ce-b19a-89d4818980c1-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-jspvh\" (UID: \"5c81aab0-939e-45ce-b19a-89d4818980c1\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jspvh" Jan 26 16:47:47 crc kubenswrapper[4754]: I0126 16:47:47.882848 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/5c81aab0-939e-45ce-b19a-89d4818980c1-env-overrides\") pod \"ovnkube-control-plane-749d76644c-jspvh\" (UID: \"5c81aab0-939e-45ce-b19a-89d4818980c1\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jspvh" Jan 26 16:47:47 crc kubenswrapper[4754]: I0126 16:47:47.887842 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/5c81aab0-939e-45ce-b19a-89d4818980c1-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-jspvh\" (UID: \"5c81aab0-939e-45ce-b19a-89d4818980c1\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jspvh" Jan 26 16:47:47 crc kubenswrapper[4754]: I0126 16:47:47.890660 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e459fd8c1093d6f3d0a671751b792b26b7df94c21ced1be1a2250821bcc82d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:47Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:47 crc kubenswrapper[4754]: I0126 16:47:47.899410 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z8bps\" (UniqueName: \"kubernetes.io/projected/5c81aab0-939e-45ce-b19a-89d4818980c1-kube-api-access-z8bps\") pod \"ovnkube-control-plane-749d76644c-jspvh\" (UID: \"5c81aab0-939e-45ce-b19a-89d4818980c1\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jspvh" Jan 26 16:47:47 crc kubenswrapper[4754]: I0126 16:47:47.901089 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2blzw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca65468e-30c6-4666-962e-cc0de05e67fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4603235bb2af495901926b805fe1c0ffc48824a4394ccf928e7fcd53a673ac3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c25z4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:33Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2blzw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:47Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:47 crc kubenswrapper[4754]: I0126 16:47:47.907432 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:47 crc kubenswrapper[4754]: I0126 16:47:47.907464 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:47 crc kubenswrapper[4754]: I0126 16:47:47.907476 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:47 crc kubenswrapper[4754]: I0126 16:47:47.907491 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:47 crc kubenswrapper[4754]: I0126 16:47:47.907503 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:47Z","lastTransitionTime":"2026-01-26T16:47:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:47 crc kubenswrapper[4754]: I0126 16:47:47.916602 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bwpd6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"351cc6c0-b8e5-4589-ae63-f98e8ab1eca3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1022f101d36873b41bbc1ed8a4a2590347b49f7cce6a172aa1d5bc9f529aed65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3ce0fe080a094316b9456057e347b90a5ff432daa12b9c56d7aea9bee1f72b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3ce0fe080a094316b9456057e347b90a5ff432daa12b9c56d7aea9bee1f72b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1772bb41eb8a36a863d3c691610bcbee5d16b260480b3e9f9d6b083e09e42799\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1772bb41eb8a36a863d3c691610bcbee5d16b260480b3e9f9d6b083e09e42799\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24848d69756024cc664affb9492764ace0638ef611f910f4f2bf86b638cc443d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://24848d69756024cc664affb9492764ace0638ef611f910f4f2bf86b638cc443d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c264fc7c3300a725566acce4464f01d9314779737624401c164d867831fd93bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c264fc7c3300a725566acce4464f01d9314779737624401c164d867831fd93bf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5acfe327eee4aad927ac3a32d24549eff788b14a1c4bdac359b6b56301d884ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5acfe327eee4aad927ac3a32d24549eff788b14a1c4bdac359b6b56301d884ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://63a9b9317494d0e0b33970ef3d461729b10d2a1ad7d3d2bd33f389b184ca56ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://63a9b9317494d0e0b33970ef3d461729b10d2a1ad7d3d2bd33f389b184ca56ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bwpd6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:47Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:47 crc kubenswrapper[4754]: I0126 16:47:47.937718 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jsbxt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"689569c4-93ee-4b82-93fc-9af0f97e6dc3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://033f8b104d4e9ff6d5f87ca998665f12e0c56696b3d30e01ede05f4da540365d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://919572e731813a98c7d8a94879bba931f75c472c7cd681e86d5197ebab43815d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1ac1f49bd564750fcff530a0f76657b0ed41573539acd020a922d19861ef834\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a300a80ad87f46ba1b4fd7ad3cf6c25767fcfe966b93bdd86d729843f53a5cda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31dcab2f9a2d218b0e49dad1df46377ff4521592824a0fbac254d5b8e6d0632e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fe4d9395447ac570eca1338e749dbefea76295d427a12487b63a142a7d6a2e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eaed406e767c0eedc024810541a22ed6ac3b1dcc2b5d41aa007b9ad282eb57fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eaed406e767c0eedc024810541a22ed6ac3b1dcc2b5d41aa007b9ad282eb57fd\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-26T16:47:45Z\\\",\\\"message\\\":\\\".EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0126 16:47:45.693395 6233 handler.go:208] Removed *v1.Node event handler 7\\\\nI0126 16:47:45.694648 6233 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI0126 16:47:45.694956 6233 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI0126 16:47:45.694961 6233 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0126 16:47:45.695763 6233 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0126 16:47:45.695841 6233 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0126 16:47:45.695852 6233 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0126 16:47:45.695893 6233 factory.go:656] Stopping watch factory\\\\nI0126 16:47:45.695913 6233 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0126 16:47:45.695920 6233 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0126 16:47:45.695935 6233 handler.go:208] Removed *v1.Namespace ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:44Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-jsbxt_openshift-ovn-kubernetes(689569c4-93ee-4b82-93fc-9af0f97e6dc3)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://07a43ff7860ff65faf51f4a5e8aeb326f06d6d7222677a92560ae2e3a0c81484\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d9a7bafabd44f8b88a6f62cbe45599af8563bc45d9d7ad3a7669f3d8d11eb9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d9a7bafabd44f8b88a6f62cbe45599af8563bc45d9d7ad3a7669f3d8d11eb9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jsbxt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:47Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:48 crc kubenswrapper[4754]: I0126 16:47:48.009878 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:48 crc kubenswrapper[4754]: I0126 16:47:48.009921 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:48 crc kubenswrapper[4754]: I0126 16:47:48.009954 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:48 crc kubenswrapper[4754]: I0126 16:47:48.009970 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:48 crc kubenswrapper[4754]: I0126 16:47:48.009978 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:48Z","lastTransitionTime":"2026-01-26T16:47:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:48 crc kubenswrapper[4754]: I0126 16:47:48.039554 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jspvh" Jan 26 16:47:48 crc kubenswrapper[4754]: W0126 16:47:48.052269 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5c81aab0_939e_45ce_b19a_89d4818980c1.slice/crio-df4451a252dd751f1aa64c359535c3580366705cb81e8c5c023f08598da4a773 WatchSource:0}: Error finding container df4451a252dd751f1aa64c359535c3580366705cb81e8c5c023f08598da4a773: Status 404 returned error can't find the container with id df4451a252dd751f1aa64c359535c3580366705cb81e8c5c023f08598da4a773 Jan 26 16:47:48 crc kubenswrapper[4754]: I0126 16:47:48.112487 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:48 crc kubenswrapper[4754]: I0126 16:47:48.112536 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:48 crc kubenswrapper[4754]: I0126 16:47:48.112548 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:48 crc kubenswrapper[4754]: I0126 16:47:48.112565 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:48 crc kubenswrapper[4754]: I0126 16:47:48.112577 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:48Z","lastTransitionTime":"2026-01-26T16:47:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:48 crc kubenswrapper[4754]: I0126 16:47:48.217077 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:48 crc kubenswrapper[4754]: I0126 16:47:48.217142 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:48 crc kubenswrapper[4754]: I0126 16:47:48.217154 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:48 crc kubenswrapper[4754]: I0126 16:47:48.217178 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:48 crc kubenswrapper[4754]: I0126 16:47:48.217191 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:48Z","lastTransitionTime":"2026-01-26T16:47:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:48 crc kubenswrapper[4754]: I0126 16:47:48.319721 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:48 crc kubenswrapper[4754]: I0126 16:47:48.319754 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:48 crc kubenswrapper[4754]: I0126 16:47:48.319765 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:48 crc kubenswrapper[4754]: I0126 16:47:48.319782 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:48 crc kubenswrapper[4754]: I0126 16:47:48.319793 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:48Z","lastTransitionTime":"2026-01-26T16:47:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:48 crc kubenswrapper[4754]: I0126 16:47:48.422016 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:48 crc kubenswrapper[4754]: I0126 16:47:48.422049 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:48 crc kubenswrapper[4754]: I0126 16:47:48.422058 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:48 crc kubenswrapper[4754]: I0126 16:47:48.422072 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:48 crc kubenswrapper[4754]: I0126 16:47:48.422081 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:48Z","lastTransitionTime":"2026-01-26T16:47:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:48 crc kubenswrapper[4754]: I0126 16:47:48.463796 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-t4ps6"] Jan 26 16:47:48 crc kubenswrapper[4754]: I0126 16:47:48.464287 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-t4ps6" Jan 26 16:47:48 crc kubenswrapper[4754]: E0126 16:47:48.464390 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-t4ps6" podUID="d45179f0-9f81-4d77-8805-6b05bdb32f8f" Jan 26 16:47:48 crc kubenswrapper[4754]: I0126 16:47:48.482126 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jsbxt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"689569c4-93ee-4b82-93fc-9af0f97e6dc3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://033f8b104d4e9ff6d5f87ca998665f12e0c56696b3d30e01ede05f4da540365d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://919572e731813a98c7d8a94879bba931f75c472c7cd681e86d5197ebab43815d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1ac1f49bd564750fcff530a0f76657b0ed41573539acd020a922d19861ef834\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a300a80ad87f46ba1b4fd7ad3cf6c25767fcfe966b93bdd86d729843f53a5cda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31dcab2f9a2d218b0e49dad1df46377ff4521592824a0fbac254d5b8e6d0632e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fe4d9395447ac570eca1338e749dbefea76295d427a12487b63a142a7d6a2e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eaed406e767c0eedc024810541a22ed6ac3b1dcc2b5d41aa007b9ad282eb57fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eaed406e767c0eedc024810541a22ed6ac3b1dcc2b5d41aa007b9ad282eb57fd\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-26T16:47:45Z\\\",\\\"message\\\":\\\".EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0126 16:47:45.693395 6233 handler.go:208] Removed *v1.Node event handler 7\\\\nI0126 16:47:45.694648 6233 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI0126 16:47:45.694956 6233 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI0126 16:47:45.694961 6233 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0126 16:47:45.695763 6233 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0126 16:47:45.695841 6233 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0126 16:47:45.695852 6233 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0126 16:47:45.695893 6233 factory.go:656] Stopping watch factory\\\\nI0126 16:47:45.695913 6233 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0126 16:47:45.695920 6233 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0126 16:47:45.695935 6233 handler.go:208] Removed *v1.Namespace ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:44Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-jsbxt_openshift-ovn-kubernetes(689569c4-93ee-4b82-93fc-9af0f97e6dc3)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://07a43ff7860ff65faf51f4a5e8aeb326f06d6d7222677a92560ae2e3a0c81484\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d9a7bafabd44f8b88a6f62cbe45599af8563bc45d9d7ad3a7669f3d8d11eb9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d9a7bafabd44f8b88a6f62cbe45599af8563bc45d9d7ad3a7669f3d8d11eb9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jsbxt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:48Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:48 crc kubenswrapper[4754]: I0126 16:47:48.487204 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vk8tg\" (UniqueName: \"kubernetes.io/projected/d45179f0-9f81-4d77-8805-6b05bdb32f8f-kube-api-access-vk8tg\") pod \"network-metrics-daemon-t4ps6\" (UID: \"d45179f0-9f81-4d77-8805-6b05bdb32f8f\") " pod="openshift-multus/network-metrics-daemon-t4ps6" Jan 26 16:47:48 crc kubenswrapper[4754]: I0126 16:47:48.487257 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d45179f0-9f81-4d77-8805-6b05bdb32f8f-metrics-certs\") pod \"network-metrics-daemon-t4ps6\" (UID: \"d45179f0-9f81-4d77-8805-6b05bdb32f8f\") " pod="openshift-multus/network-metrics-daemon-t4ps6" Jan 26 16:47:48 crc kubenswrapper[4754]: I0126 16:47:48.494967 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:48Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:48 crc kubenswrapper[4754]: I0126 16:47:48.508094 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e459fd8c1093d6f3d0a671751b792b26b7df94c21ced1be1a2250821bcc82d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:48Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:48 crc kubenswrapper[4754]: I0126 16:47:48.519408 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2blzw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca65468e-30c6-4666-962e-cc0de05e67fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4603235bb2af495901926b805fe1c0ffc48824a4394ccf928e7fcd53a673ac3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c25z4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:33Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2blzw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:48Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:48 crc kubenswrapper[4754]: I0126 16:47:48.524068 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:48 crc kubenswrapper[4754]: I0126 16:47:48.524105 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:48 crc kubenswrapper[4754]: I0126 16:47:48.524115 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:48 crc kubenswrapper[4754]: I0126 16:47:48.524130 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:48 crc kubenswrapper[4754]: I0126 16:47:48.524140 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:48Z","lastTransitionTime":"2026-01-26T16:47:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:48 crc kubenswrapper[4754]: I0126 16:47:48.533895 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bwpd6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"351cc6c0-b8e5-4589-ae63-f98e8ab1eca3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1022f101d36873b41bbc1ed8a4a2590347b49f7cce6a172aa1d5bc9f529aed65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3ce0fe080a094316b9456057e347b90a5ff432daa12b9c56d7aea9bee1f72b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3ce0fe080a094316b9456057e347b90a5ff432daa12b9c56d7aea9bee1f72b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1772bb41eb8a36a863d3c691610bcbee5d16b260480b3e9f9d6b083e09e42799\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1772bb41eb8a36a863d3c691610bcbee5d16b260480b3e9f9d6b083e09e42799\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24848d69756024cc664affb9492764ace0638ef611f910f4f2bf86b638cc443d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://24848d69756024cc664affb9492764ace0638ef611f910f4f2bf86b638cc443d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c264fc7c3300a725566acce4464f01d9314779737624401c164d867831fd93bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c264fc7c3300a725566acce4464f01d9314779737624401c164d867831fd93bf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5acfe327eee4aad927ac3a32d24549eff788b14a1c4bdac359b6b56301d884ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5acfe327eee4aad927ac3a32d24549eff788b14a1c4bdac359b6b56301d884ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://63a9b9317494d0e0b33970ef3d461729b10d2a1ad7d3d2bd33f389b184ca56ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://63a9b9317494d0e0b33970ef3d461729b10d2a1ad7d3d2bd33f389b184ca56ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bwpd6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:48Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:48 crc kubenswrapper[4754]: I0126 16:47:48.546278 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-skh5n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b619d233-b592-4b05-a0b7-dc094c88471e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9231a4191565e2e0ef53d8a394bb34615c900b601a0dfec24c2d8e5fa7b0b659\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-skh5n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:48Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:48 crc kubenswrapper[4754]: I0126 16:47:48.559240 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jspvh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5c81aab0-939e-45ce-b19a-89d4818980c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8bps\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8bps\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-jspvh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:48Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:48 crc kubenswrapper[4754]: I0126 16:47:48.570986 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"099af702-abf2-4933-b4c9-63ef2326e468\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23017e73f8450839b4c2c02fa61235ffe4604bdce11556def27f25bcadcf89dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a74507c941b9de4d0401c8f1d53bb79ed11359628d3405cdab7f5474a75c862e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4402182b3e2e35a7d5dd4c44ac53e152915b312b7ccb076c4f2b41b925a6e7f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f772a003566198519fe1849fd14f29d4346004b2e655810662507c95b4d5ffa1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:13Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:48Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:48 crc kubenswrapper[4754]: I0126 16:47:48.583385 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:48Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:48 crc kubenswrapper[4754]: I0126 16:47:48.588140 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vk8tg\" (UniqueName: \"kubernetes.io/projected/d45179f0-9f81-4d77-8805-6b05bdb32f8f-kube-api-access-vk8tg\") pod \"network-metrics-daemon-t4ps6\" (UID: \"d45179f0-9f81-4d77-8805-6b05bdb32f8f\") " pod="openshift-multus/network-metrics-daemon-t4ps6" Jan 26 16:47:48 crc kubenswrapper[4754]: I0126 16:47:48.588188 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d45179f0-9f81-4d77-8805-6b05bdb32f8f-metrics-certs\") pod \"network-metrics-daemon-t4ps6\" (UID: \"d45179f0-9f81-4d77-8805-6b05bdb32f8f\") " pod="openshift-multus/network-metrics-daemon-t4ps6" Jan 26 16:47:48 crc kubenswrapper[4754]: E0126 16:47:48.588300 4754 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Jan 26 16:47:48 crc kubenswrapper[4754]: E0126 16:47:48.588354 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d45179f0-9f81-4d77-8805-6b05bdb32f8f-metrics-certs podName:d45179f0-9f81-4d77-8805-6b05bdb32f8f nodeName:}" failed. No retries permitted until 2026-01-26 16:47:49.088341594 +0000 UTC m=+35.612522028 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/d45179f0-9f81-4d77-8805-6b05bdb32f8f-metrics-certs") pod "network-metrics-daemon-t4ps6" (UID: "d45179f0-9f81-4d77-8805-6b05bdb32f8f") : object "openshift-multus"/"metrics-daemon-secret" not registered Jan 26 16:47:48 crc kubenswrapper[4754]: I0126 16:47:48.595624 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89251fa9c64bc25d6d6f355400d7644c099f22aed1cae812fc991a03595f1e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e90508d8a53be9146f9b388750730fdae9ee200306caf5ad393a356390a69ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:48Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:48 crc kubenswrapper[4754]: I0126 16:47:48.604053 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vk8tg\" (UniqueName: \"kubernetes.io/projected/d45179f0-9f81-4d77-8805-6b05bdb32f8f-kube-api-access-vk8tg\") pod \"network-metrics-daemon-t4ps6\" (UID: \"d45179f0-9f81-4d77-8805-6b05bdb32f8f\") " pod="openshift-multus/network-metrics-daemon-t4ps6" Jan 26 16:47:48 crc kubenswrapper[4754]: I0126 16:47:48.605304 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-84lgq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6763d962-0892-4acc-bab9-7a4e93e39a85\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://953c1917fca896c3def3fc2c69f9e3dbdc9ca26fc2b3d40282be71f71759fd68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8zlw6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:33Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-84lgq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:48Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:48 crc kubenswrapper[4754]: I0126 16:47:48.616998 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b89aa081-0a2e-4757-9210-e2a6c228bee9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9c5f40188fda14b14128514fe84bddfe8dcede3d6c560283ccb2ac0c66895fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f45445cbf1d180d6d4dd9588e85bf678f043bfce8cff0812b78ca3510904f9a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://085fa7b8d4fef8b2722fac8c346107699b5a6bd33fbfc74ed35831cb8fdf9dad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c40e018a0979035827963efb9efc48a878a55b9df68ec1a9242d7cff64fe480\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8356f76fbc3a4e636c12d1da5455d3f6767d9b83ce2f2925a11435d7d7261e9f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0126 16:47:26.205208 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0126 16:47:26.206538 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3645934156/tls.crt::/tmp/serving-cert-3645934156/tls.key\\\\\\\"\\\\nI0126 16:47:31.504518 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0126 16:47:31.509366 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0126 16:47:31.509400 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0126 16:47:31.509428 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0126 16:47:31.509436 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0126 16:47:31.515912 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0126 16:47:31.516036 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 16:47:31.516068 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 16:47:31.516096 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0126 16:47:31.516121 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0126 16:47:31.516145 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0126 16:47:31.516170 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0126 16:47:31.516412 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0126 16:47:31.519294 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27b74e2e65a547e0e28fddd4f15feacb17f4d57596eeb8da82df9d39243c85e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22ba7649a1308815bfad52016000987a7a9a5bc214d7ff61e137ba099b20fc6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://22ba7649a1308815bfad52016000987a7a9a5bc214d7ff61e137ba099b20fc6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:13Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:48Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:48 crc kubenswrapper[4754]: I0126 16:47:48.626220 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:48 crc kubenswrapper[4754]: I0126 16:47:48.626253 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:48 crc kubenswrapper[4754]: I0126 16:47:48.626262 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:48 crc kubenswrapper[4754]: I0126 16:47:48.626273 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:48 crc kubenswrapper[4754]: I0126 16:47:48.626281 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:48Z","lastTransitionTime":"2026-01-26T16:47:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:48 crc kubenswrapper[4754]: I0126 16:47:48.629411 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://205ec256e46a8e08b4b0b0805fb6c5f4966fc3ccef0738df4b97cb404e74da17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:48Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:48 crc kubenswrapper[4754]: I0126 16:47:48.640463 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:48Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:48 crc kubenswrapper[4754]: I0126 16:47:48.652175 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c3718a4-f354-4284-92e0-fdfb45a692bd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b86f46f49163bca7eca9458279029af330dbd4890b03325e06fa3b1bdf030de6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxjcz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fffc03b536bfee3c7bc1bddd500e834e34b391896477a1501e3e06201a19374\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxjcz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-x65wv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:48Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:48 crc kubenswrapper[4754]: I0126 16:47:48.662994 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-t4ps6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d45179f0-9f81-4d77-8805-6b05bdb32f8f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vk8tg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vk8tg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:48Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-t4ps6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:48Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:48 crc kubenswrapper[4754]: I0126 16:47:48.727838 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:48 crc kubenswrapper[4754]: I0126 16:47:48.727875 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:48 crc kubenswrapper[4754]: I0126 16:47:48.727887 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:48 crc kubenswrapper[4754]: I0126 16:47:48.727902 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:48 crc kubenswrapper[4754]: I0126 16:47:48.727913 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:48Z","lastTransitionTime":"2026-01-26T16:47:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:48 crc kubenswrapper[4754]: I0126 16:47:48.732613 4754 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-09 01:37:58.151367719 +0000 UTC Jan 26 16:47:48 crc kubenswrapper[4754]: I0126 16:47:48.829638 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:48 crc kubenswrapper[4754]: I0126 16:47:48.829703 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:48 crc kubenswrapper[4754]: I0126 16:47:48.829716 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:48 crc kubenswrapper[4754]: I0126 16:47:48.829733 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:48 crc kubenswrapper[4754]: I0126 16:47:48.829744 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:48Z","lastTransitionTime":"2026-01-26T16:47:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:48 crc kubenswrapper[4754]: I0126 16:47:48.932313 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:48 crc kubenswrapper[4754]: I0126 16:47:48.932355 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:48 crc kubenswrapper[4754]: I0126 16:47:48.932366 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:48 crc kubenswrapper[4754]: I0126 16:47:48.932382 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:48 crc kubenswrapper[4754]: I0126 16:47:48.932392 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:48Z","lastTransitionTime":"2026-01-26T16:47:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:49 crc kubenswrapper[4754]: I0126 16:47:49.034957 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:49 crc kubenswrapper[4754]: I0126 16:47:49.034999 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:49 crc kubenswrapper[4754]: I0126 16:47:49.035010 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:49 crc kubenswrapper[4754]: I0126 16:47:49.035033 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:49 crc kubenswrapper[4754]: I0126 16:47:49.035045 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:49Z","lastTransitionTime":"2026-01-26T16:47:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:49 crc kubenswrapper[4754]: I0126 16:47:49.045403 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jspvh" event={"ID":"5c81aab0-939e-45ce-b19a-89d4818980c1","Type":"ContainerStarted","Data":"05fd57780c44cd127859dc8b163d20eafbd8b06f3348cf545ef41021ce6bc526"} Jan 26 16:47:49 crc kubenswrapper[4754]: I0126 16:47:49.045456 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jspvh" event={"ID":"5c81aab0-939e-45ce-b19a-89d4818980c1","Type":"ContainerStarted","Data":"356dc54150a495dd6f35e0667c82b807b712b07200d8a0adbb1c92fe3f7af8f4"} Jan 26 16:47:49 crc kubenswrapper[4754]: I0126 16:47:49.045471 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jspvh" event={"ID":"5c81aab0-939e-45ce-b19a-89d4818980c1","Type":"ContainerStarted","Data":"df4451a252dd751f1aa64c359535c3580366705cb81e8c5c023f08598da4a773"} Jan 26 16:47:49 crc kubenswrapper[4754]: I0126 16:47:49.059272 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e459fd8c1093d6f3d0a671751b792b26b7df94c21ced1be1a2250821bcc82d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:49Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:49 crc kubenswrapper[4754]: I0126 16:47:49.073445 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2blzw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca65468e-30c6-4666-962e-cc0de05e67fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4603235bb2af495901926b805fe1c0ffc48824a4394ccf928e7fcd53a673ac3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c25z4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:33Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2blzw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:49Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:49 crc kubenswrapper[4754]: I0126 16:47:49.089206 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bwpd6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"351cc6c0-b8e5-4589-ae63-f98e8ab1eca3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1022f101d36873b41bbc1ed8a4a2590347b49f7cce6a172aa1d5bc9f529aed65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3ce0fe080a094316b9456057e347b90a5ff432daa12b9c56d7aea9bee1f72b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3ce0fe080a094316b9456057e347b90a5ff432daa12b9c56d7aea9bee1f72b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1772bb41eb8a36a863d3c691610bcbee5d16b260480b3e9f9d6b083e09e42799\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1772bb41eb8a36a863d3c691610bcbee5d16b260480b3e9f9d6b083e09e42799\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24848d69756024cc664affb9492764ace0638ef611f910f4f2bf86b638cc443d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://24848d69756024cc664affb9492764ace0638ef611f910f4f2bf86b638cc443d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c264fc7c3300a725566acce4464f01d9314779737624401c164d867831fd93bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c264fc7c3300a725566acce4464f01d9314779737624401c164d867831fd93bf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5acfe327eee4aad927ac3a32d24549eff788b14a1c4bdac359b6b56301d884ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5acfe327eee4aad927ac3a32d24549eff788b14a1c4bdac359b6b56301d884ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://63a9b9317494d0e0b33970ef3d461729b10d2a1ad7d3d2bd33f389b184ca56ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://63a9b9317494d0e0b33970ef3d461729b10d2a1ad7d3d2bd33f389b184ca56ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bwpd6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:49Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:49 crc kubenswrapper[4754]: I0126 16:47:49.093207 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d45179f0-9f81-4d77-8805-6b05bdb32f8f-metrics-certs\") pod \"network-metrics-daemon-t4ps6\" (UID: \"d45179f0-9f81-4d77-8805-6b05bdb32f8f\") " pod="openshift-multus/network-metrics-daemon-t4ps6" Jan 26 16:47:49 crc kubenswrapper[4754]: E0126 16:47:49.093793 4754 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Jan 26 16:47:49 crc kubenswrapper[4754]: E0126 16:47:49.093850 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d45179f0-9f81-4d77-8805-6b05bdb32f8f-metrics-certs podName:d45179f0-9f81-4d77-8805-6b05bdb32f8f nodeName:}" failed. No retries permitted until 2026-01-26 16:47:50.093834004 +0000 UTC m=+36.618014528 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/d45179f0-9f81-4d77-8805-6b05bdb32f8f-metrics-certs") pod "network-metrics-daemon-t4ps6" (UID: "d45179f0-9f81-4d77-8805-6b05bdb32f8f") : object "openshift-multus"/"metrics-daemon-secret" not registered Jan 26 16:47:49 crc kubenswrapper[4754]: I0126 16:47:49.107274 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jsbxt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"689569c4-93ee-4b82-93fc-9af0f97e6dc3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://033f8b104d4e9ff6d5f87ca998665f12e0c56696b3d30e01ede05f4da540365d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://919572e731813a98c7d8a94879bba931f75c472c7cd681e86d5197ebab43815d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1ac1f49bd564750fcff530a0f76657b0ed41573539acd020a922d19861ef834\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a300a80ad87f46ba1b4fd7ad3cf6c25767fcfe966b93bdd86d729843f53a5cda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31dcab2f9a2d218b0e49dad1df46377ff4521592824a0fbac254d5b8e6d0632e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fe4d9395447ac570eca1338e749dbefea76295d427a12487b63a142a7d6a2e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eaed406e767c0eedc024810541a22ed6ac3b1dcc2b5d41aa007b9ad282eb57fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eaed406e767c0eedc024810541a22ed6ac3b1dcc2b5d41aa007b9ad282eb57fd\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-26T16:47:45Z\\\",\\\"message\\\":\\\".EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0126 16:47:45.693395 6233 handler.go:208] Removed *v1.Node event handler 7\\\\nI0126 16:47:45.694648 6233 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI0126 16:47:45.694956 6233 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI0126 16:47:45.694961 6233 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0126 16:47:45.695763 6233 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0126 16:47:45.695841 6233 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0126 16:47:45.695852 6233 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0126 16:47:45.695893 6233 factory.go:656] Stopping watch factory\\\\nI0126 16:47:45.695913 6233 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0126 16:47:45.695920 6233 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0126 16:47:45.695935 6233 handler.go:208] Removed *v1.Namespace ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:44Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-jsbxt_openshift-ovn-kubernetes(689569c4-93ee-4b82-93fc-9af0f97e6dc3)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://07a43ff7860ff65faf51f4a5e8aeb326f06d6d7222677a92560ae2e3a0c81484\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d9a7bafabd44f8b88a6f62cbe45599af8563bc45d9d7ad3a7669f3d8d11eb9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d9a7bafabd44f8b88a6f62cbe45599af8563bc45d9d7ad3a7669f3d8d11eb9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jsbxt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:49Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:49 crc kubenswrapper[4754]: I0126 16:47:49.119604 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:49Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:49 crc kubenswrapper[4754]: I0126 16:47:49.131402 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:49Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:49 crc kubenswrapper[4754]: I0126 16:47:49.137357 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:49 crc kubenswrapper[4754]: I0126 16:47:49.137392 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:49 crc kubenswrapper[4754]: I0126 16:47:49.137403 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:49 crc kubenswrapper[4754]: I0126 16:47:49.137418 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:49 crc kubenswrapper[4754]: I0126 16:47:49.137430 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:49Z","lastTransitionTime":"2026-01-26T16:47:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:49 crc kubenswrapper[4754]: I0126 16:47:49.145000 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89251fa9c64bc25d6d6f355400d7644c099f22aed1cae812fc991a03595f1e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e90508d8a53be9146f9b388750730fdae9ee200306caf5ad393a356390a69ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:49Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:49 crc kubenswrapper[4754]: I0126 16:47:49.156443 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-84lgq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6763d962-0892-4acc-bab9-7a4e93e39a85\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://953c1917fca896c3def3fc2c69f9e3dbdc9ca26fc2b3d40282be71f71759fd68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8zlw6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:33Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-84lgq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:49Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:49 crc kubenswrapper[4754]: I0126 16:47:49.169403 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-skh5n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b619d233-b592-4b05-a0b7-dc094c88471e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9231a4191565e2e0ef53d8a394bb34615c900b601a0dfec24c2d8e5fa7b0b659\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-skh5n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:49Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:49 crc kubenswrapper[4754]: I0126 16:47:49.182115 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jspvh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5c81aab0-939e-45ce-b19a-89d4818980c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://356dc54150a495dd6f35e0667c82b807b712b07200d8a0adbb1c92fe3f7af8f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8bps\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://05fd57780c44cd127859dc8b163d20eafbd8b06f3348cf545ef41021ce6bc526\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8bps\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-jspvh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:49Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:49 crc kubenswrapper[4754]: I0126 16:47:49.196121 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"099af702-abf2-4933-b4c9-63ef2326e468\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23017e73f8450839b4c2c02fa61235ffe4604bdce11556def27f25bcadcf89dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a74507c941b9de4d0401c8f1d53bb79ed11359628d3405cdab7f5474a75c862e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4402182b3e2e35a7d5dd4c44ac53e152915b312b7ccb076c4f2b41b925a6e7f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f772a003566198519fe1849fd14f29d4346004b2e655810662507c95b4d5ffa1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:13Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:49Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:49 crc kubenswrapper[4754]: I0126 16:47:49.214024 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:49Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:49 crc kubenswrapper[4754]: I0126 16:47:49.228512 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c3718a4-f354-4284-92e0-fdfb45a692bd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b86f46f49163bca7eca9458279029af330dbd4890b03325e06fa3b1bdf030de6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxjcz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fffc03b536bfee3c7bc1bddd500e834e34b391896477a1501e3e06201a19374\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxjcz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-x65wv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:49Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:49 crc kubenswrapper[4754]: I0126 16:47:49.239444 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-t4ps6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d45179f0-9f81-4d77-8805-6b05bdb32f8f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vk8tg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vk8tg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:48Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-t4ps6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:49Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:49 crc kubenswrapper[4754]: I0126 16:47:49.239962 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:49 crc kubenswrapper[4754]: I0126 16:47:49.239998 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:49 crc kubenswrapper[4754]: I0126 16:47:49.240007 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:49 crc kubenswrapper[4754]: I0126 16:47:49.240020 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:49 crc kubenswrapper[4754]: I0126 16:47:49.240029 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:49Z","lastTransitionTime":"2026-01-26T16:47:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:49 crc kubenswrapper[4754]: I0126 16:47:49.253151 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b89aa081-0a2e-4757-9210-e2a6c228bee9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9c5f40188fda14b14128514fe84bddfe8dcede3d6c560283ccb2ac0c66895fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f45445cbf1d180d6d4dd9588e85bf678f043bfce8cff0812b78ca3510904f9a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://085fa7b8d4fef8b2722fac8c346107699b5a6bd33fbfc74ed35831cb8fdf9dad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c40e018a0979035827963efb9efc48a878a55b9df68ec1a9242d7cff64fe480\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8356f76fbc3a4e636c12d1da5455d3f6767d9b83ce2f2925a11435d7d7261e9f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0126 16:47:26.205208 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0126 16:47:26.206538 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3645934156/tls.crt::/tmp/serving-cert-3645934156/tls.key\\\\\\\"\\\\nI0126 16:47:31.504518 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0126 16:47:31.509366 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0126 16:47:31.509400 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0126 16:47:31.509428 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0126 16:47:31.509436 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0126 16:47:31.515912 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0126 16:47:31.516036 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 16:47:31.516068 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 16:47:31.516096 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0126 16:47:31.516121 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0126 16:47:31.516145 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0126 16:47:31.516170 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0126 16:47:31.516412 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0126 16:47:31.519294 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27b74e2e65a547e0e28fddd4f15feacb17f4d57596eeb8da82df9d39243c85e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22ba7649a1308815bfad52016000987a7a9a5bc214d7ff61e137ba099b20fc6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://22ba7649a1308815bfad52016000987a7a9a5bc214d7ff61e137ba099b20fc6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:13Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:49Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:49 crc kubenswrapper[4754]: I0126 16:47:49.267090 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://205ec256e46a8e08b4b0b0805fb6c5f4966fc3ccef0738df4b97cb404e74da17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:49Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:49 crc kubenswrapper[4754]: I0126 16:47:49.342844 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:49 crc kubenswrapper[4754]: I0126 16:47:49.342900 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:49 crc kubenswrapper[4754]: I0126 16:47:49.342913 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:49 crc kubenswrapper[4754]: I0126 16:47:49.342940 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:49 crc kubenswrapper[4754]: I0126 16:47:49.342955 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:49Z","lastTransitionTime":"2026-01-26T16:47:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:49 crc kubenswrapper[4754]: I0126 16:47:49.446827 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:49 crc kubenswrapper[4754]: I0126 16:47:49.446879 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:49 crc kubenswrapper[4754]: I0126 16:47:49.446898 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:49 crc kubenswrapper[4754]: I0126 16:47:49.446918 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:49 crc kubenswrapper[4754]: I0126 16:47:49.446931 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:49Z","lastTransitionTime":"2026-01-26T16:47:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:49 crc kubenswrapper[4754]: I0126 16:47:49.550196 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:49 crc kubenswrapper[4754]: I0126 16:47:49.550250 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:49 crc kubenswrapper[4754]: I0126 16:47:49.550261 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:49 crc kubenswrapper[4754]: I0126 16:47:49.550274 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:49 crc kubenswrapper[4754]: I0126 16:47:49.550285 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:49Z","lastTransitionTime":"2026-01-26T16:47:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:49 crc kubenswrapper[4754]: I0126 16:47:49.652564 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:49 crc kubenswrapper[4754]: I0126 16:47:49.652625 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:49 crc kubenswrapper[4754]: I0126 16:47:49.652635 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:49 crc kubenswrapper[4754]: I0126 16:47:49.652649 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:49 crc kubenswrapper[4754]: I0126 16:47:49.652658 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:49Z","lastTransitionTime":"2026-01-26T16:47:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:49 crc kubenswrapper[4754]: I0126 16:47:49.732984 4754 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-27 05:50:27.07216726 +0000 UTC Jan 26 16:47:49 crc kubenswrapper[4754]: I0126 16:47:49.754918 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:49 crc kubenswrapper[4754]: I0126 16:47:49.754958 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:49 crc kubenswrapper[4754]: I0126 16:47:49.754968 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:49 crc kubenswrapper[4754]: I0126 16:47:49.754985 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:49 crc kubenswrapper[4754]: I0126 16:47:49.754995 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:49Z","lastTransitionTime":"2026-01-26T16:47:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:49 crc kubenswrapper[4754]: I0126 16:47:49.766435 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-t4ps6" Jan 26 16:47:49 crc kubenswrapper[4754]: E0126 16:47:49.766557 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-t4ps6" podUID="d45179f0-9f81-4d77-8805-6b05bdb32f8f" Jan 26 16:47:49 crc kubenswrapper[4754]: I0126 16:47:49.766792 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 16:47:49 crc kubenswrapper[4754]: I0126 16:47:49.766873 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 16:47:49 crc kubenswrapper[4754]: E0126 16:47:49.767018 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 16:47:49 crc kubenswrapper[4754]: I0126 16:47:49.767030 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 16:47:49 crc kubenswrapper[4754]: E0126 16:47:49.767205 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 16:47:49 crc kubenswrapper[4754]: E0126 16:47:49.767271 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 16:47:49 crc kubenswrapper[4754]: I0126 16:47:49.857866 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:49 crc kubenswrapper[4754]: I0126 16:47:49.857904 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:49 crc kubenswrapper[4754]: I0126 16:47:49.857914 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:49 crc kubenswrapper[4754]: I0126 16:47:49.857930 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:49 crc kubenswrapper[4754]: I0126 16:47:49.857941 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:49Z","lastTransitionTime":"2026-01-26T16:47:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:49 crc kubenswrapper[4754]: I0126 16:47:49.959980 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:49 crc kubenswrapper[4754]: I0126 16:47:49.960024 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:49 crc kubenswrapper[4754]: I0126 16:47:49.960035 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:49 crc kubenswrapper[4754]: I0126 16:47:49.960049 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:49 crc kubenswrapper[4754]: I0126 16:47:49.960059 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:49Z","lastTransitionTime":"2026-01-26T16:47:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:50 crc kubenswrapper[4754]: I0126 16:47:50.062688 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:50 crc kubenswrapper[4754]: I0126 16:47:50.062724 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:50 crc kubenswrapper[4754]: I0126 16:47:50.062737 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:50 crc kubenswrapper[4754]: I0126 16:47:50.062753 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:50 crc kubenswrapper[4754]: I0126 16:47:50.062765 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:50Z","lastTransitionTime":"2026-01-26T16:47:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:50 crc kubenswrapper[4754]: I0126 16:47:50.104801 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d45179f0-9f81-4d77-8805-6b05bdb32f8f-metrics-certs\") pod \"network-metrics-daemon-t4ps6\" (UID: \"d45179f0-9f81-4d77-8805-6b05bdb32f8f\") " pod="openshift-multus/network-metrics-daemon-t4ps6" Jan 26 16:47:50 crc kubenswrapper[4754]: E0126 16:47:50.104957 4754 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Jan 26 16:47:50 crc kubenswrapper[4754]: E0126 16:47:50.105022 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d45179f0-9f81-4d77-8805-6b05bdb32f8f-metrics-certs podName:d45179f0-9f81-4d77-8805-6b05bdb32f8f nodeName:}" failed. No retries permitted until 2026-01-26 16:47:52.105004831 +0000 UTC m=+38.629185275 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/d45179f0-9f81-4d77-8805-6b05bdb32f8f-metrics-certs") pod "network-metrics-daemon-t4ps6" (UID: "d45179f0-9f81-4d77-8805-6b05bdb32f8f") : object "openshift-multus"/"metrics-daemon-secret" not registered Jan 26 16:47:50 crc kubenswrapper[4754]: I0126 16:47:50.166009 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:50 crc kubenswrapper[4754]: I0126 16:47:50.166054 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:50 crc kubenswrapper[4754]: I0126 16:47:50.166062 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:50 crc kubenswrapper[4754]: I0126 16:47:50.166081 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:50 crc kubenswrapper[4754]: I0126 16:47:50.166090 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:50Z","lastTransitionTime":"2026-01-26T16:47:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:50 crc kubenswrapper[4754]: I0126 16:47:50.268026 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:50 crc kubenswrapper[4754]: I0126 16:47:50.268066 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:50 crc kubenswrapper[4754]: I0126 16:47:50.268075 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:50 crc kubenswrapper[4754]: I0126 16:47:50.268090 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:50 crc kubenswrapper[4754]: I0126 16:47:50.268100 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:50Z","lastTransitionTime":"2026-01-26T16:47:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:50 crc kubenswrapper[4754]: I0126 16:47:50.370466 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:50 crc kubenswrapper[4754]: I0126 16:47:50.370520 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:50 crc kubenswrapper[4754]: I0126 16:47:50.370532 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:50 crc kubenswrapper[4754]: I0126 16:47:50.370550 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:50 crc kubenswrapper[4754]: I0126 16:47:50.370561 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:50Z","lastTransitionTime":"2026-01-26T16:47:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:50 crc kubenswrapper[4754]: I0126 16:47:50.473201 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:50 crc kubenswrapper[4754]: I0126 16:47:50.473346 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:50 crc kubenswrapper[4754]: I0126 16:47:50.473361 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:50 crc kubenswrapper[4754]: I0126 16:47:50.473376 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:50 crc kubenswrapper[4754]: I0126 16:47:50.473386 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:50Z","lastTransitionTime":"2026-01-26T16:47:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:50 crc kubenswrapper[4754]: I0126 16:47:50.575764 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:50 crc kubenswrapper[4754]: I0126 16:47:50.575806 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:50 crc kubenswrapper[4754]: I0126 16:47:50.575816 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:50 crc kubenswrapper[4754]: I0126 16:47:50.575835 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:50 crc kubenswrapper[4754]: I0126 16:47:50.575846 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:50Z","lastTransitionTime":"2026-01-26T16:47:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:50 crc kubenswrapper[4754]: I0126 16:47:50.678804 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:50 crc kubenswrapper[4754]: I0126 16:47:50.678853 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:50 crc kubenswrapper[4754]: I0126 16:47:50.678864 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:50 crc kubenswrapper[4754]: I0126 16:47:50.678880 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:50 crc kubenswrapper[4754]: I0126 16:47:50.678896 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:50Z","lastTransitionTime":"2026-01-26T16:47:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:50 crc kubenswrapper[4754]: I0126 16:47:50.733901 4754 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-28 22:35:27.682464076 +0000 UTC Jan 26 16:47:50 crc kubenswrapper[4754]: I0126 16:47:50.781303 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:50 crc kubenswrapper[4754]: I0126 16:47:50.781353 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:50 crc kubenswrapper[4754]: I0126 16:47:50.781366 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:50 crc kubenswrapper[4754]: I0126 16:47:50.781381 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:50 crc kubenswrapper[4754]: I0126 16:47:50.781390 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:50Z","lastTransitionTime":"2026-01-26T16:47:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:50 crc kubenswrapper[4754]: I0126 16:47:50.883713 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:50 crc kubenswrapper[4754]: I0126 16:47:50.883793 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:50 crc kubenswrapper[4754]: I0126 16:47:50.883808 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:50 crc kubenswrapper[4754]: I0126 16:47:50.883829 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:50 crc kubenswrapper[4754]: I0126 16:47:50.883843 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:50Z","lastTransitionTime":"2026-01-26T16:47:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:50 crc kubenswrapper[4754]: I0126 16:47:50.985726 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:50 crc kubenswrapper[4754]: I0126 16:47:50.985763 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:50 crc kubenswrapper[4754]: I0126 16:47:50.985774 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:50 crc kubenswrapper[4754]: I0126 16:47:50.985790 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:50 crc kubenswrapper[4754]: I0126 16:47:50.985803 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:50Z","lastTransitionTime":"2026-01-26T16:47:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:51 crc kubenswrapper[4754]: I0126 16:47:51.088639 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:51 crc kubenswrapper[4754]: I0126 16:47:51.088739 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:51 crc kubenswrapper[4754]: I0126 16:47:51.088754 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:51 crc kubenswrapper[4754]: I0126 16:47:51.088770 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:51 crc kubenswrapper[4754]: I0126 16:47:51.088781 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:51Z","lastTransitionTime":"2026-01-26T16:47:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:51 crc kubenswrapper[4754]: I0126 16:47:51.190729 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:51 crc kubenswrapper[4754]: I0126 16:47:51.190773 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:51 crc kubenswrapper[4754]: I0126 16:47:51.190784 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:51 crc kubenswrapper[4754]: I0126 16:47:51.190802 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:51 crc kubenswrapper[4754]: I0126 16:47:51.190816 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:51Z","lastTransitionTime":"2026-01-26T16:47:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:51 crc kubenswrapper[4754]: I0126 16:47:51.293893 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:51 crc kubenswrapper[4754]: I0126 16:47:51.293936 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:51 crc kubenswrapper[4754]: I0126 16:47:51.293947 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:51 crc kubenswrapper[4754]: I0126 16:47:51.293961 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:51 crc kubenswrapper[4754]: I0126 16:47:51.293970 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:51Z","lastTransitionTime":"2026-01-26T16:47:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:51 crc kubenswrapper[4754]: I0126 16:47:51.396527 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:51 crc kubenswrapper[4754]: I0126 16:47:51.396616 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:51 crc kubenswrapper[4754]: I0126 16:47:51.396625 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:51 crc kubenswrapper[4754]: I0126 16:47:51.396641 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:51 crc kubenswrapper[4754]: I0126 16:47:51.396652 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:51Z","lastTransitionTime":"2026-01-26T16:47:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:51 crc kubenswrapper[4754]: I0126 16:47:51.499195 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:51 crc kubenswrapper[4754]: I0126 16:47:51.499251 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:51 crc kubenswrapper[4754]: I0126 16:47:51.499262 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:51 crc kubenswrapper[4754]: I0126 16:47:51.499277 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:51 crc kubenswrapper[4754]: I0126 16:47:51.499288 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:51Z","lastTransitionTime":"2026-01-26T16:47:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:51 crc kubenswrapper[4754]: I0126 16:47:51.602124 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:51 crc kubenswrapper[4754]: I0126 16:47:51.602174 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:51 crc kubenswrapper[4754]: I0126 16:47:51.602186 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:51 crc kubenswrapper[4754]: I0126 16:47:51.602209 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:51 crc kubenswrapper[4754]: I0126 16:47:51.602223 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:51Z","lastTransitionTime":"2026-01-26T16:47:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:51 crc kubenswrapper[4754]: I0126 16:47:51.704164 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:51 crc kubenswrapper[4754]: I0126 16:47:51.704201 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:51 crc kubenswrapper[4754]: I0126 16:47:51.704211 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:51 crc kubenswrapper[4754]: I0126 16:47:51.704227 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:51 crc kubenswrapper[4754]: I0126 16:47:51.704238 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:51Z","lastTransitionTime":"2026-01-26T16:47:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:51 crc kubenswrapper[4754]: I0126 16:47:51.735004 4754 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-11 11:04:25.039797243 +0000 UTC Jan 26 16:47:51 crc kubenswrapper[4754]: I0126 16:47:51.766340 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 16:47:51 crc kubenswrapper[4754]: I0126 16:47:51.766364 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 16:47:51 crc kubenswrapper[4754]: I0126 16:47:51.766407 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-t4ps6" Jan 26 16:47:51 crc kubenswrapper[4754]: I0126 16:47:51.766466 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 16:47:51 crc kubenswrapper[4754]: E0126 16:47:51.766560 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 16:47:51 crc kubenswrapper[4754]: E0126 16:47:51.766701 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 16:47:51 crc kubenswrapper[4754]: E0126 16:47:51.766738 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-t4ps6" podUID="d45179f0-9f81-4d77-8805-6b05bdb32f8f" Jan 26 16:47:51 crc kubenswrapper[4754]: E0126 16:47:51.766795 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 16:47:51 crc kubenswrapper[4754]: I0126 16:47:51.806285 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:51 crc kubenswrapper[4754]: I0126 16:47:51.806568 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:51 crc kubenswrapper[4754]: I0126 16:47:51.806697 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:51 crc kubenswrapper[4754]: I0126 16:47:51.806799 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:51 crc kubenswrapper[4754]: I0126 16:47:51.806893 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:51Z","lastTransitionTime":"2026-01-26T16:47:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:51 crc kubenswrapper[4754]: I0126 16:47:51.909657 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:51 crc kubenswrapper[4754]: I0126 16:47:51.909741 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:51 crc kubenswrapper[4754]: I0126 16:47:51.909754 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:51 crc kubenswrapper[4754]: I0126 16:47:51.909778 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:51 crc kubenswrapper[4754]: I0126 16:47:51.909792 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:51Z","lastTransitionTime":"2026-01-26T16:47:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:52 crc kubenswrapper[4754]: I0126 16:47:52.011408 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:52 crc kubenswrapper[4754]: I0126 16:47:52.011447 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:52 crc kubenswrapper[4754]: I0126 16:47:52.011458 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:52 crc kubenswrapper[4754]: I0126 16:47:52.011475 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:52 crc kubenswrapper[4754]: I0126 16:47:52.011487 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:52Z","lastTransitionTime":"2026-01-26T16:47:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:52 crc kubenswrapper[4754]: I0126 16:47:52.113868 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:52 crc kubenswrapper[4754]: I0126 16:47:52.114747 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:52 crc kubenswrapper[4754]: I0126 16:47:52.114852 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:52 crc kubenswrapper[4754]: I0126 16:47:52.114947 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:52 crc kubenswrapper[4754]: I0126 16:47:52.115024 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:52Z","lastTransitionTime":"2026-01-26T16:47:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:52 crc kubenswrapper[4754]: I0126 16:47:52.124452 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d45179f0-9f81-4d77-8805-6b05bdb32f8f-metrics-certs\") pod \"network-metrics-daemon-t4ps6\" (UID: \"d45179f0-9f81-4d77-8805-6b05bdb32f8f\") " pod="openshift-multus/network-metrics-daemon-t4ps6" Jan 26 16:47:52 crc kubenswrapper[4754]: E0126 16:47:52.124692 4754 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Jan 26 16:47:52 crc kubenswrapper[4754]: E0126 16:47:52.124791 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d45179f0-9f81-4d77-8805-6b05bdb32f8f-metrics-certs podName:d45179f0-9f81-4d77-8805-6b05bdb32f8f nodeName:}" failed. No retries permitted until 2026-01-26 16:47:56.124767746 +0000 UTC m=+42.648948250 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/d45179f0-9f81-4d77-8805-6b05bdb32f8f-metrics-certs") pod "network-metrics-daemon-t4ps6" (UID: "d45179f0-9f81-4d77-8805-6b05bdb32f8f") : object "openshift-multus"/"metrics-daemon-secret" not registered Jan 26 16:47:52 crc kubenswrapper[4754]: I0126 16:47:52.217162 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:52 crc kubenswrapper[4754]: I0126 16:47:52.217418 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:52 crc kubenswrapper[4754]: I0126 16:47:52.217494 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:52 crc kubenswrapper[4754]: I0126 16:47:52.217571 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:52 crc kubenswrapper[4754]: I0126 16:47:52.217647 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:52Z","lastTransitionTime":"2026-01-26T16:47:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:52 crc kubenswrapper[4754]: I0126 16:47:52.320270 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:52 crc kubenswrapper[4754]: I0126 16:47:52.320525 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:52 crc kubenswrapper[4754]: I0126 16:47:52.320617 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:52 crc kubenswrapper[4754]: I0126 16:47:52.320760 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:52 crc kubenswrapper[4754]: I0126 16:47:52.320836 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:52Z","lastTransitionTime":"2026-01-26T16:47:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:52 crc kubenswrapper[4754]: I0126 16:47:52.423194 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:52 crc kubenswrapper[4754]: I0126 16:47:52.423228 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:52 crc kubenswrapper[4754]: I0126 16:47:52.423238 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:52 crc kubenswrapper[4754]: I0126 16:47:52.423251 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:52 crc kubenswrapper[4754]: I0126 16:47:52.423262 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:52Z","lastTransitionTime":"2026-01-26T16:47:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:52 crc kubenswrapper[4754]: I0126 16:47:52.525816 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:52 crc kubenswrapper[4754]: I0126 16:47:52.525859 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:52 crc kubenswrapper[4754]: I0126 16:47:52.525871 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:52 crc kubenswrapper[4754]: I0126 16:47:52.525888 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:52 crc kubenswrapper[4754]: I0126 16:47:52.525900 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:52Z","lastTransitionTime":"2026-01-26T16:47:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:52 crc kubenswrapper[4754]: I0126 16:47:52.629192 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:52 crc kubenswrapper[4754]: I0126 16:47:52.629240 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:52 crc kubenswrapper[4754]: I0126 16:47:52.629252 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:52 crc kubenswrapper[4754]: I0126 16:47:52.629270 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:52 crc kubenswrapper[4754]: I0126 16:47:52.629283 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:52Z","lastTransitionTime":"2026-01-26T16:47:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:52 crc kubenswrapper[4754]: I0126 16:47:52.731233 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:52 crc kubenswrapper[4754]: I0126 16:47:52.731265 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:52 crc kubenswrapper[4754]: I0126 16:47:52.731275 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:52 crc kubenswrapper[4754]: I0126 16:47:52.731288 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:52 crc kubenswrapper[4754]: I0126 16:47:52.731299 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:52Z","lastTransitionTime":"2026-01-26T16:47:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:52 crc kubenswrapper[4754]: I0126 16:47:52.735574 4754 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-02 14:48:08.714746939 +0000 UTC Jan 26 16:47:52 crc kubenswrapper[4754]: I0126 16:47:52.832729 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:52 crc kubenswrapper[4754]: I0126 16:47:52.832833 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:52 crc kubenswrapper[4754]: I0126 16:47:52.832842 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:52 crc kubenswrapper[4754]: I0126 16:47:52.832855 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:52 crc kubenswrapper[4754]: I0126 16:47:52.832865 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:52Z","lastTransitionTime":"2026-01-26T16:47:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:52 crc kubenswrapper[4754]: I0126 16:47:52.935871 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:52 crc kubenswrapper[4754]: I0126 16:47:52.935908 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:52 crc kubenswrapper[4754]: I0126 16:47:52.935964 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:52 crc kubenswrapper[4754]: I0126 16:47:52.935984 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:52 crc kubenswrapper[4754]: I0126 16:47:52.935998 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:52Z","lastTransitionTime":"2026-01-26T16:47:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:53 crc kubenswrapper[4754]: I0126 16:47:53.038657 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:53 crc kubenswrapper[4754]: I0126 16:47:53.038726 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:53 crc kubenswrapper[4754]: I0126 16:47:53.038736 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:53 crc kubenswrapper[4754]: I0126 16:47:53.038751 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:53 crc kubenswrapper[4754]: I0126 16:47:53.038761 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:53Z","lastTransitionTime":"2026-01-26T16:47:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:53 crc kubenswrapper[4754]: I0126 16:47:53.142165 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:53 crc kubenswrapper[4754]: I0126 16:47:53.142223 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:53 crc kubenswrapper[4754]: I0126 16:47:53.142235 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:53 crc kubenswrapper[4754]: I0126 16:47:53.142257 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:53 crc kubenswrapper[4754]: I0126 16:47:53.142272 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:53Z","lastTransitionTime":"2026-01-26T16:47:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:53 crc kubenswrapper[4754]: I0126 16:47:53.244610 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:53 crc kubenswrapper[4754]: I0126 16:47:53.244655 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:53 crc kubenswrapper[4754]: I0126 16:47:53.244687 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:53 crc kubenswrapper[4754]: I0126 16:47:53.244702 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:53 crc kubenswrapper[4754]: I0126 16:47:53.244712 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:53Z","lastTransitionTime":"2026-01-26T16:47:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:53 crc kubenswrapper[4754]: I0126 16:47:53.348734 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:53 crc kubenswrapper[4754]: I0126 16:47:53.349030 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:53 crc kubenswrapper[4754]: I0126 16:47:53.349092 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:53 crc kubenswrapper[4754]: I0126 16:47:53.349185 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:53 crc kubenswrapper[4754]: I0126 16:47:53.349287 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:53Z","lastTransitionTime":"2026-01-26T16:47:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:53 crc kubenswrapper[4754]: I0126 16:47:53.454221 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:53 crc kubenswrapper[4754]: I0126 16:47:53.454293 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:53 crc kubenswrapper[4754]: I0126 16:47:53.454307 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:53 crc kubenswrapper[4754]: I0126 16:47:53.454325 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:53 crc kubenswrapper[4754]: I0126 16:47:53.454337 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:53Z","lastTransitionTime":"2026-01-26T16:47:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:53 crc kubenswrapper[4754]: I0126 16:47:53.556209 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:53 crc kubenswrapper[4754]: I0126 16:47:53.556243 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:53 crc kubenswrapper[4754]: I0126 16:47:53.556252 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:53 crc kubenswrapper[4754]: I0126 16:47:53.556265 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:53 crc kubenswrapper[4754]: I0126 16:47:53.556275 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:53Z","lastTransitionTime":"2026-01-26T16:47:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:53 crc kubenswrapper[4754]: I0126 16:47:53.659049 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:53 crc kubenswrapper[4754]: I0126 16:47:53.659114 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:53 crc kubenswrapper[4754]: I0126 16:47:53.659131 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:53 crc kubenswrapper[4754]: I0126 16:47:53.659152 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:53 crc kubenswrapper[4754]: I0126 16:47:53.659168 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:53Z","lastTransitionTime":"2026-01-26T16:47:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:53 crc kubenswrapper[4754]: I0126 16:47:53.735842 4754 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-03 03:10:32.010486149 +0000 UTC Jan 26 16:47:53 crc kubenswrapper[4754]: I0126 16:47:53.762255 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:53 crc kubenswrapper[4754]: I0126 16:47:53.762315 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:53 crc kubenswrapper[4754]: I0126 16:47:53.762326 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:53 crc kubenswrapper[4754]: I0126 16:47:53.762340 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:53 crc kubenswrapper[4754]: I0126 16:47:53.762351 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:53Z","lastTransitionTime":"2026-01-26T16:47:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:53 crc kubenswrapper[4754]: I0126 16:47:53.766547 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 16:47:53 crc kubenswrapper[4754]: E0126 16:47:53.766655 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 16:47:53 crc kubenswrapper[4754]: I0126 16:47:53.766690 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-t4ps6" Jan 26 16:47:53 crc kubenswrapper[4754]: I0126 16:47:53.766723 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 16:47:53 crc kubenswrapper[4754]: E0126 16:47:53.766788 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-t4ps6" podUID="d45179f0-9f81-4d77-8805-6b05bdb32f8f" Jan 26 16:47:53 crc kubenswrapper[4754]: I0126 16:47:53.767021 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 16:47:53 crc kubenswrapper[4754]: E0126 16:47:53.767011 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 16:47:53 crc kubenswrapper[4754]: E0126 16:47:53.767079 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 16:47:53 crc kubenswrapper[4754]: I0126 16:47:53.783334 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bwpd6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"351cc6c0-b8e5-4589-ae63-f98e8ab1eca3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1022f101d36873b41bbc1ed8a4a2590347b49f7cce6a172aa1d5bc9f529aed65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3ce0fe080a094316b9456057e347b90a5ff432daa12b9c56d7aea9bee1f72b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3ce0fe080a094316b9456057e347b90a5ff432daa12b9c56d7aea9bee1f72b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1772bb41eb8a36a863d3c691610bcbee5d16b260480b3e9f9d6b083e09e42799\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1772bb41eb8a36a863d3c691610bcbee5d16b260480b3e9f9d6b083e09e42799\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24848d69756024cc664affb9492764ace0638ef611f910f4f2bf86b638cc443d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://24848d69756024cc664affb9492764ace0638ef611f910f4f2bf86b638cc443d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c264fc7c3300a725566acce4464f01d9314779737624401c164d867831fd93bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c264fc7c3300a725566acce4464f01d9314779737624401c164d867831fd93bf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5acfe327eee4aad927ac3a32d24549eff788b14a1c4bdac359b6b56301d884ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5acfe327eee4aad927ac3a32d24549eff788b14a1c4bdac359b6b56301d884ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://63a9b9317494d0e0b33970ef3d461729b10d2a1ad7d3d2bd33f389b184ca56ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://63a9b9317494d0e0b33970ef3d461729b10d2a1ad7d3d2bd33f389b184ca56ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bwpd6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:53Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:53 crc kubenswrapper[4754]: I0126 16:47:53.803276 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jsbxt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"689569c4-93ee-4b82-93fc-9af0f97e6dc3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://033f8b104d4e9ff6d5f87ca998665f12e0c56696b3d30e01ede05f4da540365d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://919572e731813a98c7d8a94879bba931f75c472c7cd681e86d5197ebab43815d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1ac1f49bd564750fcff530a0f76657b0ed41573539acd020a922d19861ef834\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a300a80ad87f46ba1b4fd7ad3cf6c25767fcfe966b93bdd86d729843f53a5cda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31dcab2f9a2d218b0e49dad1df46377ff4521592824a0fbac254d5b8e6d0632e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fe4d9395447ac570eca1338e749dbefea76295d427a12487b63a142a7d6a2e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eaed406e767c0eedc024810541a22ed6ac3b1dcc2b5d41aa007b9ad282eb57fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eaed406e767c0eedc024810541a22ed6ac3b1dcc2b5d41aa007b9ad282eb57fd\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-26T16:47:45Z\\\",\\\"message\\\":\\\".EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0126 16:47:45.693395 6233 handler.go:208] Removed *v1.Node event handler 7\\\\nI0126 16:47:45.694648 6233 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI0126 16:47:45.694956 6233 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI0126 16:47:45.694961 6233 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0126 16:47:45.695763 6233 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0126 16:47:45.695841 6233 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0126 16:47:45.695852 6233 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0126 16:47:45.695893 6233 factory.go:656] Stopping watch factory\\\\nI0126 16:47:45.695913 6233 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0126 16:47:45.695920 6233 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0126 16:47:45.695935 6233 handler.go:208] Removed *v1.Namespace ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:44Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-jsbxt_openshift-ovn-kubernetes(689569c4-93ee-4b82-93fc-9af0f97e6dc3)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://07a43ff7860ff65faf51f4a5e8aeb326f06d6d7222677a92560ae2e3a0c81484\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d9a7bafabd44f8b88a6f62cbe45599af8563bc45d9d7ad3a7669f3d8d11eb9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d9a7bafabd44f8b88a6f62cbe45599af8563bc45d9d7ad3a7669f3d8d11eb9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jsbxt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:53Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:53 crc kubenswrapper[4754]: I0126 16:47:53.816448 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:53Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:53 crc kubenswrapper[4754]: I0126 16:47:53.830713 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e459fd8c1093d6f3d0a671751b792b26b7df94c21ced1be1a2250821bcc82d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:53Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:53 crc kubenswrapper[4754]: I0126 16:47:53.842992 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2blzw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca65468e-30c6-4666-962e-cc0de05e67fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4603235bb2af495901926b805fe1c0ffc48824a4394ccf928e7fcd53a673ac3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c25z4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:33Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2blzw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:53Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:53 crc kubenswrapper[4754]: I0126 16:47:53.852268 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-84lgq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6763d962-0892-4acc-bab9-7a4e93e39a85\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://953c1917fca896c3def3fc2c69f9e3dbdc9ca26fc2b3d40282be71f71759fd68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8zlw6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:33Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-84lgq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:53Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:53 crc kubenswrapper[4754]: I0126 16:47:53.864232 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-skh5n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b619d233-b592-4b05-a0b7-dc094c88471e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9231a4191565e2e0ef53d8a394bb34615c900b601a0dfec24c2d8e5fa7b0b659\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-skh5n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:53Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:53 crc kubenswrapper[4754]: I0126 16:47:53.864571 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:53 crc kubenswrapper[4754]: I0126 16:47:53.864604 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:53 crc kubenswrapper[4754]: I0126 16:47:53.864614 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:53 crc kubenswrapper[4754]: I0126 16:47:53.864630 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:53 crc kubenswrapper[4754]: I0126 16:47:53.864640 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:53Z","lastTransitionTime":"2026-01-26T16:47:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:53 crc kubenswrapper[4754]: I0126 16:47:53.874242 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jspvh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5c81aab0-939e-45ce-b19a-89d4818980c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://356dc54150a495dd6f35e0667c82b807b712b07200d8a0adbb1c92fe3f7af8f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8bps\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://05fd57780c44cd127859dc8b163d20eafbd8b06f3348cf545ef41021ce6bc526\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8bps\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-jspvh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:53Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:53 crc kubenswrapper[4754]: I0126 16:47:53.885749 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"099af702-abf2-4933-b4c9-63ef2326e468\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23017e73f8450839b4c2c02fa61235ffe4604bdce11556def27f25bcadcf89dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a74507c941b9de4d0401c8f1d53bb79ed11359628d3405cdab7f5474a75c862e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4402182b3e2e35a7d5dd4c44ac53e152915b312b7ccb076c4f2b41b925a6e7f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f772a003566198519fe1849fd14f29d4346004b2e655810662507c95b4d5ffa1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:13Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:53Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:53 crc kubenswrapper[4754]: I0126 16:47:53.900131 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:53Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:53 crc kubenswrapper[4754]: I0126 16:47:53.911617 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89251fa9c64bc25d6d6f355400d7644c099f22aed1cae812fc991a03595f1e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e90508d8a53be9146f9b388750730fdae9ee200306caf5ad393a356390a69ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:53Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:53 crc kubenswrapper[4754]: I0126 16:47:53.922941 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-t4ps6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d45179f0-9f81-4d77-8805-6b05bdb32f8f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vk8tg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vk8tg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:48Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-t4ps6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:53Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:53 crc kubenswrapper[4754]: I0126 16:47:53.936519 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b89aa081-0a2e-4757-9210-e2a6c228bee9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9c5f40188fda14b14128514fe84bddfe8dcede3d6c560283ccb2ac0c66895fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f45445cbf1d180d6d4dd9588e85bf678f043bfce8cff0812b78ca3510904f9a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://085fa7b8d4fef8b2722fac8c346107699b5a6bd33fbfc74ed35831cb8fdf9dad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c40e018a0979035827963efb9efc48a878a55b9df68ec1a9242d7cff64fe480\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8356f76fbc3a4e636c12d1da5455d3f6767d9b83ce2f2925a11435d7d7261e9f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0126 16:47:26.205208 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0126 16:47:26.206538 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3645934156/tls.crt::/tmp/serving-cert-3645934156/tls.key\\\\\\\"\\\\nI0126 16:47:31.504518 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0126 16:47:31.509366 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0126 16:47:31.509400 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0126 16:47:31.509428 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0126 16:47:31.509436 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0126 16:47:31.515912 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0126 16:47:31.516036 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 16:47:31.516068 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 16:47:31.516096 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0126 16:47:31.516121 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0126 16:47:31.516145 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0126 16:47:31.516170 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0126 16:47:31.516412 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0126 16:47:31.519294 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27b74e2e65a547e0e28fddd4f15feacb17f4d57596eeb8da82df9d39243c85e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22ba7649a1308815bfad52016000987a7a9a5bc214d7ff61e137ba099b20fc6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://22ba7649a1308815bfad52016000987a7a9a5bc214d7ff61e137ba099b20fc6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:13Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:53Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:53 crc kubenswrapper[4754]: I0126 16:47:53.948927 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://205ec256e46a8e08b4b0b0805fb6c5f4966fc3ccef0738df4b97cb404e74da17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:53Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:53 crc kubenswrapper[4754]: I0126 16:47:53.960904 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:53Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:53 crc kubenswrapper[4754]: I0126 16:47:53.967387 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:53 crc kubenswrapper[4754]: I0126 16:47:53.967432 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:53 crc kubenswrapper[4754]: I0126 16:47:53.967443 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:53 crc kubenswrapper[4754]: I0126 16:47:53.967459 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:53 crc kubenswrapper[4754]: I0126 16:47:53.967498 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:53Z","lastTransitionTime":"2026-01-26T16:47:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:53 crc kubenswrapper[4754]: I0126 16:47:53.972646 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c3718a4-f354-4284-92e0-fdfb45a692bd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b86f46f49163bca7eca9458279029af330dbd4890b03325e06fa3b1bdf030de6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxjcz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fffc03b536bfee3c7bc1bddd500e834e34b391896477a1501e3e06201a19374\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxjcz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-x65wv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:53Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:54 crc kubenswrapper[4754]: I0126 16:47:54.069724 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:54 crc kubenswrapper[4754]: I0126 16:47:54.069769 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:54 crc kubenswrapper[4754]: I0126 16:47:54.069780 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:54 crc kubenswrapper[4754]: I0126 16:47:54.069795 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:54 crc kubenswrapper[4754]: I0126 16:47:54.069807 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:54Z","lastTransitionTime":"2026-01-26T16:47:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:54 crc kubenswrapper[4754]: I0126 16:47:54.172363 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:54 crc kubenswrapper[4754]: I0126 16:47:54.172427 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:54 crc kubenswrapper[4754]: I0126 16:47:54.172440 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:54 crc kubenswrapper[4754]: I0126 16:47:54.172460 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:54 crc kubenswrapper[4754]: I0126 16:47:54.172479 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:54Z","lastTransitionTime":"2026-01-26T16:47:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:54 crc kubenswrapper[4754]: I0126 16:47:54.275479 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:54 crc kubenswrapper[4754]: I0126 16:47:54.275522 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:54 crc kubenswrapper[4754]: I0126 16:47:54.275536 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:54 crc kubenswrapper[4754]: I0126 16:47:54.275552 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:54 crc kubenswrapper[4754]: I0126 16:47:54.275564 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:54Z","lastTransitionTime":"2026-01-26T16:47:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:54 crc kubenswrapper[4754]: I0126 16:47:54.379507 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:54 crc kubenswrapper[4754]: I0126 16:47:54.379563 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:54 crc kubenswrapper[4754]: I0126 16:47:54.379578 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:54 crc kubenswrapper[4754]: I0126 16:47:54.379599 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:54 crc kubenswrapper[4754]: I0126 16:47:54.379613 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:54Z","lastTransitionTime":"2026-01-26T16:47:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:54 crc kubenswrapper[4754]: I0126 16:47:54.482362 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:54 crc kubenswrapper[4754]: I0126 16:47:54.482414 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:54 crc kubenswrapper[4754]: I0126 16:47:54.482425 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:54 crc kubenswrapper[4754]: I0126 16:47:54.482443 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:54 crc kubenswrapper[4754]: I0126 16:47:54.482457 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:54Z","lastTransitionTime":"2026-01-26T16:47:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:54 crc kubenswrapper[4754]: I0126 16:47:54.585848 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:54 crc kubenswrapper[4754]: I0126 16:47:54.585929 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:54 crc kubenswrapper[4754]: I0126 16:47:54.585945 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:54 crc kubenswrapper[4754]: I0126 16:47:54.585972 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:54 crc kubenswrapper[4754]: I0126 16:47:54.586052 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:54Z","lastTransitionTime":"2026-01-26T16:47:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:54 crc kubenswrapper[4754]: I0126 16:47:54.688802 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:54 crc kubenswrapper[4754]: I0126 16:47:54.688838 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:54 crc kubenswrapper[4754]: I0126 16:47:54.688851 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:54 crc kubenswrapper[4754]: I0126 16:47:54.688868 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:54 crc kubenswrapper[4754]: I0126 16:47:54.688879 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:54Z","lastTransitionTime":"2026-01-26T16:47:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:54 crc kubenswrapper[4754]: I0126 16:47:54.736875 4754 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-07 13:49:09.166790117 +0000 UTC Jan 26 16:47:54 crc kubenswrapper[4754]: I0126 16:47:54.790869 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:54 crc kubenswrapper[4754]: I0126 16:47:54.790911 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:54 crc kubenswrapper[4754]: I0126 16:47:54.790921 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:54 crc kubenswrapper[4754]: I0126 16:47:54.790935 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:54 crc kubenswrapper[4754]: I0126 16:47:54.790945 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:54Z","lastTransitionTime":"2026-01-26T16:47:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:54 crc kubenswrapper[4754]: I0126 16:47:54.894128 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:54 crc kubenswrapper[4754]: I0126 16:47:54.894170 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:54 crc kubenswrapper[4754]: I0126 16:47:54.894180 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:54 crc kubenswrapper[4754]: I0126 16:47:54.894196 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:54 crc kubenswrapper[4754]: I0126 16:47:54.894207 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:54Z","lastTransitionTime":"2026-01-26T16:47:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:54 crc kubenswrapper[4754]: I0126 16:47:54.964505 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:54 crc kubenswrapper[4754]: I0126 16:47:54.964559 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:54 crc kubenswrapper[4754]: I0126 16:47:54.964568 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:54 crc kubenswrapper[4754]: I0126 16:47:54.964588 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:54 crc kubenswrapper[4754]: I0126 16:47:54.964601 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:54Z","lastTransitionTime":"2026-01-26T16:47:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:54 crc kubenswrapper[4754]: E0126 16:47:54.977809 4754 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:47:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:47:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:54Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:47:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:47:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:54Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d8506764-e7fa-45cb-a13d-6f527164f548\\\",\\\"systemUUID\\\":\\\"bfbb6c33-ce46-4e81-a8a8-b44409b03821\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:54Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:54 crc kubenswrapper[4754]: I0126 16:47:54.981880 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:54 crc kubenswrapper[4754]: I0126 16:47:54.981941 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:54 crc kubenswrapper[4754]: I0126 16:47:54.981957 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:54 crc kubenswrapper[4754]: I0126 16:47:54.981976 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:54 crc kubenswrapper[4754]: I0126 16:47:54.981987 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:54Z","lastTransitionTime":"2026-01-26T16:47:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:54 crc kubenswrapper[4754]: E0126 16:47:54.997444 4754 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:47:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:47:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:54Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:47:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:47:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:54Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d8506764-e7fa-45cb-a13d-6f527164f548\\\",\\\"systemUUID\\\":\\\"bfbb6c33-ce46-4e81-a8a8-b44409b03821\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:54Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:55 crc kubenswrapper[4754]: I0126 16:47:55.003220 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:55 crc kubenswrapper[4754]: I0126 16:47:55.003347 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:55 crc kubenswrapper[4754]: I0126 16:47:55.003370 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:55 crc kubenswrapper[4754]: I0126 16:47:55.003395 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:55 crc kubenswrapper[4754]: I0126 16:47:55.003415 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:55Z","lastTransitionTime":"2026-01-26T16:47:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:55 crc kubenswrapper[4754]: E0126 16:47:55.020468 4754 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:47:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:47:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:55Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:47:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:47:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:55Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d8506764-e7fa-45cb-a13d-6f527164f548\\\",\\\"systemUUID\\\":\\\"bfbb6c33-ce46-4e81-a8a8-b44409b03821\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:55Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:55 crc kubenswrapper[4754]: I0126 16:47:55.027782 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:55 crc kubenswrapper[4754]: I0126 16:47:55.028022 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:55 crc kubenswrapper[4754]: I0126 16:47:55.028139 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:55 crc kubenswrapper[4754]: I0126 16:47:55.028237 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:55 crc kubenswrapper[4754]: I0126 16:47:55.028319 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:55Z","lastTransitionTime":"2026-01-26T16:47:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:55 crc kubenswrapper[4754]: E0126 16:47:55.042107 4754 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:47:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:47:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:55Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:47:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:47:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:55Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d8506764-e7fa-45cb-a13d-6f527164f548\\\",\\\"systemUUID\\\":\\\"bfbb6c33-ce46-4e81-a8a8-b44409b03821\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:55Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:55 crc kubenswrapper[4754]: I0126 16:47:55.046464 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:55 crc kubenswrapper[4754]: I0126 16:47:55.046591 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:55 crc kubenswrapper[4754]: I0126 16:47:55.046736 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:55 crc kubenswrapper[4754]: I0126 16:47:55.046872 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:55 crc kubenswrapper[4754]: I0126 16:47:55.046957 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:55Z","lastTransitionTime":"2026-01-26T16:47:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:55 crc kubenswrapper[4754]: E0126 16:47:55.059971 4754 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:47:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:47:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:55Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:47:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:47:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:55Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d8506764-e7fa-45cb-a13d-6f527164f548\\\",\\\"systemUUID\\\":\\\"bfbb6c33-ce46-4e81-a8a8-b44409b03821\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:55Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:55 crc kubenswrapper[4754]: E0126 16:47:55.060204 4754 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Jan 26 16:47:55 crc kubenswrapper[4754]: I0126 16:47:55.061615 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:55 crc kubenswrapper[4754]: I0126 16:47:55.061658 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:55 crc kubenswrapper[4754]: I0126 16:47:55.061721 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:55 crc kubenswrapper[4754]: I0126 16:47:55.061753 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:55 crc kubenswrapper[4754]: I0126 16:47:55.061764 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:55Z","lastTransitionTime":"2026-01-26T16:47:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:55 crc kubenswrapper[4754]: I0126 16:47:55.164035 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:55 crc kubenswrapper[4754]: I0126 16:47:55.164099 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:55 crc kubenswrapper[4754]: I0126 16:47:55.164111 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:55 crc kubenswrapper[4754]: I0126 16:47:55.164157 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:55 crc kubenswrapper[4754]: I0126 16:47:55.164167 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:55Z","lastTransitionTime":"2026-01-26T16:47:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:55 crc kubenswrapper[4754]: I0126 16:47:55.266969 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:55 crc kubenswrapper[4754]: I0126 16:47:55.267057 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:55 crc kubenswrapper[4754]: I0126 16:47:55.267074 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:55 crc kubenswrapper[4754]: I0126 16:47:55.267100 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:55 crc kubenswrapper[4754]: I0126 16:47:55.267114 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:55Z","lastTransitionTime":"2026-01-26T16:47:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:55 crc kubenswrapper[4754]: I0126 16:47:55.369184 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:55 crc kubenswrapper[4754]: I0126 16:47:55.369219 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:55 crc kubenswrapper[4754]: I0126 16:47:55.369227 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:55 crc kubenswrapper[4754]: I0126 16:47:55.369241 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:55 crc kubenswrapper[4754]: I0126 16:47:55.369252 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:55Z","lastTransitionTime":"2026-01-26T16:47:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:55 crc kubenswrapper[4754]: I0126 16:47:55.472752 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:55 crc kubenswrapper[4754]: I0126 16:47:55.472817 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:55 crc kubenswrapper[4754]: I0126 16:47:55.472828 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:55 crc kubenswrapper[4754]: I0126 16:47:55.472852 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:55 crc kubenswrapper[4754]: I0126 16:47:55.472865 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:55Z","lastTransitionTime":"2026-01-26T16:47:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:55 crc kubenswrapper[4754]: I0126 16:47:55.574872 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:55 crc kubenswrapper[4754]: I0126 16:47:55.574922 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:55 crc kubenswrapper[4754]: I0126 16:47:55.574932 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:55 crc kubenswrapper[4754]: I0126 16:47:55.574949 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:55 crc kubenswrapper[4754]: I0126 16:47:55.574965 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:55Z","lastTransitionTime":"2026-01-26T16:47:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:55 crc kubenswrapper[4754]: I0126 16:47:55.678809 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:55 crc kubenswrapper[4754]: I0126 16:47:55.678867 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:55 crc kubenswrapper[4754]: I0126 16:47:55.678881 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:55 crc kubenswrapper[4754]: I0126 16:47:55.678899 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:55 crc kubenswrapper[4754]: I0126 16:47:55.678911 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:55Z","lastTransitionTime":"2026-01-26T16:47:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:55 crc kubenswrapper[4754]: I0126 16:47:55.737994 4754 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-21 04:48:31.693308924 +0000 UTC Jan 26 16:47:55 crc kubenswrapper[4754]: I0126 16:47:55.766974 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 16:47:55 crc kubenswrapper[4754]: I0126 16:47:55.767015 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 16:47:55 crc kubenswrapper[4754]: I0126 16:47:55.766993 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 16:47:55 crc kubenswrapper[4754]: E0126 16:47:55.767100 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 16:47:55 crc kubenswrapper[4754]: I0126 16:47:55.766988 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-t4ps6" Jan 26 16:47:55 crc kubenswrapper[4754]: E0126 16:47:55.767279 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 16:47:55 crc kubenswrapper[4754]: E0126 16:47:55.767376 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-t4ps6" podUID="d45179f0-9f81-4d77-8805-6b05bdb32f8f" Jan 26 16:47:55 crc kubenswrapper[4754]: E0126 16:47:55.767515 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 16:47:55 crc kubenswrapper[4754]: I0126 16:47:55.780713 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:55 crc kubenswrapper[4754]: I0126 16:47:55.780745 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:55 crc kubenswrapper[4754]: I0126 16:47:55.780756 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:55 crc kubenswrapper[4754]: I0126 16:47:55.780773 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:55 crc kubenswrapper[4754]: I0126 16:47:55.780784 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:55Z","lastTransitionTime":"2026-01-26T16:47:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:55 crc kubenswrapper[4754]: I0126 16:47:55.883043 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:55 crc kubenswrapper[4754]: I0126 16:47:55.883099 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:55 crc kubenswrapper[4754]: I0126 16:47:55.883115 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:55 crc kubenswrapper[4754]: I0126 16:47:55.883133 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:55 crc kubenswrapper[4754]: I0126 16:47:55.883145 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:55Z","lastTransitionTime":"2026-01-26T16:47:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:55 crc kubenswrapper[4754]: I0126 16:47:55.985484 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:55 crc kubenswrapper[4754]: I0126 16:47:55.985526 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:55 crc kubenswrapper[4754]: I0126 16:47:55.985538 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:55 crc kubenswrapper[4754]: I0126 16:47:55.985554 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:55 crc kubenswrapper[4754]: I0126 16:47:55.985566 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:55Z","lastTransitionTime":"2026-01-26T16:47:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:56 crc kubenswrapper[4754]: I0126 16:47:56.088164 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:56 crc kubenswrapper[4754]: I0126 16:47:56.088215 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:56 crc kubenswrapper[4754]: I0126 16:47:56.088227 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:56 crc kubenswrapper[4754]: I0126 16:47:56.088242 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:56 crc kubenswrapper[4754]: I0126 16:47:56.088250 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:56Z","lastTransitionTime":"2026-01-26T16:47:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:56 crc kubenswrapper[4754]: I0126 16:47:56.171247 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d45179f0-9f81-4d77-8805-6b05bdb32f8f-metrics-certs\") pod \"network-metrics-daemon-t4ps6\" (UID: \"d45179f0-9f81-4d77-8805-6b05bdb32f8f\") " pod="openshift-multus/network-metrics-daemon-t4ps6" Jan 26 16:47:56 crc kubenswrapper[4754]: E0126 16:47:56.171363 4754 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Jan 26 16:47:56 crc kubenswrapper[4754]: E0126 16:47:56.171406 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d45179f0-9f81-4d77-8805-6b05bdb32f8f-metrics-certs podName:d45179f0-9f81-4d77-8805-6b05bdb32f8f nodeName:}" failed. No retries permitted until 2026-01-26 16:48:04.171393365 +0000 UTC m=+50.695573799 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/d45179f0-9f81-4d77-8805-6b05bdb32f8f-metrics-certs") pod "network-metrics-daemon-t4ps6" (UID: "d45179f0-9f81-4d77-8805-6b05bdb32f8f") : object "openshift-multus"/"metrics-daemon-secret" not registered Jan 26 16:47:56 crc kubenswrapper[4754]: I0126 16:47:56.190620 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:56 crc kubenswrapper[4754]: I0126 16:47:56.190710 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:56 crc kubenswrapper[4754]: I0126 16:47:56.190721 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:56 crc kubenswrapper[4754]: I0126 16:47:56.190737 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:56 crc kubenswrapper[4754]: I0126 16:47:56.190754 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:56Z","lastTransitionTime":"2026-01-26T16:47:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:56 crc kubenswrapper[4754]: I0126 16:47:56.293499 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:56 crc kubenswrapper[4754]: I0126 16:47:56.293545 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:56 crc kubenswrapper[4754]: I0126 16:47:56.293554 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:56 crc kubenswrapper[4754]: I0126 16:47:56.293568 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:56 crc kubenswrapper[4754]: I0126 16:47:56.293579 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:56Z","lastTransitionTime":"2026-01-26T16:47:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:56 crc kubenswrapper[4754]: I0126 16:47:56.396062 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:56 crc kubenswrapper[4754]: I0126 16:47:56.396124 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:56 crc kubenswrapper[4754]: I0126 16:47:56.396137 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:56 crc kubenswrapper[4754]: I0126 16:47:56.396153 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:56 crc kubenswrapper[4754]: I0126 16:47:56.396163 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:56Z","lastTransitionTime":"2026-01-26T16:47:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:56 crc kubenswrapper[4754]: I0126 16:47:56.498508 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:56 crc kubenswrapper[4754]: I0126 16:47:56.498580 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:56 crc kubenswrapper[4754]: I0126 16:47:56.498591 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:56 crc kubenswrapper[4754]: I0126 16:47:56.498611 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:56 crc kubenswrapper[4754]: I0126 16:47:56.498624 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:56Z","lastTransitionTime":"2026-01-26T16:47:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:56 crc kubenswrapper[4754]: I0126 16:47:56.601511 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:56 crc kubenswrapper[4754]: I0126 16:47:56.601576 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:56 crc kubenswrapper[4754]: I0126 16:47:56.601591 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:56 crc kubenswrapper[4754]: I0126 16:47:56.601756 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:56 crc kubenswrapper[4754]: I0126 16:47:56.601792 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:56Z","lastTransitionTime":"2026-01-26T16:47:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:56 crc kubenswrapper[4754]: I0126 16:47:56.704269 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:56 crc kubenswrapper[4754]: I0126 16:47:56.704350 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:56 crc kubenswrapper[4754]: I0126 16:47:56.704373 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:56 crc kubenswrapper[4754]: I0126 16:47:56.704404 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:56 crc kubenswrapper[4754]: I0126 16:47:56.704427 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:56Z","lastTransitionTime":"2026-01-26T16:47:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:56 crc kubenswrapper[4754]: I0126 16:47:56.738727 4754 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-10 05:07:54.503898111 +0000 UTC Jan 26 16:47:56 crc kubenswrapper[4754]: I0126 16:47:56.806469 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:56 crc kubenswrapper[4754]: I0126 16:47:56.806511 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:56 crc kubenswrapper[4754]: I0126 16:47:56.806523 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:56 crc kubenswrapper[4754]: I0126 16:47:56.806545 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:56 crc kubenswrapper[4754]: I0126 16:47:56.806557 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:56Z","lastTransitionTime":"2026-01-26T16:47:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:56 crc kubenswrapper[4754]: I0126 16:47:56.909471 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:56 crc kubenswrapper[4754]: I0126 16:47:56.909765 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:56 crc kubenswrapper[4754]: I0126 16:47:56.909856 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:56 crc kubenswrapper[4754]: I0126 16:47:56.909937 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:56 crc kubenswrapper[4754]: I0126 16:47:56.910016 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:56Z","lastTransitionTime":"2026-01-26T16:47:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:57 crc kubenswrapper[4754]: I0126 16:47:57.012216 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:57 crc kubenswrapper[4754]: I0126 16:47:57.012255 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:57 crc kubenswrapper[4754]: I0126 16:47:57.012264 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:57 crc kubenswrapper[4754]: I0126 16:47:57.012281 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:57 crc kubenswrapper[4754]: I0126 16:47:57.012292 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:57Z","lastTransitionTime":"2026-01-26T16:47:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:57 crc kubenswrapper[4754]: I0126 16:47:57.114818 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:57 crc kubenswrapper[4754]: I0126 16:47:57.114857 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:57 crc kubenswrapper[4754]: I0126 16:47:57.114871 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:57 crc kubenswrapper[4754]: I0126 16:47:57.114890 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:57 crc kubenswrapper[4754]: I0126 16:47:57.114903 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:57Z","lastTransitionTime":"2026-01-26T16:47:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:57 crc kubenswrapper[4754]: I0126 16:47:57.217285 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:57 crc kubenswrapper[4754]: I0126 16:47:57.217337 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:57 crc kubenswrapper[4754]: I0126 16:47:57.217349 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:57 crc kubenswrapper[4754]: I0126 16:47:57.217368 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:57 crc kubenswrapper[4754]: I0126 16:47:57.217379 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:57Z","lastTransitionTime":"2026-01-26T16:47:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:57 crc kubenswrapper[4754]: I0126 16:47:57.319535 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:57 crc kubenswrapper[4754]: I0126 16:47:57.319831 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:57 crc kubenswrapper[4754]: I0126 16:47:57.319923 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:57 crc kubenswrapper[4754]: I0126 16:47:57.319991 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:57 crc kubenswrapper[4754]: I0126 16:47:57.320062 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:57Z","lastTransitionTime":"2026-01-26T16:47:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:57 crc kubenswrapper[4754]: I0126 16:47:57.422900 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:57 crc kubenswrapper[4754]: I0126 16:47:57.422953 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:57 crc kubenswrapper[4754]: I0126 16:47:57.422969 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:57 crc kubenswrapper[4754]: I0126 16:47:57.422992 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:57 crc kubenswrapper[4754]: I0126 16:47:57.423007 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:57Z","lastTransitionTime":"2026-01-26T16:47:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:57 crc kubenswrapper[4754]: I0126 16:47:57.525770 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:57 crc kubenswrapper[4754]: I0126 16:47:57.525807 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:57 crc kubenswrapper[4754]: I0126 16:47:57.525818 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:57 crc kubenswrapper[4754]: I0126 16:47:57.525834 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:57 crc kubenswrapper[4754]: I0126 16:47:57.525846 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:57Z","lastTransitionTime":"2026-01-26T16:47:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:57 crc kubenswrapper[4754]: I0126 16:47:57.628719 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:57 crc kubenswrapper[4754]: I0126 16:47:57.628765 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:57 crc kubenswrapper[4754]: I0126 16:47:57.628776 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:57 crc kubenswrapper[4754]: I0126 16:47:57.628791 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:57 crc kubenswrapper[4754]: I0126 16:47:57.628805 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:57Z","lastTransitionTime":"2026-01-26T16:47:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:57 crc kubenswrapper[4754]: I0126 16:47:57.731773 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:57 crc kubenswrapper[4754]: I0126 16:47:57.731820 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:57 crc kubenswrapper[4754]: I0126 16:47:57.731829 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:57 crc kubenswrapper[4754]: I0126 16:47:57.731844 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:57 crc kubenswrapper[4754]: I0126 16:47:57.731854 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:57Z","lastTransitionTime":"2026-01-26T16:47:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:57 crc kubenswrapper[4754]: I0126 16:47:57.739182 4754 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-08 02:32:27.48874416 +0000 UTC Jan 26 16:47:57 crc kubenswrapper[4754]: I0126 16:47:57.766733 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 16:47:57 crc kubenswrapper[4754]: I0126 16:47:57.766782 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 16:47:57 crc kubenswrapper[4754]: E0126 16:47:57.766861 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 16:47:57 crc kubenswrapper[4754]: I0126 16:47:57.766958 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-t4ps6" Jan 26 16:47:57 crc kubenswrapper[4754]: I0126 16:47:57.767041 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 16:47:57 crc kubenswrapper[4754]: E0126 16:47:57.767431 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 16:47:57 crc kubenswrapper[4754]: E0126 16:47:57.767495 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-t4ps6" podUID="d45179f0-9f81-4d77-8805-6b05bdb32f8f" Jan 26 16:47:57 crc kubenswrapper[4754]: E0126 16:47:57.767581 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 16:47:57 crc kubenswrapper[4754]: I0126 16:47:57.767760 4754 scope.go:117] "RemoveContainer" containerID="eaed406e767c0eedc024810541a22ed6ac3b1dcc2b5d41aa007b9ad282eb57fd" Jan 26 16:47:57 crc kubenswrapper[4754]: I0126 16:47:57.834653 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:57 crc kubenswrapper[4754]: I0126 16:47:57.834711 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:57 crc kubenswrapper[4754]: I0126 16:47:57.834722 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:57 crc kubenswrapper[4754]: I0126 16:47:57.834737 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:57 crc kubenswrapper[4754]: I0126 16:47:57.834751 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:57Z","lastTransitionTime":"2026-01-26T16:47:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:57 crc kubenswrapper[4754]: I0126 16:47:57.936984 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:57 crc kubenswrapper[4754]: I0126 16:47:57.937014 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:57 crc kubenswrapper[4754]: I0126 16:47:57.937024 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:57 crc kubenswrapper[4754]: I0126 16:47:57.937041 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:57 crc kubenswrapper[4754]: I0126 16:47:57.937050 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:57Z","lastTransitionTime":"2026-01-26T16:47:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:58 crc kubenswrapper[4754]: I0126 16:47:58.040366 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:58 crc kubenswrapper[4754]: I0126 16:47:58.040417 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:58 crc kubenswrapper[4754]: I0126 16:47:58.040429 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:58 crc kubenswrapper[4754]: I0126 16:47:58.040445 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:58 crc kubenswrapper[4754]: I0126 16:47:58.040458 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:58Z","lastTransitionTime":"2026-01-26T16:47:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:58 crc kubenswrapper[4754]: I0126 16:47:58.072371 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-jsbxt_689569c4-93ee-4b82-93fc-9af0f97e6dc3/ovnkube-controller/1.log" Jan 26 16:47:58 crc kubenswrapper[4754]: I0126 16:47:58.074938 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jsbxt" event={"ID":"689569c4-93ee-4b82-93fc-9af0f97e6dc3","Type":"ContainerStarted","Data":"0e9d45c28e67d857b1c50ed78197ed052c1ac2b98f1c8a94addd63698bf0bb23"} Jan 26 16:47:58 crc kubenswrapper[4754]: I0126 16:47:58.075443 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-jsbxt" Jan 26 16:47:58 crc kubenswrapper[4754]: I0126 16:47:58.092268 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:58Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:58 crc kubenswrapper[4754]: I0126 16:47:58.110476 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e459fd8c1093d6f3d0a671751b792b26b7df94c21ced1be1a2250821bcc82d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:58Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:58 crc kubenswrapper[4754]: I0126 16:47:58.124311 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2blzw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca65468e-30c6-4666-962e-cc0de05e67fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4603235bb2af495901926b805fe1c0ffc48824a4394ccf928e7fcd53a673ac3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c25z4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:33Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2blzw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:58Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:58 crc kubenswrapper[4754]: I0126 16:47:58.141784 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bwpd6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"351cc6c0-b8e5-4589-ae63-f98e8ab1eca3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1022f101d36873b41bbc1ed8a4a2590347b49f7cce6a172aa1d5bc9f529aed65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3ce0fe080a094316b9456057e347b90a5ff432daa12b9c56d7aea9bee1f72b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3ce0fe080a094316b9456057e347b90a5ff432daa12b9c56d7aea9bee1f72b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1772bb41eb8a36a863d3c691610bcbee5d16b260480b3e9f9d6b083e09e42799\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1772bb41eb8a36a863d3c691610bcbee5d16b260480b3e9f9d6b083e09e42799\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24848d69756024cc664affb9492764ace0638ef611f910f4f2bf86b638cc443d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://24848d69756024cc664affb9492764ace0638ef611f910f4f2bf86b638cc443d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c264fc7c3300a725566acce4464f01d9314779737624401c164d867831fd93bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c264fc7c3300a725566acce4464f01d9314779737624401c164d867831fd93bf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5acfe327eee4aad927ac3a32d24549eff788b14a1c4bdac359b6b56301d884ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5acfe327eee4aad927ac3a32d24549eff788b14a1c4bdac359b6b56301d884ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://63a9b9317494d0e0b33970ef3d461729b10d2a1ad7d3d2bd33f389b184ca56ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://63a9b9317494d0e0b33970ef3d461729b10d2a1ad7d3d2bd33f389b184ca56ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bwpd6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:58Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:58 crc kubenswrapper[4754]: I0126 16:47:58.142985 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:58 crc kubenswrapper[4754]: I0126 16:47:58.143037 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:58 crc kubenswrapper[4754]: I0126 16:47:58.143052 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:58 crc kubenswrapper[4754]: I0126 16:47:58.143070 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:58 crc kubenswrapper[4754]: I0126 16:47:58.143085 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:58Z","lastTransitionTime":"2026-01-26T16:47:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:58 crc kubenswrapper[4754]: I0126 16:47:58.171712 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jsbxt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"689569c4-93ee-4b82-93fc-9af0f97e6dc3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://033f8b104d4e9ff6d5f87ca998665f12e0c56696b3d30e01ede05f4da540365d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://919572e731813a98c7d8a94879bba931f75c472c7cd681e86d5197ebab43815d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1ac1f49bd564750fcff530a0f76657b0ed41573539acd020a922d19861ef834\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a300a80ad87f46ba1b4fd7ad3cf6c25767fcfe966b93bdd86d729843f53a5cda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31dcab2f9a2d218b0e49dad1df46377ff4521592824a0fbac254d5b8e6d0632e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fe4d9395447ac570eca1338e749dbefea76295d427a12487b63a142a7d6a2e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e9d45c28e67d857b1c50ed78197ed052c1ac2b98f1c8a94addd63698bf0bb23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eaed406e767c0eedc024810541a22ed6ac3b1dcc2b5d41aa007b9ad282eb57fd\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-26T16:47:45Z\\\",\\\"message\\\":\\\".EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0126 16:47:45.693395 6233 handler.go:208] Removed *v1.Node event handler 7\\\\nI0126 16:47:45.694648 6233 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI0126 16:47:45.694956 6233 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI0126 16:47:45.694961 6233 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0126 16:47:45.695763 6233 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0126 16:47:45.695841 6233 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0126 16:47:45.695852 6233 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0126 16:47:45.695893 6233 factory.go:656] Stopping watch factory\\\\nI0126 16:47:45.695913 6233 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0126 16:47:45.695920 6233 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0126 16:47:45.695935 6233 handler.go:208] Removed *v1.Namespace ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:44Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://07a43ff7860ff65faf51f4a5e8aeb326f06d6d7222677a92560ae2e3a0c81484\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d9a7bafabd44f8b88a6f62cbe45599af8563bc45d9d7ad3a7669f3d8d11eb9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d9a7bafabd44f8b88a6f62cbe45599af8563bc45d9d7ad3a7669f3d8d11eb9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jsbxt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:58Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:58 crc kubenswrapper[4754]: I0126 16:47:58.187782 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"099af702-abf2-4933-b4c9-63ef2326e468\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23017e73f8450839b4c2c02fa61235ffe4604bdce11556def27f25bcadcf89dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a74507c941b9de4d0401c8f1d53bb79ed11359628d3405cdab7f5474a75c862e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4402182b3e2e35a7d5dd4c44ac53e152915b312b7ccb076c4f2b41b925a6e7f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f772a003566198519fe1849fd14f29d4346004b2e655810662507c95b4d5ffa1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:13Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:58Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:58 crc kubenswrapper[4754]: I0126 16:47:58.210870 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:58Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:58 crc kubenswrapper[4754]: I0126 16:47:58.227284 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89251fa9c64bc25d6d6f355400d7644c099f22aed1cae812fc991a03595f1e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e90508d8a53be9146f9b388750730fdae9ee200306caf5ad393a356390a69ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:58Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:58 crc kubenswrapper[4754]: I0126 16:47:58.238923 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-84lgq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6763d962-0892-4acc-bab9-7a4e93e39a85\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://953c1917fca896c3def3fc2c69f9e3dbdc9ca26fc2b3d40282be71f71759fd68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8zlw6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:33Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-84lgq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:58Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:58 crc kubenswrapper[4754]: I0126 16:47:58.245649 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:58 crc kubenswrapper[4754]: I0126 16:47:58.245702 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:58 crc kubenswrapper[4754]: I0126 16:47:58.245713 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:58 crc kubenswrapper[4754]: I0126 16:47:58.245729 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:58 crc kubenswrapper[4754]: I0126 16:47:58.245742 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:58Z","lastTransitionTime":"2026-01-26T16:47:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:58 crc kubenswrapper[4754]: I0126 16:47:58.252358 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-skh5n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b619d233-b592-4b05-a0b7-dc094c88471e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9231a4191565e2e0ef53d8a394bb34615c900b601a0dfec24c2d8e5fa7b0b659\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-skh5n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:58Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:58 crc kubenswrapper[4754]: I0126 16:47:58.264212 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jspvh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5c81aab0-939e-45ce-b19a-89d4818980c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://356dc54150a495dd6f35e0667c82b807b712b07200d8a0adbb1c92fe3f7af8f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8bps\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://05fd57780c44cd127859dc8b163d20eafbd8b06f3348cf545ef41021ce6bc526\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8bps\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-jspvh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:58Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:58 crc kubenswrapper[4754]: I0126 16:47:58.279396 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b89aa081-0a2e-4757-9210-e2a6c228bee9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9c5f40188fda14b14128514fe84bddfe8dcede3d6c560283ccb2ac0c66895fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f45445cbf1d180d6d4dd9588e85bf678f043bfce8cff0812b78ca3510904f9a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://085fa7b8d4fef8b2722fac8c346107699b5a6bd33fbfc74ed35831cb8fdf9dad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c40e018a0979035827963efb9efc48a878a55b9df68ec1a9242d7cff64fe480\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8356f76fbc3a4e636c12d1da5455d3f6767d9b83ce2f2925a11435d7d7261e9f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0126 16:47:26.205208 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0126 16:47:26.206538 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3645934156/tls.crt::/tmp/serving-cert-3645934156/tls.key\\\\\\\"\\\\nI0126 16:47:31.504518 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0126 16:47:31.509366 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0126 16:47:31.509400 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0126 16:47:31.509428 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0126 16:47:31.509436 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0126 16:47:31.515912 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0126 16:47:31.516036 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 16:47:31.516068 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 16:47:31.516096 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0126 16:47:31.516121 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0126 16:47:31.516145 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0126 16:47:31.516170 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0126 16:47:31.516412 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0126 16:47:31.519294 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27b74e2e65a547e0e28fddd4f15feacb17f4d57596eeb8da82df9d39243c85e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22ba7649a1308815bfad52016000987a7a9a5bc214d7ff61e137ba099b20fc6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://22ba7649a1308815bfad52016000987a7a9a5bc214d7ff61e137ba099b20fc6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:13Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:58Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:58 crc kubenswrapper[4754]: I0126 16:47:58.294588 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://205ec256e46a8e08b4b0b0805fb6c5f4966fc3ccef0738df4b97cb404e74da17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:58Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:58 crc kubenswrapper[4754]: I0126 16:47:58.308992 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:58Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:58 crc kubenswrapper[4754]: I0126 16:47:58.322760 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c3718a4-f354-4284-92e0-fdfb45a692bd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b86f46f49163bca7eca9458279029af330dbd4890b03325e06fa3b1bdf030de6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxjcz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fffc03b536bfee3c7bc1bddd500e834e34b391896477a1501e3e06201a19374\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxjcz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-x65wv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:58Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:58 crc kubenswrapper[4754]: I0126 16:47:58.337717 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-t4ps6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d45179f0-9f81-4d77-8805-6b05bdb32f8f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vk8tg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vk8tg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:48Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-t4ps6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:58Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:58 crc kubenswrapper[4754]: I0126 16:47:58.348425 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:58 crc kubenswrapper[4754]: I0126 16:47:58.348491 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:58 crc kubenswrapper[4754]: I0126 16:47:58.348546 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:58 crc kubenswrapper[4754]: I0126 16:47:58.348584 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:58 crc kubenswrapper[4754]: I0126 16:47:58.348595 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:58Z","lastTransitionTime":"2026-01-26T16:47:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:58 crc kubenswrapper[4754]: I0126 16:47:58.451300 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:58 crc kubenswrapper[4754]: I0126 16:47:58.451345 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:58 crc kubenswrapper[4754]: I0126 16:47:58.451356 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:58 crc kubenswrapper[4754]: I0126 16:47:58.451375 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:58 crc kubenswrapper[4754]: I0126 16:47:58.451387 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:58Z","lastTransitionTime":"2026-01-26T16:47:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:58 crc kubenswrapper[4754]: I0126 16:47:58.554341 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:58 crc kubenswrapper[4754]: I0126 16:47:58.554384 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:58 crc kubenswrapper[4754]: I0126 16:47:58.554394 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:58 crc kubenswrapper[4754]: I0126 16:47:58.554409 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:58 crc kubenswrapper[4754]: I0126 16:47:58.554420 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:58Z","lastTransitionTime":"2026-01-26T16:47:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:58 crc kubenswrapper[4754]: I0126 16:47:58.656154 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:58 crc kubenswrapper[4754]: I0126 16:47:58.656189 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:58 crc kubenswrapper[4754]: I0126 16:47:58.656197 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:58 crc kubenswrapper[4754]: I0126 16:47:58.656211 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:58 crc kubenswrapper[4754]: I0126 16:47:58.656221 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:58Z","lastTransitionTime":"2026-01-26T16:47:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:58 crc kubenswrapper[4754]: I0126 16:47:58.739777 4754 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-23 03:56:18.313908489 +0000 UTC Jan 26 16:47:58 crc kubenswrapper[4754]: I0126 16:47:58.759884 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:58 crc kubenswrapper[4754]: I0126 16:47:58.759927 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:58 crc kubenswrapper[4754]: I0126 16:47:58.759938 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:58 crc kubenswrapper[4754]: I0126 16:47:58.759955 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:58 crc kubenswrapper[4754]: I0126 16:47:58.759967 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:58Z","lastTransitionTime":"2026-01-26T16:47:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:58 crc kubenswrapper[4754]: I0126 16:47:58.863387 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:58 crc kubenswrapper[4754]: I0126 16:47:58.863448 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:58 crc kubenswrapper[4754]: I0126 16:47:58.863461 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:58 crc kubenswrapper[4754]: I0126 16:47:58.863480 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:58 crc kubenswrapper[4754]: I0126 16:47:58.863492 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:58Z","lastTransitionTime":"2026-01-26T16:47:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:58 crc kubenswrapper[4754]: I0126 16:47:58.965854 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:58 crc kubenswrapper[4754]: I0126 16:47:58.965903 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:58 crc kubenswrapper[4754]: I0126 16:47:58.965912 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:58 crc kubenswrapper[4754]: I0126 16:47:58.965929 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:58 crc kubenswrapper[4754]: I0126 16:47:58.965942 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:58Z","lastTransitionTime":"2026-01-26T16:47:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:59 crc kubenswrapper[4754]: I0126 16:47:59.068342 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:59 crc kubenswrapper[4754]: I0126 16:47:59.068380 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:59 crc kubenswrapper[4754]: I0126 16:47:59.068390 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:59 crc kubenswrapper[4754]: I0126 16:47:59.068404 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:59 crc kubenswrapper[4754]: I0126 16:47:59.068413 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:59Z","lastTransitionTime":"2026-01-26T16:47:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:59 crc kubenswrapper[4754]: I0126 16:47:59.079904 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-jsbxt_689569c4-93ee-4b82-93fc-9af0f97e6dc3/ovnkube-controller/2.log" Jan 26 16:47:59 crc kubenswrapper[4754]: I0126 16:47:59.080485 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-jsbxt_689569c4-93ee-4b82-93fc-9af0f97e6dc3/ovnkube-controller/1.log" Jan 26 16:47:59 crc kubenswrapper[4754]: I0126 16:47:59.083129 4754 generic.go:334] "Generic (PLEG): container finished" podID="689569c4-93ee-4b82-93fc-9af0f97e6dc3" containerID="0e9d45c28e67d857b1c50ed78197ed052c1ac2b98f1c8a94addd63698bf0bb23" exitCode=1 Jan 26 16:47:59 crc kubenswrapper[4754]: I0126 16:47:59.083170 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jsbxt" event={"ID":"689569c4-93ee-4b82-93fc-9af0f97e6dc3","Type":"ContainerDied","Data":"0e9d45c28e67d857b1c50ed78197ed052c1ac2b98f1c8a94addd63698bf0bb23"} Jan 26 16:47:59 crc kubenswrapper[4754]: I0126 16:47:59.083211 4754 scope.go:117] "RemoveContainer" containerID="eaed406e767c0eedc024810541a22ed6ac3b1dcc2b5d41aa007b9ad282eb57fd" Jan 26 16:47:59 crc kubenswrapper[4754]: I0126 16:47:59.084640 4754 scope.go:117] "RemoveContainer" containerID="0e9d45c28e67d857b1c50ed78197ed052c1ac2b98f1c8a94addd63698bf0bb23" Jan 26 16:47:59 crc kubenswrapper[4754]: E0126 16:47:59.085009 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-jsbxt_openshift-ovn-kubernetes(689569c4-93ee-4b82-93fc-9af0f97e6dc3)\"" pod="openshift-ovn-kubernetes/ovnkube-node-jsbxt" podUID="689569c4-93ee-4b82-93fc-9af0f97e6dc3" Jan 26 16:47:59 crc kubenswrapper[4754]: I0126 16:47:59.108443 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jsbxt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"689569c4-93ee-4b82-93fc-9af0f97e6dc3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://033f8b104d4e9ff6d5f87ca998665f12e0c56696b3d30e01ede05f4da540365d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://919572e731813a98c7d8a94879bba931f75c472c7cd681e86d5197ebab43815d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1ac1f49bd564750fcff530a0f76657b0ed41573539acd020a922d19861ef834\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a300a80ad87f46ba1b4fd7ad3cf6c25767fcfe966b93bdd86d729843f53a5cda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31dcab2f9a2d218b0e49dad1df46377ff4521592824a0fbac254d5b8e6d0632e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fe4d9395447ac570eca1338e749dbefea76295d427a12487b63a142a7d6a2e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e9d45c28e67d857b1c50ed78197ed052c1ac2b98f1c8a94addd63698bf0bb23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eaed406e767c0eedc024810541a22ed6ac3b1dcc2b5d41aa007b9ad282eb57fd\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-26T16:47:45Z\\\",\\\"message\\\":\\\".EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0126 16:47:45.693395 6233 handler.go:208] Removed *v1.Node event handler 7\\\\nI0126 16:47:45.694648 6233 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI0126 16:47:45.694956 6233 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI0126 16:47:45.694961 6233 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0126 16:47:45.695763 6233 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0126 16:47:45.695841 6233 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0126 16:47:45.695852 6233 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0126 16:47:45.695893 6233 factory.go:656] Stopping watch factory\\\\nI0126 16:47:45.695913 6233 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0126 16:47:45.695920 6233 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0126 16:47:45.695935 6233 handler.go:208] Removed *v1.Namespace ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:44Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0e9d45c28e67d857b1c50ed78197ed052c1ac2b98f1c8a94addd63698bf0bb23\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-26T16:47:58Z\\\",\\\"message\\\":\\\":303] Retry object setup: *v1.Pod openshift-machine-config-operator/machine-config-daemon-x65wv\\\\nI0126 16:47:58.625627 6452 obj_retry.go:303] Retry object setup: *v1.Pod openshift-multus/multus-skh5n\\\\nI0126 16:47:58.625718 6452 obj_retry.go:365] Adding new object: *v1.Pod openshift-machine-config-operator/machine-config-daemon-x65wv\\\\nI0126 16:47:58.625617 6452 obj_retry.go:303] Retry object setup: *v1.Pod openshift-multus/multus-additional-cni-plugins-bwpd6\\\\nI0126 16:47:58.625726 6452 ovn.go:134] Ensuring zone local for Pod openshift-machine-config-operator/machine-config-daemon-x65wv in node crc\\\\nI0126 16:47:58.625727 6452 obj_retry.go:365] Adding new object: *v1.Pod openshift-multus/multus-skh5n\\\\nI0126 16:47:58.625732 6452 obj_retry.go:365] Adding new object: *v1.Pod openshift-multus/multus-additional-cni-plugins-bwpd6\\\\nI0126 16:47:58.625706 6452 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {eb8eef51-1a8d-43f9-ae2e-3b2cc00ded60}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0126 16:47:58.625740 6452 ovn.go:134] Ensuring zone local for Pod openshift-multus/multus-additional-cni-plugins-bwpd6 in no\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://07a43ff7860ff65faf51f4a5e8aeb326f06d6d7222677a92560ae2e3a0c81484\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d9a7bafabd44f8b88a6f62cbe45599af8563bc45d9d7ad3a7669f3d8d11eb9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d9a7bafabd44f8b88a6f62cbe45599af8563bc45d9d7ad3a7669f3d8d11eb9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jsbxt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:59Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:59 crc kubenswrapper[4754]: I0126 16:47:59.121156 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:59Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:59 crc kubenswrapper[4754]: I0126 16:47:59.132499 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e459fd8c1093d6f3d0a671751b792b26b7df94c21ced1be1a2250821bcc82d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:59Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:59 crc kubenswrapper[4754]: I0126 16:47:59.143792 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2blzw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca65468e-30c6-4666-962e-cc0de05e67fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4603235bb2af495901926b805fe1c0ffc48824a4394ccf928e7fcd53a673ac3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c25z4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:33Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2blzw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:59Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:59 crc kubenswrapper[4754]: I0126 16:47:59.157525 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bwpd6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"351cc6c0-b8e5-4589-ae63-f98e8ab1eca3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1022f101d36873b41bbc1ed8a4a2590347b49f7cce6a172aa1d5bc9f529aed65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3ce0fe080a094316b9456057e347b90a5ff432daa12b9c56d7aea9bee1f72b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3ce0fe080a094316b9456057e347b90a5ff432daa12b9c56d7aea9bee1f72b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1772bb41eb8a36a863d3c691610bcbee5d16b260480b3e9f9d6b083e09e42799\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1772bb41eb8a36a863d3c691610bcbee5d16b260480b3e9f9d6b083e09e42799\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24848d69756024cc664affb9492764ace0638ef611f910f4f2bf86b638cc443d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://24848d69756024cc664affb9492764ace0638ef611f910f4f2bf86b638cc443d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c264fc7c3300a725566acce4464f01d9314779737624401c164d867831fd93bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c264fc7c3300a725566acce4464f01d9314779737624401c164d867831fd93bf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5acfe327eee4aad927ac3a32d24549eff788b14a1c4bdac359b6b56301d884ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5acfe327eee4aad927ac3a32d24549eff788b14a1c4bdac359b6b56301d884ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://63a9b9317494d0e0b33970ef3d461729b10d2a1ad7d3d2bd33f389b184ca56ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://63a9b9317494d0e0b33970ef3d461729b10d2a1ad7d3d2bd33f389b184ca56ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bwpd6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:59Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:59 crc kubenswrapper[4754]: I0126 16:47:59.171794 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-skh5n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b619d233-b592-4b05-a0b7-dc094c88471e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9231a4191565e2e0ef53d8a394bb34615c900b601a0dfec24c2d8e5fa7b0b659\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-skh5n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:59Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:59 crc kubenswrapper[4754]: I0126 16:47:59.171870 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:59 crc kubenswrapper[4754]: I0126 16:47:59.171899 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:59 crc kubenswrapper[4754]: I0126 16:47:59.171910 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:59 crc kubenswrapper[4754]: I0126 16:47:59.171934 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:59 crc kubenswrapper[4754]: I0126 16:47:59.171946 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:59Z","lastTransitionTime":"2026-01-26T16:47:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:59 crc kubenswrapper[4754]: I0126 16:47:59.183305 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jspvh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5c81aab0-939e-45ce-b19a-89d4818980c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://356dc54150a495dd6f35e0667c82b807b712b07200d8a0adbb1c92fe3f7af8f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8bps\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://05fd57780c44cd127859dc8b163d20eafbd8b06f3348cf545ef41021ce6bc526\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8bps\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-jspvh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:59Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:59 crc kubenswrapper[4754]: I0126 16:47:59.194502 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"099af702-abf2-4933-b4c9-63ef2326e468\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23017e73f8450839b4c2c02fa61235ffe4604bdce11556def27f25bcadcf89dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a74507c941b9de4d0401c8f1d53bb79ed11359628d3405cdab7f5474a75c862e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4402182b3e2e35a7d5dd4c44ac53e152915b312b7ccb076c4f2b41b925a6e7f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f772a003566198519fe1849fd14f29d4346004b2e655810662507c95b4d5ffa1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:13Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:59Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:59 crc kubenswrapper[4754]: I0126 16:47:59.206706 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:59Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:59 crc kubenswrapper[4754]: I0126 16:47:59.220693 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89251fa9c64bc25d6d6f355400d7644c099f22aed1cae812fc991a03595f1e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e90508d8a53be9146f9b388750730fdae9ee200306caf5ad393a356390a69ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:59Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:59 crc kubenswrapper[4754]: I0126 16:47:59.231408 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-84lgq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6763d962-0892-4acc-bab9-7a4e93e39a85\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://953c1917fca896c3def3fc2c69f9e3dbdc9ca26fc2b3d40282be71f71759fd68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8zlw6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:33Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-84lgq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:59Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:59 crc kubenswrapper[4754]: I0126 16:47:59.244968 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b89aa081-0a2e-4757-9210-e2a6c228bee9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9c5f40188fda14b14128514fe84bddfe8dcede3d6c560283ccb2ac0c66895fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f45445cbf1d180d6d4dd9588e85bf678f043bfce8cff0812b78ca3510904f9a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://085fa7b8d4fef8b2722fac8c346107699b5a6bd33fbfc74ed35831cb8fdf9dad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c40e018a0979035827963efb9efc48a878a55b9df68ec1a9242d7cff64fe480\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8356f76fbc3a4e636c12d1da5455d3f6767d9b83ce2f2925a11435d7d7261e9f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0126 16:47:26.205208 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0126 16:47:26.206538 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3645934156/tls.crt::/tmp/serving-cert-3645934156/tls.key\\\\\\\"\\\\nI0126 16:47:31.504518 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0126 16:47:31.509366 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0126 16:47:31.509400 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0126 16:47:31.509428 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0126 16:47:31.509436 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0126 16:47:31.515912 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0126 16:47:31.516036 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 16:47:31.516068 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 16:47:31.516096 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0126 16:47:31.516121 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0126 16:47:31.516145 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0126 16:47:31.516170 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0126 16:47:31.516412 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0126 16:47:31.519294 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27b74e2e65a547e0e28fddd4f15feacb17f4d57596eeb8da82df9d39243c85e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22ba7649a1308815bfad52016000987a7a9a5bc214d7ff61e137ba099b20fc6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://22ba7649a1308815bfad52016000987a7a9a5bc214d7ff61e137ba099b20fc6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:13Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:59Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:59 crc kubenswrapper[4754]: I0126 16:47:59.261467 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://205ec256e46a8e08b4b0b0805fb6c5f4966fc3ccef0738df4b97cb404e74da17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:59Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:59 crc kubenswrapper[4754]: I0126 16:47:59.273661 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:59 crc kubenswrapper[4754]: I0126 16:47:59.273712 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:59 crc kubenswrapper[4754]: I0126 16:47:59.273723 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:59 crc kubenswrapper[4754]: I0126 16:47:59.273737 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:59 crc kubenswrapper[4754]: I0126 16:47:59.273748 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:59Z","lastTransitionTime":"2026-01-26T16:47:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:59 crc kubenswrapper[4754]: I0126 16:47:59.275149 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:59Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:59 crc kubenswrapper[4754]: I0126 16:47:59.285193 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c3718a4-f354-4284-92e0-fdfb45a692bd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b86f46f49163bca7eca9458279029af330dbd4890b03325e06fa3b1bdf030de6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxjcz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fffc03b536bfee3c7bc1bddd500e834e34b391896477a1501e3e06201a19374\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxjcz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-x65wv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:59Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:59 crc kubenswrapper[4754]: I0126 16:47:59.294648 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-t4ps6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d45179f0-9f81-4d77-8805-6b05bdb32f8f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vk8tg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vk8tg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:48Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-t4ps6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:47:59Z is after 2025-08-24T17:21:41Z" Jan 26 16:47:59 crc kubenswrapper[4754]: I0126 16:47:59.375830 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:59 crc kubenswrapper[4754]: I0126 16:47:59.375872 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:59 crc kubenswrapper[4754]: I0126 16:47:59.375881 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:59 crc kubenswrapper[4754]: I0126 16:47:59.375897 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:59 crc kubenswrapper[4754]: I0126 16:47:59.375907 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:59Z","lastTransitionTime":"2026-01-26T16:47:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:59 crc kubenswrapper[4754]: I0126 16:47:59.478163 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:59 crc kubenswrapper[4754]: I0126 16:47:59.478190 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:59 crc kubenswrapper[4754]: I0126 16:47:59.478200 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:59 crc kubenswrapper[4754]: I0126 16:47:59.478215 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:59 crc kubenswrapper[4754]: I0126 16:47:59.478225 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:59Z","lastTransitionTime":"2026-01-26T16:47:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:59 crc kubenswrapper[4754]: I0126 16:47:59.580811 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:59 crc kubenswrapper[4754]: I0126 16:47:59.580860 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:59 crc kubenswrapper[4754]: I0126 16:47:59.580873 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:59 crc kubenswrapper[4754]: I0126 16:47:59.580894 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:59 crc kubenswrapper[4754]: I0126 16:47:59.580907 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:59Z","lastTransitionTime":"2026-01-26T16:47:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:59 crc kubenswrapper[4754]: I0126 16:47:59.684302 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:59 crc kubenswrapper[4754]: I0126 16:47:59.684355 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:59 crc kubenswrapper[4754]: I0126 16:47:59.684373 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:59 crc kubenswrapper[4754]: I0126 16:47:59.684396 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:59 crc kubenswrapper[4754]: I0126 16:47:59.684414 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:59Z","lastTransitionTime":"2026-01-26T16:47:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:59 crc kubenswrapper[4754]: I0126 16:47:59.740802 4754 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-05 12:56:54.090948453 +0000 UTC Jan 26 16:47:59 crc kubenswrapper[4754]: I0126 16:47:59.767245 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 16:47:59 crc kubenswrapper[4754]: I0126 16:47:59.767309 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-t4ps6" Jan 26 16:47:59 crc kubenswrapper[4754]: I0126 16:47:59.767405 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 16:47:59 crc kubenswrapper[4754]: E0126 16:47:59.767400 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 16:47:59 crc kubenswrapper[4754]: I0126 16:47:59.767450 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 16:47:59 crc kubenswrapper[4754]: E0126 16:47:59.767535 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 16:47:59 crc kubenswrapper[4754]: E0126 16:47:59.767587 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 16:47:59 crc kubenswrapper[4754]: E0126 16:47:59.767631 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-t4ps6" podUID="d45179f0-9f81-4d77-8805-6b05bdb32f8f" Jan 26 16:47:59 crc kubenswrapper[4754]: I0126 16:47:59.786477 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:59 crc kubenswrapper[4754]: I0126 16:47:59.786529 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:59 crc kubenswrapper[4754]: I0126 16:47:59.786543 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:59 crc kubenswrapper[4754]: I0126 16:47:59.786561 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:59 crc kubenswrapper[4754]: I0126 16:47:59.786575 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:59Z","lastTransitionTime":"2026-01-26T16:47:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:59 crc kubenswrapper[4754]: I0126 16:47:59.889221 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:59 crc kubenswrapper[4754]: I0126 16:47:59.889277 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:59 crc kubenswrapper[4754]: I0126 16:47:59.889287 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:59 crc kubenswrapper[4754]: I0126 16:47:59.889301 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:59 crc kubenswrapper[4754]: I0126 16:47:59.889315 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:59Z","lastTransitionTime":"2026-01-26T16:47:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:47:59 crc kubenswrapper[4754]: I0126 16:47:59.991239 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:47:59 crc kubenswrapper[4754]: I0126 16:47:59.991289 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:47:59 crc kubenswrapper[4754]: I0126 16:47:59.991302 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:47:59 crc kubenswrapper[4754]: I0126 16:47:59.991318 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:47:59 crc kubenswrapper[4754]: I0126 16:47:59.991330 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:47:59Z","lastTransitionTime":"2026-01-26T16:47:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:00 crc kubenswrapper[4754]: I0126 16:48:00.088580 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-jsbxt_689569c4-93ee-4b82-93fc-9af0f97e6dc3/ovnkube-controller/2.log" Jan 26 16:48:00 crc kubenswrapper[4754]: I0126 16:48:00.092172 4754 scope.go:117] "RemoveContainer" containerID="0e9d45c28e67d857b1c50ed78197ed052c1ac2b98f1c8a94addd63698bf0bb23" Jan 26 16:48:00 crc kubenswrapper[4754]: E0126 16:48:00.092313 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-jsbxt_openshift-ovn-kubernetes(689569c4-93ee-4b82-93fc-9af0f97e6dc3)\"" pod="openshift-ovn-kubernetes/ovnkube-node-jsbxt" podUID="689569c4-93ee-4b82-93fc-9af0f97e6dc3" Jan 26 16:48:00 crc kubenswrapper[4754]: I0126 16:48:00.093650 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:00 crc kubenswrapper[4754]: I0126 16:48:00.093704 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:00 crc kubenswrapper[4754]: I0126 16:48:00.093713 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:00 crc kubenswrapper[4754]: I0126 16:48:00.093727 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:00 crc kubenswrapper[4754]: I0126 16:48:00.093736 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:00Z","lastTransitionTime":"2026-01-26T16:48:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:00 crc kubenswrapper[4754]: I0126 16:48:00.105525 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:00Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:00 crc kubenswrapper[4754]: I0126 16:48:00.120562 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e459fd8c1093d6f3d0a671751b792b26b7df94c21ced1be1a2250821bcc82d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:00Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:00 crc kubenswrapper[4754]: I0126 16:48:00.130481 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2blzw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca65468e-30c6-4666-962e-cc0de05e67fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4603235bb2af495901926b805fe1c0ffc48824a4394ccf928e7fcd53a673ac3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c25z4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:33Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2blzw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:00Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:00 crc kubenswrapper[4754]: I0126 16:48:00.146106 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bwpd6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"351cc6c0-b8e5-4589-ae63-f98e8ab1eca3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1022f101d36873b41bbc1ed8a4a2590347b49f7cce6a172aa1d5bc9f529aed65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3ce0fe080a094316b9456057e347b90a5ff432daa12b9c56d7aea9bee1f72b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3ce0fe080a094316b9456057e347b90a5ff432daa12b9c56d7aea9bee1f72b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1772bb41eb8a36a863d3c691610bcbee5d16b260480b3e9f9d6b083e09e42799\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1772bb41eb8a36a863d3c691610bcbee5d16b260480b3e9f9d6b083e09e42799\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24848d69756024cc664affb9492764ace0638ef611f910f4f2bf86b638cc443d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://24848d69756024cc664affb9492764ace0638ef611f910f4f2bf86b638cc443d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c264fc7c3300a725566acce4464f01d9314779737624401c164d867831fd93bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c264fc7c3300a725566acce4464f01d9314779737624401c164d867831fd93bf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5acfe327eee4aad927ac3a32d24549eff788b14a1c4bdac359b6b56301d884ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5acfe327eee4aad927ac3a32d24549eff788b14a1c4bdac359b6b56301d884ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://63a9b9317494d0e0b33970ef3d461729b10d2a1ad7d3d2bd33f389b184ca56ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://63a9b9317494d0e0b33970ef3d461729b10d2a1ad7d3d2bd33f389b184ca56ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bwpd6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:00Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:00 crc kubenswrapper[4754]: I0126 16:48:00.166743 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jsbxt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"689569c4-93ee-4b82-93fc-9af0f97e6dc3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://033f8b104d4e9ff6d5f87ca998665f12e0c56696b3d30e01ede05f4da540365d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://919572e731813a98c7d8a94879bba931f75c472c7cd681e86d5197ebab43815d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1ac1f49bd564750fcff530a0f76657b0ed41573539acd020a922d19861ef834\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a300a80ad87f46ba1b4fd7ad3cf6c25767fcfe966b93bdd86d729843f53a5cda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31dcab2f9a2d218b0e49dad1df46377ff4521592824a0fbac254d5b8e6d0632e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fe4d9395447ac570eca1338e749dbefea76295d427a12487b63a142a7d6a2e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e9d45c28e67d857b1c50ed78197ed052c1ac2b98f1c8a94addd63698bf0bb23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0e9d45c28e67d857b1c50ed78197ed052c1ac2b98f1c8a94addd63698bf0bb23\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-26T16:47:58Z\\\",\\\"message\\\":\\\":303] Retry object setup: *v1.Pod openshift-machine-config-operator/machine-config-daemon-x65wv\\\\nI0126 16:47:58.625627 6452 obj_retry.go:303] Retry object setup: *v1.Pod openshift-multus/multus-skh5n\\\\nI0126 16:47:58.625718 6452 obj_retry.go:365] Adding new object: *v1.Pod openshift-machine-config-operator/machine-config-daemon-x65wv\\\\nI0126 16:47:58.625617 6452 obj_retry.go:303] Retry object setup: *v1.Pod openshift-multus/multus-additional-cni-plugins-bwpd6\\\\nI0126 16:47:58.625726 6452 ovn.go:134] Ensuring zone local for Pod openshift-machine-config-operator/machine-config-daemon-x65wv in node crc\\\\nI0126 16:47:58.625727 6452 obj_retry.go:365] Adding new object: *v1.Pod openshift-multus/multus-skh5n\\\\nI0126 16:47:58.625732 6452 obj_retry.go:365] Adding new object: *v1.Pod openshift-multus/multus-additional-cni-plugins-bwpd6\\\\nI0126 16:47:58.625706 6452 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {eb8eef51-1a8d-43f9-ae2e-3b2cc00ded60}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0126 16:47:58.625740 6452 ovn.go:134] Ensuring zone local for Pod openshift-multus/multus-additional-cni-plugins-bwpd6 in no\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:57Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-jsbxt_openshift-ovn-kubernetes(689569c4-93ee-4b82-93fc-9af0f97e6dc3)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://07a43ff7860ff65faf51f4a5e8aeb326f06d6d7222677a92560ae2e3a0c81484\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d9a7bafabd44f8b88a6f62cbe45599af8563bc45d9d7ad3a7669f3d8d11eb9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d9a7bafabd44f8b88a6f62cbe45599af8563bc45d9d7ad3a7669f3d8d11eb9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jsbxt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:00Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:00 crc kubenswrapper[4754]: I0126 16:48:00.185265 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"099af702-abf2-4933-b4c9-63ef2326e468\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23017e73f8450839b4c2c02fa61235ffe4604bdce11556def27f25bcadcf89dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a74507c941b9de4d0401c8f1d53bb79ed11359628d3405cdab7f5474a75c862e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4402182b3e2e35a7d5dd4c44ac53e152915b312b7ccb076c4f2b41b925a6e7f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f772a003566198519fe1849fd14f29d4346004b2e655810662507c95b4d5ffa1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:13Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:00Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:00 crc kubenswrapper[4754]: I0126 16:48:00.196774 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:00 crc kubenswrapper[4754]: I0126 16:48:00.196825 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:00 crc kubenswrapper[4754]: I0126 16:48:00.196839 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:00 crc kubenswrapper[4754]: I0126 16:48:00.196856 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:00 crc kubenswrapper[4754]: I0126 16:48:00.196867 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:00Z","lastTransitionTime":"2026-01-26T16:48:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:00 crc kubenswrapper[4754]: I0126 16:48:00.208060 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:00Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:00 crc kubenswrapper[4754]: I0126 16:48:00.223653 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89251fa9c64bc25d6d6f355400d7644c099f22aed1cae812fc991a03595f1e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e90508d8a53be9146f9b388750730fdae9ee200306caf5ad393a356390a69ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:00Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:00 crc kubenswrapper[4754]: I0126 16:48:00.239931 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-84lgq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6763d962-0892-4acc-bab9-7a4e93e39a85\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://953c1917fca896c3def3fc2c69f9e3dbdc9ca26fc2b3d40282be71f71759fd68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8zlw6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:33Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-84lgq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:00Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:00 crc kubenswrapper[4754]: I0126 16:48:00.254118 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-skh5n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b619d233-b592-4b05-a0b7-dc094c88471e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9231a4191565e2e0ef53d8a394bb34615c900b601a0dfec24c2d8e5fa7b0b659\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-skh5n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:00Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:00 crc kubenswrapper[4754]: I0126 16:48:00.268554 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jspvh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5c81aab0-939e-45ce-b19a-89d4818980c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://356dc54150a495dd6f35e0667c82b807b712b07200d8a0adbb1c92fe3f7af8f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8bps\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://05fd57780c44cd127859dc8b163d20eafbd8b06f3348cf545ef41021ce6bc526\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8bps\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-jspvh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:00Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:00 crc kubenswrapper[4754]: I0126 16:48:00.283177 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b89aa081-0a2e-4757-9210-e2a6c228bee9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9c5f40188fda14b14128514fe84bddfe8dcede3d6c560283ccb2ac0c66895fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f45445cbf1d180d6d4dd9588e85bf678f043bfce8cff0812b78ca3510904f9a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://085fa7b8d4fef8b2722fac8c346107699b5a6bd33fbfc74ed35831cb8fdf9dad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c40e018a0979035827963efb9efc48a878a55b9df68ec1a9242d7cff64fe480\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8356f76fbc3a4e636c12d1da5455d3f6767d9b83ce2f2925a11435d7d7261e9f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0126 16:47:26.205208 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0126 16:47:26.206538 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3645934156/tls.crt::/tmp/serving-cert-3645934156/tls.key\\\\\\\"\\\\nI0126 16:47:31.504518 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0126 16:47:31.509366 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0126 16:47:31.509400 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0126 16:47:31.509428 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0126 16:47:31.509436 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0126 16:47:31.515912 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0126 16:47:31.516036 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 16:47:31.516068 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 16:47:31.516096 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0126 16:47:31.516121 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0126 16:47:31.516145 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0126 16:47:31.516170 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0126 16:47:31.516412 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0126 16:47:31.519294 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27b74e2e65a547e0e28fddd4f15feacb17f4d57596eeb8da82df9d39243c85e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22ba7649a1308815bfad52016000987a7a9a5bc214d7ff61e137ba099b20fc6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://22ba7649a1308815bfad52016000987a7a9a5bc214d7ff61e137ba099b20fc6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:13Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:00Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:00 crc kubenswrapper[4754]: I0126 16:48:00.298066 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://205ec256e46a8e08b4b0b0805fb6c5f4966fc3ccef0738df4b97cb404e74da17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:00Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:00 crc kubenswrapper[4754]: I0126 16:48:00.299557 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:00 crc kubenswrapper[4754]: I0126 16:48:00.299612 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:00 crc kubenswrapper[4754]: I0126 16:48:00.299625 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:00 crc kubenswrapper[4754]: I0126 16:48:00.299644 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:00 crc kubenswrapper[4754]: I0126 16:48:00.299657 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:00Z","lastTransitionTime":"2026-01-26T16:48:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:00 crc kubenswrapper[4754]: I0126 16:48:00.315098 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:00Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:00 crc kubenswrapper[4754]: I0126 16:48:00.331890 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c3718a4-f354-4284-92e0-fdfb45a692bd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b86f46f49163bca7eca9458279029af330dbd4890b03325e06fa3b1bdf030de6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxjcz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fffc03b536bfee3c7bc1bddd500e834e34b391896477a1501e3e06201a19374\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxjcz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-x65wv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:00Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:00 crc kubenswrapper[4754]: I0126 16:48:00.346371 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-t4ps6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d45179f0-9f81-4d77-8805-6b05bdb32f8f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vk8tg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vk8tg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:48Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-t4ps6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:00Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:00 crc kubenswrapper[4754]: I0126 16:48:00.403037 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:00 crc kubenswrapper[4754]: I0126 16:48:00.403135 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:00 crc kubenswrapper[4754]: I0126 16:48:00.403339 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:00 crc kubenswrapper[4754]: I0126 16:48:00.403369 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:00 crc kubenswrapper[4754]: I0126 16:48:00.403384 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:00Z","lastTransitionTime":"2026-01-26T16:48:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:00 crc kubenswrapper[4754]: I0126 16:48:00.506176 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:00 crc kubenswrapper[4754]: I0126 16:48:00.506218 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:00 crc kubenswrapper[4754]: I0126 16:48:00.506227 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:00 crc kubenswrapper[4754]: I0126 16:48:00.506242 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:00 crc kubenswrapper[4754]: I0126 16:48:00.506257 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:00Z","lastTransitionTime":"2026-01-26T16:48:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:00 crc kubenswrapper[4754]: I0126 16:48:00.609268 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:00 crc kubenswrapper[4754]: I0126 16:48:00.609319 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:00 crc kubenswrapper[4754]: I0126 16:48:00.609331 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:00 crc kubenswrapper[4754]: I0126 16:48:00.609349 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:00 crc kubenswrapper[4754]: I0126 16:48:00.609360 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:00Z","lastTransitionTime":"2026-01-26T16:48:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:00 crc kubenswrapper[4754]: I0126 16:48:00.711630 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:00 crc kubenswrapper[4754]: I0126 16:48:00.711695 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:00 crc kubenswrapper[4754]: I0126 16:48:00.711705 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:00 crc kubenswrapper[4754]: I0126 16:48:00.711724 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:00 crc kubenswrapper[4754]: I0126 16:48:00.711735 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:00Z","lastTransitionTime":"2026-01-26T16:48:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:00 crc kubenswrapper[4754]: I0126 16:48:00.741465 4754 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-22 03:33:32.136136635 +0000 UTC Jan 26 16:48:00 crc kubenswrapper[4754]: I0126 16:48:00.813507 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:00 crc kubenswrapper[4754]: I0126 16:48:00.813553 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:00 crc kubenswrapper[4754]: I0126 16:48:00.813564 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:00 crc kubenswrapper[4754]: I0126 16:48:00.813579 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:00 crc kubenswrapper[4754]: I0126 16:48:00.813591 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:00Z","lastTransitionTime":"2026-01-26T16:48:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:00 crc kubenswrapper[4754]: I0126 16:48:00.916304 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:00 crc kubenswrapper[4754]: I0126 16:48:00.916364 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:00 crc kubenswrapper[4754]: I0126 16:48:00.916373 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:00 crc kubenswrapper[4754]: I0126 16:48:00.916388 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:00 crc kubenswrapper[4754]: I0126 16:48:00.916400 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:00Z","lastTransitionTime":"2026-01-26T16:48:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:01 crc kubenswrapper[4754]: I0126 16:48:01.019340 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:01 crc kubenswrapper[4754]: I0126 16:48:01.019916 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:01 crc kubenswrapper[4754]: I0126 16:48:01.019927 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:01 crc kubenswrapper[4754]: I0126 16:48:01.019946 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:01 crc kubenswrapper[4754]: I0126 16:48:01.019957 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:01Z","lastTransitionTime":"2026-01-26T16:48:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:01 crc kubenswrapper[4754]: I0126 16:48:01.123358 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:01 crc kubenswrapper[4754]: I0126 16:48:01.123405 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:01 crc kubenswrapper[4754]: I0126 16:48:01.123415 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:01 crc kubenswrapper[4754]: I0126 16:48:01.123432 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:01 crc kubenswrapper[4754]: I0126 16:48:01.123443 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:01Z","lastTransitionTime":"2026-01-26T16:48:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:01 crc kubenswrapper[4754]: I0126 16:48:01.226070 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:01 crc kubenswrapper[4754]: I0126 16:48:01.226134 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:01 crc kubenswrapper[4754]: I0126 16:48:01.226143 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:01 crc kubenswrapper[4754]: I0126 16:48:01.226158 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:01 crc kubenswrapper[4754]: I0126 16:48:01.226169 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:01Z","lastTransitionTime":"2026-01-26T16:48:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:01 crc kubenswrapper[4754]: I0126 16:48:01.328789 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:01 crc kubenswrapper[4754]: I0126 16:48:01.328849 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:01 crc kubenswrapper[4754]: I0126 16:48:01.328861 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:01 crc kubenswrapper[4754]: I0126 16:48:01.328891 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:01 crc kubenswrapper[4754]: I0126 16:48:01.328905 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:01Z","lastTransitionTime":"2026-01-26T16:48:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:01 crc kubenswrapper[4754]: I0126 16:48:01.431351 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:01 crc kubenswrapper[4754]: I0126 16:48:01.431395 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:01 crc kubenswrapper[4754]: I0126 16:48:01.431407 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:01 crc kubenswrapper[4754]: I0126 16:48:01.431422 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:01 crc kubenswrapper[4754]: I0126 16:48:01.431433 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:01Z","lastTransitionTime":"2026-01-26T16:48:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:01 crc kubenswrapper[4754]: I0126 16:48:01.533848 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:01 crc kubenswrapper[4754]: I0126 16:48:01.533890 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:01 crc kubenswrapper[4754]: I0126 16:48:01.533906 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:01 crc kubenswrapper[4754]: I0126 16:48:01.533921 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:01 crc kubenswrapper[4754]: I0126 16:48:01.533932 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:01Z","lastTransitionTime":"2026-01-26T16:48:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:01 crc kubenswrapper[4754]: I0126 16:48:01.637630 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:01 crc kubenswrapper[4754]: I0126 16:48:01.637794 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:01 crc kubenswrapper[4754]: I0126 16:48:01.637810 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:01 crc kubenswrapper[4754]: I0126 16:48:01.637838 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:01 crc kubenswrapper[4754]: I0126 16:48:01.637856 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:01Z","lastTransitionTime":"2026-01-26T16:48:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:01 crc kubenswrapper[4754]: I0126 16:48:01.741417 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:01 crc kubenswrapper[4754]: I0126 16:48:01.741456 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:01 crc kubenswrapper[4754]: I0126 16:48:01.741466 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:01 crc kubenswrapper[4754]: I0126 16:48:01.741485 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:01 crc kubenswrapper[4754]: I0126 16:48:01.741495 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:01Z","lastTransitionTime":"2026-01-26T16:48:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:01 crc kubenswrapper[4754]: I0126 16:48:01.741717 4754 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-21 23:19:49.060194375 +0000 UTC Jan 26 16:48:01 crc kubenswrapper[4754]: I0126 16:48:01.766822 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 16:48:01 crc kubenswrapper[4754]: I0126 16:48:01.766870 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 16:48:01 crc kubenswrapper[4754]: I0126 16:48:01.766887 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-t4ps6" Jan 26 16:48:01 crc kubenswrapper[4754]: E0126 16:48:01.766979 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 16:48:01 crc kubenswrapper[4754]: E0126 16:48:01.767129 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 16:48:01 crc kubenswrapper[4754]: I0126 16:48:01.767203 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 16:48:01 crc kubenswrapper[4754]: E0126 16:48:01.767250 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-t4ps6" podUID="d45179f0-9f81-4d77-8805-6b05bdb32f8f" Jan 26 16:48:01 crc kubenswrapper[4754]: E0126 16:48:01.767383 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 16:48:01 crc kubenswrapper[4754]: I0126 16:48:01.845056 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:01 crc kubenswrapper[4754]: I0126 16:48:01.845125 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:01 crc kubenswrapper[4754]: I0126 16:48:01.845137 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:01 crc kubenswrapper[4754]: I0126 16:48:01.845159 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:01 crc kubenswrapper[4754]: I0126 16:48:01.845176 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:01Z","lastTransitionTime":"2026-01-26T16:48:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:01 crc kubenswrapper[4754]: I0126 16:48:01.948363 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:01 crc kubenswrapper[4754]: I0126 16:48:01.948419 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:01 crc kubenswrapper[4754]: I0126 16:48:01.948431 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:01 crc kubenswrapper[4754]: I0126 16:48:01.948447 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:01 crc kubenswrapper[4754]: I0126 16:48:01.948459 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:01Z","lastTransitionTime":"2026-01-26T16:48:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:02 crc kubenswrapper[4754]: I0126 16:48:02.050948 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:02 crc kubenswrapper[4754]: I0126 16:48:02.050996 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:02 crc kubenswrapper[4754]: I0126 16:48:02.051007 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:02 crc kubenswrapper[4754]: I0126 16:48:02.051023 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:02 crc kubenswrapper[4754]: I0126 16:48:02.051034 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:02Z","lastTransitionTime":"2026-01-26T16:48:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:02 crc kubenswrapper[4754]: I0126 16:48:02.153160 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:02 crc kubenswrapper[4754]: I0126 16:48:02.153212 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:02 crc kubenswrapper[4754]: I0126 16:48:02.153220 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:02 crc kubenswrapper[4754]: I0126 16:48:02.153233 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:02 crc kubenswrapper[4754]: I0126 16:48:02.153241 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:02Z","lastTransitionTime":"2026-01-26T16:48:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:02 crc kubenswrapper[4754]: I0126 16:48:02.255846 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:02 crc kubenswrapper[4754]: I0126 16:48:02.255906 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:02 crc kubenswrapper[4754]: I0126 16:48:02.255918 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:02 crc kubenswrapper[4754]: I0126 16:48:02.255934 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:02 crc kubenswrapper[4754]: I0126 16:48:02.255944 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:02Z","lastTransitionTime":"2026-01-26T16:48:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:02 crc kubenswrapper[4754]: I0126 16:48:02.358575 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:02 crc kubenswrapper[4754]: I0126 16:48:02.358623 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:02 crc kubenswrapper[4754]: I0126 16:48:02.358633 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:02 crc kubenswrapper[4754]: I0126 16:48:02.358650 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:02 crc kubenswrapper[4754]: I0126 16:48:02.358675 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:02Z","lastTransitionTime":"2026-01-26T16:48:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:02 crc kubenswrapper[4754]: I0126 16:48:02.461426 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:02 crc kubenswrapper[4754]: I0126 16:48:02.461468 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:02 crc kubenswrapper[4754]: I0126 16:48:02.461482 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:02 crc kubenswrapper[4754]: I0126 16:48:02.461497 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:02 crc kubenswrapper[4754]: I0126 16:48:02.461508 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:02Z","lastTransitionTime":"2026-01-26T16:48:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:02 crc kubenswrapper[4754]: I0126 16:48:02.564184 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:02 crc kubenswrapper[4754]: I0126 16:48:02.564225 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:02 crc kubenswrapper[4754]: I0126 16:48:02.564236 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:02 crc kubenswrapper[4754]: I0126 16:48:02.564251 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:02 crc kubenswrapper[4754]: I0126 16:48:02.564262 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:02Z","lastTransitionTime":"2026-01-26T16:48:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:02 crc kubenswrapper[4754]: I0126 16:48:02.666801 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:02 crc kubenswrapper[4754]: I0126 16:48:02.666865 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:02 crc kubenswrapper[4754]: I0126 16:48:02.666877 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:02 crc kubenswrapper[4754]: I0126 16:48:02.666896 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:02 crc kubenswrapper[4754]: I0126 16:48:02.666909 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:02Z","lastTransitionTime":"2026-01-26T16:48:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:02 crc kubenswrapper[4754]: I0126 16:48:02.742061 4754 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-27 08:48:51.683988928 +0000 UTC Jan 26 16:48:02 crc kubenswrapper[4754]: I0126 16:48:02.769147 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:02 crc kubenswrapper[4754]: I0126 16:48:02.769202 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:02 crc kubenswrapper[4754]: I0126 16:48:02.769213 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:02 crc kubenswrapper[4754]: I0126 16:48:02.769229 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:02 crc kubenswrapper[4754]: I0126 16:48:02.769241 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:02Z","lastTransitionTime":"2026-01-26T16:48:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:02 crc kubenswrapper[4754]: I0126 16:48:02.873997 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:02 crc kubenswrapper[4754]: I0126 16:48:02.874042 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:02 crc kubenswrapper[4754]: I0126 16:48:02.874052 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:02 crc kubenswrapper[4754]: I0126 16:48:02.874067 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:02 crc kubenswrapper[4754]: I0126 16:48:02.874078 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:02Z","lastTransitionTime":"2026-01-26T16:48:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:02 crc kubenswrapper[4754]: I0126 16:48:02.976644 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:02 crc kubenswrapper[4754]: I0126 16:48:02.976715 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:02 crc kubenswrapper[4754]: I0126 16:48:02.976724 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:02 crc kubenswrapper[4754]: I0126 16:48:02.976737 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:02 crc kubenswrapper[4754]: I0126 16:48:02.976746 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:02Z","lastTransitionTime":"2026-01-26T16:48:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:03 crc kubenswrapper[4754]: I0126 16:48:03.079704 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:03 crc kubenswrapper[4754]: I0126 16:48:03.079883 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:03 crc kubenswrapper[4754]: I0126 16:48:03.079918 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:03 crc kubenswrapper[4754]: I0126 16:48:03.079938 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:03 crc kubenswrapper[4754]: I0126 16:48:03.079949 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:03Z","lastTransitionTime":"2026-01-26T16:48:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:03 crc kubenswrapper[4754]: I0126 16:48:03.183427 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:03 crc kubenswrapper[4754]: I0126 16:48:03.183480 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:03 crc kubenswrapper[4754]: I0126 16:48:03.183493 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:03 crc kubenswrapper[4754]: I0126 16:48:03.183511 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:03 crc kubenswrapper[4754]: I0126 16:48:03.183521 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:03Z","lastTransitionTime":"2026-01-26T16:48:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:03 crc kubenswrapper[4754]: I0126 16:48:03.287068 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:03 crc kubenswrapper[4754]: I0126 16:48:03.287134 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:03 crc kubenswrapper[4754]: I0126 16:48:03.287146 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:03 crc kubenswrapper[4754]: I0126 16:48:03.287164 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:03 crc kubenswrapper[4754]: I0126 16:48:03.287176 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:03Z","lastTransitionTime":"2026-01-26T16:48:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:03 crc kubenswrapper[4754]: I0126 16:48:03.389062 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:03 crc kubenswrapper[4754]: I0126 16:48:03.389102 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:03 crc kubenswrapper[4754]: I0126 16:48:03.389116 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:03 crc kubenswrapper[4754]: I0126 16:48:03.389132 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:03 crc kubenswrapper[4754]: I0126 16:48:03.389142 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:03Z","lastTransitionTime":"2026-01-26T16:48:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:03 crc kubenswrapper[4754]: I0126 16:48:03.493255 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:03 crc kubenswrapper[4754]: I0126 16:48:03.493293 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:03 crc kubenswrapper[4754]: I0126 16:48:03.493302 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:03 crc kubenswrapper[4754]: I0126 16:48:03.493315 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:03 crc kubenswrapper[4754]: I0126 16:48:03.493328 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:03Z","lastTransitionTime":"2026-01-26T16:48:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:03 crc kubenswrapper[4754]: I0126 16:48:03.595697 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:03 crc kubenswrapper[4754]: I0126 16:48:03.595736 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:03 crc kubenswrapper[4754]: I0126 16:48:03.595747 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:03 crc kubenswrapper[4754]: I0126 16:48:03.595763 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:03 crc kubenswrapper[4754]: I0126 16:48:03.595773 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:03Z","lastTransitionTime":"2026-01-26T16:48:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:03 crc kubenswrapper[4754]: I0126 16:48:03.698480 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:03 crc kubenswrapper[4754]: I0126 16:48:03.698512 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:03 crc kubenswrapper[4754]: I0126 16:48:03.698523 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:03 crc kubenswrapper[4754]: I0126 16:48:03.698534 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:03 crc kubenswrapper[4754]: I0126 16:48:03.698542 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:03Z","lastTransitionTime":"2026-01-26T16:48:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:03 crc kubenswrapper[4754]: I0126 16:48:03.742625 4754 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-18 09:43:30.207976385 +0000 UTC Jan 26 16:48:03 crc kubenswrapper[4754]: I0126 16:48:03.750290 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 16:48:03 crc kubenswrapper[4754]: I0126 16:48:03.750476 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 16:48:03 crc kubenswrapper[4754]: E0126 16:48:03.750552 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 16:48:35.750515461 +0000 UTC m=+82.274695895 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 16:48:03 crc kubenswrapper[4754]: I0126 16:48:03.750625 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 16:48:03 crc kubenswrapper[4754]: E0126 16:48:03.750654 4754 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 26 16:48:03 crc kubenswrapper[4754]: I0126 16:48:03.750719 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 16:48:03 crc kubenswrapper[4754]: I0126 16:48:03.750751 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 16:48:03 crc kubenswrapper[4754]: E0126 16:48:03.750726 4754 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 26 16:48:03 crc kubenswrapper[4754]: E0126 16:48:03.750836 4754 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 26 16:48:03 crc kubenswrapper[4754]: E0126 16:48:03.750885 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-01-26 16:48:35.75087585 +0000 UTC m=+82.275056284 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 26 16:48:03 crc kubenswrapper[4754]: E0126 16:48:03.750749 4754 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 26 16:48:03 crc kubenswrapper[4754]: E0126 16:48:03.750952 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-26 16:48:35.750935202 +0000 UTC m=+82.275115636 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 26 16:48:03 crc kubenswrapper[4754]: E0126 16:48:03.750799 4754 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Jan 26 16:48:03 crc kubenswrapper[4754]: E0126 16:48:03.751001 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-26 16:48:35.750993204 +0000 UTC m=+82.275173738 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Jan 26 16:48:03 crc kubenswrapper[4754]: E0126 16:48:03.750796 4754 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 26 16:48:03 crc kubenswrapper[4754]: E0126 16:48:03.751021 4754 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 26 16:48:03 crc kubenswrapper[4754]: E0126 16:48:03.751032 4754 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 26 16:48:03 crc kubenswrapper[4754]: E0126 16:48:03.751058 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-01-26 16:48:35.751048815 +0000 UTC m=+82.275229339 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 26 16:48:03 crc kubenswrapper[4754]: I0126 16:48:03.767413 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-t4ps6" Jan 26 16:48:03 crc kubenswrapper[4754]: E0126 16:48:03.767596 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-t4ps6" podUID="d45179f0-9f81-4d77-8805-6b05bdb32f8f" Jan 26 16:48:03 crc kubenswrapper[4754]: I0126 16:48:03.767732 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 16:48:03 crc kubenswrapper[4754]: I0126 16:48:03.767782 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 16:48:03 crc kubenswrapper[4754]: E0126 16:48:03.767899 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 16:48:03 crc kubenswrapper[4754]: I0126 16:48:03.768129 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 16:48:03 crc kubenswrapper[4754]: E0126 16:48:03.768122 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 16:48:03 crc kubenswrapper[4754]: E0126 16:48:03.768228 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 16:48:03 crc kubenswrapper[4754]: I0126 16:48:03.784085 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2blzw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca65468e-30c6-4666-962e-cc0de05e67fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4603235bb2af495901926b805fe1c0ffc48824a4394ccf928e7fcd53a673ac3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c25z4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:33Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2blzw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:03Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:03 crc kubenswrapper[4754]: I0126 16:48:03.800849 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:03 crc kubenswrapper[4754]: I0126 16:48:03.801189 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:03 crc kubenswrapper[4754]: I0126 16:48:03.801266 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:03 crc kubenswrapper[4754]: I0126 16:48:03.801346 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:03 crc kubenswrapper[4754]: I0126 16:48:03.801419 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:03Z","lastTransitionTime":"2026-01-26T16:48:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:03 crc kubenswrapper[4754]: I0126 16:48:03.804457 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bwpd6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"351cc6c0-b8e5-4589-ae63-f98e8ab1eca3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1022f101d36873b41bbc1ed8a4a2590347b49f7cce6a172aa1d5bc9f529aed65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3ce0fe080a094316b9456057e347b90a5ff432daa12b9c56d7aea9bee1f72b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3ce0fe080a094316b9456057e347b90a5ff432daa12b9c56d7aea9bee1f72b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1772bb41eb8a36a863d3c691610bcbee5d16b260480b3e9f9d6b083e09e42799\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1772bb41eb8a36a863d3c691610bcbee5d16b260480b3e9f9d6b083e09e42799\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24848d69756024cc664affb9492764ace0638ef611f910f4f2bf86b638cc443d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://24848d69756024cc664affb9492764ace0638ef611f910f4f2bf86b638cc443d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c264fc7c3300a725566acce4464f01d9314779737624401c164d867831fd93bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c264fc7c3300a725566acce4464f01d9314779737624401c164d867831fd93bf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5acfe327eee4aad927ac3a32d24549eff788b14a1c4bdac359b6b56301d884ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5acfe327eee4aad927ac3a32d24549eff788b14a1c4bdac359b6b56301d884ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://63a9b9317494d0e0b33970ef3d461729b10d2a1ad7d3d2bd33f389b184ca56ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://63a9b9317494d0e0b33970ef3d461729b10d2a1ad7d3d2bd33f389b184ca56ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bwpd6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:03Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:03 crc kubenswrapper[4754]: I0126 16:48:03.835572 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jsbxt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"689569c4-93ee-4b82-93fc-9af0f97e6dc3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://033f8b104d4e9ff6d5f87ca998665f12e0c56696b3d30e01ede05f4da540365d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://919572e731813a98c7d8a94879bba931f75c472c7cd681e86d5197ebab43815d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1ac1f49bd564750fcff530a0f76657b0ed41573539acd020a922d19861ef834\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a300a80ad87f46ba1b4fd7ad3cf6c25767fcfe966b93bdd86d729843f53a5cda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31dcab2f9a2d218b0e49dad1df46377ff4521592824a0fbac254d5b8e6d0632e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fe4d9395447ac570eca1338e749dbefea76295d427a12487b63a142a7d6a2e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e9d45c28e67d857b1c50ed78197ed052c1ac2b98f1c8a94addd63698bf0bb23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0e9d45c28e67d857b1c50ed78197ed052c1ac2b98f1c8a94addd63698bf0bb23\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-26T16:47:58Z\\\",\\\"message\\\":\\\":303] Retry object setup: *v1.Pod openshift-machine-config-operator/machine-config-daemon-x65wv\\\\nI0126 16:47:58.625627 6452 obj_retry.go:303] Retry object setup: *v1.Pod openshift-multus/multus-skh5n\\\\nI0126 16:47:58.625718 6452 obj_retry.go:365] Adding new object: *v1.Pod openshift-machine-config-operator/machine-config-daemon-x65wv\\\\nI0126 16:47:58.625617 6452 obj_retry.go:303] Retry object setup: *v1.Pod openshift-multus/multus-additional-cni-plugins-bwpd6\\\\nI0126 16:47:58.625726 6452 ovn.go:134] Ensuring zone local for Pod openshift-machine-config-operator/machine-config-daemon-x65wv in node crc\\\\nI0126 16:47:58.625727 6452 obj_retry.go:365] Adding new object: *v1.Pod openshift-multus/multus-skh5n\\\\nI0126 16:47:58.625732 6452 obj_retry.go:365] Adding new object: *v1.Pod openshift-multus/multus-additional-cni-plugins-bwpd6\\\\nI0126 16:47:58.625706 6452 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {eb8eef51-1a8d-43f9-ae2e-3b2cc00ded60}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0126 16:47:58.625740 6452 ovn.go:134] Ensuring zone local for Pod openshift-multus/multus-additional-cni-plugins-bwpd6 in no\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:57Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-jsbxt_openshift-ovn-kubernetes(689569c4-93ee-4b82-93fc-9af0f97e6dc3)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://07a43ff7860ff65faf51f4a5e8aeb326f06d6d7222677a92560ae2e3a0c81484\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d9a7bafabd44f8b88a6f62cbe45599af8563bc45d9d7ad3a7669f3d8d11eb9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d9a7bafabd44f8b88a6f62cbe45599af8563bc45d9d7ad3a7669f3d8d11eb9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jsbxt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:03Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:03 crc kubenswrapper[4754]: I0126 16:48:03.852871 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:03Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:03 crc kubenswrapper[4754]: I0126 16:48:03.864191 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e459fd8c1093d6f3d0a671751b792b26b7df94c21ced1be1a2250821bcc82d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:03Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:03 crc kubenswrapper[4754]: I0126 16:48:03.877063 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89251fa9c64bc25d6d6f355400d7644c099f22aed1cae812fc991a03595f1e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e90508d8a53be9146f9b388750730fdae9ee200306caf5ad393a356390a69ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:03Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:03 crc kubenswrapper[4754]: I0126 16:48:03.888329 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-84lgq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6763d962-0892-4acc-bab9-7a4e93e39a85\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://953c1917fca896c3def3fc2c69f9e3dbdc9ca26fc2b3d40282be71f71759fd68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8zlw6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:33Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-84lgq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:03Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:03 crc kubenswrapper[4754]: I0126 16:48:03.899833 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-skh5n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b619d233-b592-4b05-a0b7-dc094c88471e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9231a4191565e2e0ef53d8a394bb34615c900b601a0dfec24c2d8e5fa7b0b659\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-skh5n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:03Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:03 crc kubenswrapper[4754]: I0126 16:48:03.903022 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:03 crc kubenswrapper[4754]: I0126 16:48:03.903056 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:03 crc kubenswrapper[4754]: I0126 16:48:03.903065 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:03 crc kubenswrapper[4754]: I0126 16:48:03.903080 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:03 crc kubenswrapper[4754]: I0126 16:48:03.903089 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:03Z","lastTransitionTime":"2026-01-26T16:48:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:03 crc kubenswrapper[4754]: I0126 16:48:03.910012 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jspvh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5c81aab0-939e-45ce-b19a-89d4818980c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://356dc54150a495dd6f35e0667c82b807b712b07200d8a0adbb1c92fe3f7af8f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8bps\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://05fd57780c44cd127859dc8b163d20eafbd8b06f3348cf545ef41021ce6bc526\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8bps\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-jspvh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:03Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:03 crc kubenswrapper[4754]: I0126 16:48:03.922312 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"099af702-abf2-4933-b4c9-63ef2326e468\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23017e73f8450839b4c2c02fa61235ffe4604bdce11556def27f25bcadcf89dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a74507c941b9de4d0401c8f1d53bb79ed11359628d3405cdab7f5474a75c862e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4402182b3e2e35a7d5dd4c44ac53e152915b312b7ccb076c4f2b41b925a6e7f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f772a003566198519fe1849fd14f29d4346004b2e655810662507c95b4d5ffa1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:13Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:03Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:03 crc kubenswrapper[4754]: I0126 16:48:03.933189 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:03Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:03 crc kubenswrapper[4754]: I0126 16:48:03.944291 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c3718a4-f354-4284-92e0-fdfb45a692bd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b86f46f49163bca7eca9458279029af330dbd4890b03325e06fa3b1bdf030de6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxjcz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fffc03b536bfee3c7bc1bddd500e834e34b391896477a1501e3e06201a19374\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxjcz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-x65wv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:03Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:03 crc kubenswrapper[4754]: I0126 16:48:03.953521 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-t4ps6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d45179f0-9f81-4d77-8805-6b05bdb32f8f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vk8tg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vk8tg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:48Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-t4ps6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:03Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:03 crc kubenswrapper[4754]: I0126 16:48:03.965802 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b89aa081-0a2e-4757-9210-e2a6c228bee9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9c5f40188fda14b14128514fe84bddfe8dcede3d6c560283ccb2ac0c66895fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f45445cbf1d180d6d4dd9588e85bf678f043bfce8cff0812b78ca3510904f9a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://085fa7b8d4fef8b2722fac8c346107699b5a6bd33fbfc74ed35831cb8fdf9dad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c40e018a0979035827963efb9efc48a878a55b9df68ec1a9242d7cff64fe480\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8356f76fbc3a4e636c12d1da5455d3f6767d9b83ce2f2925a11435d7d7261e9f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0126 16:47:26.205208 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0126 16:47:26.206538 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3645934156/tls.crt::/tmp/serving-cert-3645934156/tls.key\\\\\\\"\\\\nI0126 16:47:31.504518 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0126 16:47:31.509366 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0126 16:47:31.509400 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0126 16:47:31.509428 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0126 16:47:31.509436 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0126 16:47:31.515912 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0126 16:47:31.516036 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 16:47:31.516068 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 16:47:31.516096 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0126 16:47:31.516121 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0126 16:47:31.516145 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0126 16:47:31.516170 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0126 16:47:31.516412 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0126 16:47:31.519294 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27b74e2e65a547e0e28fddd4f15feacb17f4d57596eeb8da82df9d39243c85e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22ba7649a1308815bfad52016000987a7a9a5bc214d7ff61e137ba099b20fc6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://22ba7649a1308815bfad52016000987a7a9a5bc214d7ff61e137ba099b20fc6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:13Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:03Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:03 crc kubenswrapper[4754]: I0126 16:48:03.980013 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://205ec256e46a8e08b4b0b0805fb6c5f4966fc3ccef0738df4b97cb404e74da17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:03Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:03 crc kubenswrapper[4754]: I0126 16:48:03.992247 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:03Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:04 crc kubenswrapper[4754]: I0126 16:48:04.005548 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:04 crc kubenswrapper[4754]: I0126 16:48:04.005581 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:04 crc kubenswrapper[4754]: I0126 16:48:04.005589 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:04 crc kubenswrapper[4754]: I0126 16:48:04.005604 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:04 crc kubenswrapper[4754]: I0126 16:48:04.005614 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:04Z","lastTransitionTime":"2026-01-26T16:48:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:04 crc kubenswrapper[4754]: I0126 16:48:04.107729 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:04 crc kubenswrapper[4754]: I0126 16:48:04.108031 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:04 crc kubenswrapper[4754]: I0126 16:48:04.108041 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:04 crc kubenswrapper[4754]: I0126 16:48:04.108058 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:04 crc kubenswrapper[4754]: I0126 16:48:04.108070 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:04Z","lastTransitionTime":"2026-01-26T16:48:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:04 crc kubenswrapper[4754]: I0126 16:48:04.210472 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:04 crc kubenswrapper[4754]: I0126 16:48:04.210552 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:04 crc kubenswrapper[4754]: I0126 16:48:04.210569 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:04 crc kubenswrapper[4754]: I0126 16:48:04.210586 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:04 crc kubenswrapper[4754]: I0126 16:48:04.210629 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:04Z","lastTransitionTime":"2026-01-26T16:48:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:04 crc kubenswrapper[4754]: I0126 16:48:04.255229 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d45179f0-9f81-4d77-8805-6b05bdb32f8f-metrics-certs\") pod \"network-metrics-daemon-t4ps6\" (UID: \"d45179f0-9f81-4d77-8805-6b05bdb32f8f\") " pod="openshift-multus/network-metrics-daemon-t4ps6" Jan 26 16:48:04 crc kubenswrapper[4754]: E0126 16:48:04.255373 4754 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Jan 26 16:48:04 crc kubenswrapper[4754]: E0126 16:48:04.255427 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d45179f0-9f81-4d77-8805-6b05bdb32f8f-metrics-certs podName:d45179f0-9f81-4d77-8805-6b05bdb32f8f nodeName:}" failed. No retries permitted until 2026-01-26 16:48:20.255412125 +0000 UTC m=+66.779592559 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/d45179f0-9f81-4d77-8805-6b05bdb32f8f-metrics-certs") pod "network-metrics-daemon-t4ps6" (UID: "d45179f0-9f81-4d77-8805-6b05bdb32f8f") : object "openshift-multus"/"metrics-daemon-secret" not registered Jan 26 16:48:04 crc kubenswrapper[4754]: I0126 16:48:04.312989 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:04 crc kubenswrapper[4754]: I0126 16:48:04.313022 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:04 crc kubenswrapper[4754]: I0126 16:48:04.313032 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:04 crc kubenswrapper[4754]: I0126 16:48:04.313045 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:04 crc kubenswrapper[4754]: I0126 16:48:04.313054 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:04Z","lastTransitionTime":"2026-01-26T16:48:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:04 crc kubenswrapper[4754]: I0126 16:48:04.415993 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:04 crc kubenswrapper[4754]: I0126 16:48:04.416030 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:04 crc kubenswrapper[4754]: I0126 16:48:04.416039 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:04 crc kubenswrapper[4754]: I0126 16:48:04.416052 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:04 crc kubenswrapper[4754]: I0126 16:48:04.416062 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:04Z","lastTransitionTime":"2026-01-26T16:48:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:04 crc kubenswrapper[4754]: I0126 16:48:04.519261 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:04 crc kubenswrapper[4754]: I0126 16:48:04.519378 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:04 crc kubenswrapper[4754]: I0126 16:48:04.519396 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:04 crc kubenswrapper[4754]: I0126 16:48:04.519426 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:04 crc kubenswrapper[4754]: I0126 16:48:04.519445 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:04Z","lastTransitionTime":"2026-01-26T16:48:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:04 crc kubenswrapper[4754]: I0126 16:48:04.622970 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:04 crc kubenswrapper[4754]: I0126 16:48:04.623040 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:04 crc kubenswrapper[4754]: I0126 16:48:04.623075 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:04 crc kubenswrapper[4754]: I0126 16:48:04.623096 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:04 crc kubenswrapper[4754]: I0126 16:48:04.623107 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:04Z","lastTransitionTime":"2026-01-26T16:48:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:04 crc kubenswrapper[4754]: I0126 16:48:04.724931 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:04 crc kubenswrapper[4754]: I0126 16:48:04.724972 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:04 crc kubenswrapper[4754]: I0126 16:48:04.724983 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:04 crc kubenswrapper[4754]: I0126 16:48:04.724999 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:04 crc kubenswrapper[4754]: I0126 16:48:04.725009 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:04Z","lastTransitionTime":"2026-01-26T16:48:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:04 crc kubenswrapper[4754]: I0126 16:48:04.743727 4754 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-01 22:06:19.127591089 +0000 UTC Jan 26 16:48:04 crc kubenswrapper[4754]: I0126 16:48:04.826880 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:04 crc kubenswrapper[4754]: I0126 16:48:04.826943 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:04 crc kubenswrapper[4754]: I0126 16:48:04.826966 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:04 crc kubenswrapper[4754]: I0126 16:48:04.826999 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:04 crc kubenswrapper[4754]: I0126 16:48:04.827022 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:04Z","lastTransitionTime":"2026-01-26T16:48:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:04 crc kubenswrapper[4754]: I0126 16:48:04.929609 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:04 crc kubenswrapper[4754]: I0126 16:48:04.929683 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:04 crc kubenswrapper[4754]: I0126 16:48:04.929696 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:04 crc kubenswrapper[4754]: I0126 16:48:04.929715 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:04 crc kubenswrapper[4754]: I0126 16:48:04.929729 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:04Z","lastTransitionTime":"2026-01-26T16:48:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:05 crc kubenswrapper[4754]: I0126 16:48:05.033144 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:05 crc kubenswrapper[4754]: I0126 16:48:05.033183 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:05 crc kubenswrapper[4754]: I0126 16:48:05.033192 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:05 crc kubenswrapper[4754]: I0126 16:48:05.033207 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:05 crc kubenswrapper[4754]: I0126 16:48:05.033219 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:05Z","lastTransitionTime":"2026-01-26T16:48:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:05 crc kubenswrapper[4754]: I0126 16:48:05.135339 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:05 crc kubenswrapper[4754]: I0126 16:48:05.135404 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:05 crc kubenswrapper[4754]: I0126 16:48:05.135420 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:05 crc kubenswrapper[4754]: I0126 16:48:05.135440 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:05 crc kubenswrapper[4754]: I0126 16:48:05.135454 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:05Z","lastTransitionTime":"2026-01-26T16:48:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:05 crc kubenswrapper[4754]: I0126 16:48:05.158522 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:05 crc kubenswrapper[4754]: I0126 16:48:05.158572 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:05 crc kubenswrapper[4754]: I0126 16:48:05.158581 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:05 crc kubenswrapper[4754]: I0126 16:48:05.158595 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:05 crc kubenswrapper[4754]: I0126 16:48:05.158606 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:05Z","lastTransitionTime":"2026-01-26T16:48:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:05 crc kubenswrapper[4754]: E0126 16:48:05.171471 4754 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:48:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:48:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:48:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:48:05Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:48:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:48:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:48:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:48:05Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d8506764-e7fa-45cb-a13d-6f527164f548\\\",\\\"systemUUID\\\":\\\"bfbb6c33-ce46-4e81-a8a8-b44409b03821\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:05Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:05 crc kubenswrapper[4754]: I0126 16:48:05.175481 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:05 crc kubenswrapper[4754]: I0126 16:48:05.175519 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:05 crc kubenswrapper[4754]: I0126 16:48:05.175530 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:05 crc kubenswrapper[4754]: I0126 16:48:05.175545 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:05 crc kubenswrapper[4754]: I0126 16:48:05.175556 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:05Z","lastTransitionTime":"2026-01-26T16:48:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:05 crc kubenswrapper[4754]: E0126 16:48:05.188657 4754 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:48:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:48:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:48:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:48:05Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:48:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:48:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:48:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:48:05Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d8506764-e7fa-45cb-a13d-6f527164f548\\\",\\\"systemUUID\\\":\\\"bfbb6c33-ce46-4e81-a8a8-b44409b03821\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:05Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:05 crc kubenswrapper[4754]: I0126 16:48:05.192902 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:05 crc kubenswrapper[4754]: I0126 16:48:05.192937 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:05 crc kubenswrapper[4754]: I0126 16:48:05.192947 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:05 crc kubenswrapper[4754]: I0126 16:48:05.192963 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:05 crc kubenswrapper[4754]: I0126 16:48:05.192972 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:05Z","lastTransitionTime":"2026-01-26T16:48:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:05 crc kubenswrapper[4754]: E0126 16:48:05.205290 4754 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:48:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:48:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:48:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:48:05Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:48:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:48:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:48:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:48:05Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d8506764-e7fa-45cb-a13d-6f527164f548\\\",\\\"systemUUID\\\":\\\"bfbb6c33-ce46-4e81-a8a8-b44409b03821\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:05Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:05 crc kubenswrapper[4754]: I0126 16:48:05.208880 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:05 crc kubenswrapper[4754]: I0126 16:48:05.208918 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:05 crc kubenswrapper[4754]: I0126 16:48:05.208929 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:05 crc kubenswrapper[4754]: I0126 16:48:05.208944 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:05 crc kubenswrapper[4754]: I0126 16:48:05.208955 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:05Z","lastTransitionTime":"2026-01-26T16:48:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:05 crc kubenswrapper[4754]: E0126 16:48:05.222191 4754 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:48:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:48:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:48:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:48:05Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:48:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:48:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:48:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:48:05Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d8506764-e7fa-45cb-a13d-6f527164f548\\\",\\\"systemUUID\\\":\\\"bfbb6c33-ce46-4e81-a8a8-b44409b03821\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:05Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:05 crc kubenswrapper[4754]: I0126 16:48:05.226973 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:05 crc kubenswrapper[4754]: I0126 16:48:05.227022 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:05 crc kubenswrapper[4754]: I0126 16:48:05.227037 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:05 crc kubenswrapper[4754]: I0126 16:48:05.227059 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:05 crc kubenswrapper[4754]: I0126 16:48:05.227074 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:05Z","lastTransitionTime":"2026-01-26T16:48:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:05 crc kubenswrapper[4754]: E0126 16:48:05.240378 4754 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:48:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:48:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:48:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:48:05Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:48:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:48:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:48:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:48:05Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d8506764-e7fa-45cb-a13d-6f527164f548\\\",\\\"systemUUID\\\":\\\"bfbb6c33-ce46-4e81-a8a8-b44409b03821\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:05Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:05 crc kubenswrapper[4754]: E0126 16:48:05.240550 4754 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Jan 26 16:48:05 crc kubenswrapper[4754]: I0126 16:48:05.242067 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:05 crc kubenswrapper[4754]: I0126 16:48:05.242095 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:05 crc kubenswrapper[4754]: I0126 16:48:05.242106 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:05 crc kubenswrapper[4754]: I0126 16:48:05.242121 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:05 crc kubenswrapper[4754]: I0126 16:48:05.242134 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:05Z","lastTransitionTime":"2026-01-26T16:48:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:05 crc kubenswrapper[4754]: I0126 16:48:05.344943 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:05 crc kubenswrapper[4754]: I0126 16:48:05.345034 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:05 crc kubenswrapper[4754]: I0126 16:48:05.345048 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:05 crc kubenswrapper[4754]: I0126 16:48:05.345063 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:05 crc kubenswrapper[4754]: I0126 16:48:05.345074 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:05Z","lastTransitionTime":"2026-01-26T16:48:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:05 crc kubenswrapper[4754]: I0126 16:48:05.448450 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:05 crc kubenswrapper[4754]: I0126 16:48:05.448499 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:05 crc kubenswrapper[4754]: I0126 16:48:05.448551 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:05 crc kubenswrapper[4754]: I0126 16:48:05.448573 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:05 crc kubenswrapper[4754]: I0126 16:48:05.448590 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:05Z","lastTransitionTime":"2026-01-26T16:48:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:05 crc kubenswrapper[4754]: I0126 16:48:05.552005 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:05 crc kubenswrapper[4754]: I0126 16:48:05.552063 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:05 crc kubenswrapper[4754]: I0126 16:48:05.552076 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:05 crc kubenswrapper[4754]: I0126 16:48:05.552091 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:05 crc kubenswrapper[4754]: I0126 16:48:05.552103 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:05Z","lastTransitionTime":"2026-01-26T16:48:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:05 crc kubenswrapper[4754]: I0126 16:48:05.654398 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:05 crc kubenswrapper[4754]: I0126 16:48:05.654436 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:05 crc kubenswrapper[4754]: I0126 16:48:05.654447 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:05 crc kubenswrapper[4754]: I0126 16:48:05.654462 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:05 crc kubenswrapper[4754]: I0126 16:48:05.654472 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:05Z","lastTransitionTime":"2026-01-26T16:48:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:05 crc kubenswrapper[4754]: I0126 16:48:05.744321 4754 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-09 12:51:19.968876734 +0000 UTC Jan 26 16:48:05 crc kubenswrapper[4754]: I0126 16:48:05.756236 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:05 crc kubenswrapper[4754]: I0126 16:48:05.756289 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:05 crc kubenswrapper[4754]: I0126 16:48:05.756300 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:05 crc kubenswrapper[4754]: I0126 16:48:05.756315 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:05 crc kubenswrapper[4754]: I0126 16:48:05.756326 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:05Z","lastTransitionTime":"2026-01-26T16:48:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:05 crc kubenswrapper[4754]: I0126 16:48:05.766911 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 16:48:05 crc kubenswrapper[4754]: I0126 16:48:05.766942 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-t4ps6" Jan 26 16:48:05 crc kubenswrapper[4754]: I0126 16:48:05.766911 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 16:48:05 crc kubenswrapper[4754]: E0126 16:48:05.767045 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 16:48:05 crc kubenswrapper[4754]: I0126 16:48:05.767164 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 16:48:05 crc kubenswrapper[4754]: E0126 16:48:05.767160 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-t4ps6" podUID="d45179f0-9f81-4d77-8805-6b05bdb32f8f" Jan 26 16:48:05 crc kubenswrapper[4754]: E0126 16:48:05.767236 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 16:48:05 crc kubenswrapper[4754]: E0126 16:48:05.767319 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 16:48:05 crc kubenswrapper[4754]: I0126 16:48:05.859343 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:05 crc kubenswrapper[4754]: I0126 16:48:05.859386 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:05 crc kubenswrapper[4754]: I0126 16:48:05.859397 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:05 crc kubenswrapper[4754]: I0126 16:48:05.859414 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:05 crc kubenswrapper[4754]: I0126 16:48:05.859428 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:05Z","lastTransitionTime":"2026-01-26T16:48:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:05 crc kubenswrapper[4754]: I0126 16:48:05.962091 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:05 crc kubenswrapper[4754]: I0126 16:48:05.962140 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:05 crc kubenswrapper[4754]: I0126 16:48:05.962152 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:05 crc kubenswrapper[4754]: I0126 16:48:05.962171 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:05 crc kubenswrapper[4754]: I0126 16:48:05.962182 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:05Z","lastTransitionTime":"2026-01-26T16:48:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:06 crc kubenswrapper[4754]: I0126 16:48:06.065124 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:06 crc kubenswrapper[4754]: I0126 16:48:06.065150 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:06 crc kubenswrapper[4754]: I0126 16:48:06.065160 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:06 crc kubenswrapper[4754]: I0126 16:48:06.065175 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:06 crc kubenswrapper[4754]: I0126 16:48:06.065186 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:06Z","lastTransitionTime":"2026-01-26T16:48:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:06 crc kubenswrapper[4754]: I0126 16:48:06.168934 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:06 crc kubenswrapper[4754]: I0126 16:48:06.168966 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:06 crc kubenswrapper[4754]: I0126 16:48:06.168975 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:06 crc kubenswrapper[4754]: I0126 16:48:06.168988 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:06 crc kubenswrapper[4754]: I0126 16:48:06.168997 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:06Z","lastTransitionTime":"2026-01-26T16:48:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:06 crc kubenswrapper[4754]: I0126 16:48:06.271583 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:06 crc kubenswrapper[4754]: I0126 16:48:06.271621 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:06 crc kubenswrapper[4754]: I0126 16:48:06.271630 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:06 crc kubenswrapper[4754]: I0126 16:48:06.271647 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:06 crc kubenswrapper[4754]: I0126 16:48:06.271658 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:06Z","lastTransitionTime":"2026-01-26T16:48:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:06 crc kubenswrapper[4754]: I0126 16:48:06.374608 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:06 crc kubenswrapper[4754]: I0126 16:48:06.374657 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:06 crc kubenswrapper[4754]: I0126 16:48:06.374686 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:06 crc kubenswrapper[4754]: I0126 16:48:06.374703 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:06 crc kubenswrapper[4754]: I0126 16:48:06.374715 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:06Z","lastTransitionTime":"2026-01-26T16:48:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:06 crc kubenswrapper[4754]: I0126 16:48:06.477436 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:06 crc kubenswrapper[4754]: I0126 16:48:06.477524 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:06 crc kubenswrapper[4754]: I0126 16:48:06.477540 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:06 crc kubenswrapper[4754]: I0126 16:48:06.477563 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:06 crc kubenswrapper[4754]: I0126 16:48:06.477579 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:06Z","lastTransitionTime":"2026-01-26T16:48:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:06 crc kubenswrapper[4754]: I0126 16:48:06.579808 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:06 crc kubenswrapper[4754]: I0126 16:48:06.579858 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:06 crc kubenswrapper[4754]: I0126 16:48:06.579871 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:06 crc kubenswrapper[4754]: I0126 16:48:06.579889 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:06 crc kubenswrapper[4754]: I0126 16:48:06.579902 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:06Z","lastTransitionTime":"2026-01-26T16:48:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:06 crc kubenswrapper[4754]: I0126 16:48:06.682956 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:06 crc kubenswrapper[4754]: I0126 16:48:06.683013 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:06 crc kubenswrapper[4754]: I0126 16:48:06.683021 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:06 crc kubenswrapper[4754]: I0126 16:48:06.683037 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:06 crc kubenswrapper[4754]: I0126 16:48:06.683048 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:06Z","lastTransitionTime":"2026-01-26T16:48:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:06 crc kubenswrapper[4754]: I0126 16:48:06.744520 4754 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-15 18:17:03.311693668 +0000 UTC Jan 26 16:48:06 crc kubenswrapper[4754]: I0126 16:48:06.786448 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:06 crc kubenswrapper[4754]: I0126 16:48:06.786501 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:06 crc kubenswrapper[4754]: I0126 16:48:06.786510 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:06 crc kubenswrapper[4754]: I0126 16:48:06.786525 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:06 crc kubenswrapper[4754]: I0126 16:48:06.786534 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:06Z","lastTransitionTime":"2026-01-26T16:48:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:06 crc kubenswrapper[4754]: I0126 16:48:06.889089 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:06 crc kubenswrapper[4754]: I0126 16:48:06.889140 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:06 crc kubenswrapper[4754]: I0126 16:48:06.889150 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:06 crc kubenswrapper[4754]: I0126 16:48:06.889166 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:06 crc kubenswrapper[4754]: I0126 16:48:06.889176 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:06Z","lastTransitionTime":"2026-01-26T16:48:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:06 crc kubenswrapper[4754]: I0126 16:48:06.991938 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:06 crc kubenswrapper[4754]: I0126 16:48:06.991994 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:06 crc kubenswrapper[4754]: I0126 16:48:06.992008 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:06 crc kubenswrapper[4754]: I0126 16:48:06.992028 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:06 crc kubenswrapper[4754]: I0126 16:48:06.992043 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:06Z","lastTransitionTime":"2026-01-26T16:48:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:07 crc kubenswrapper[4754]: I0126 16:48:07.094407 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:07 crc kubenswrapper[4754]: I0126 16:48:07.094456 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:07 crc kubenswrapper[4754]: I0126 16:48:07.094466 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:07 crc kubenswrapper[4754]: I0126 16:48:07.094492 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:07 crc kubenswrapper[4754]: I0126 16:48:07.094503 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:07Z","lastTransitionTime":"2026-01-26T16:48:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:07 crc kubenswrapper[4754]: I0126 16:48:07.196809 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:07 crc kubenswrapper[4754]: I0126 16:48:07.196854 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:07 crc kubenswrapper[4754]: I0126 16:48:07.196869 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:07 crc kubenswrapper[4754]: I0126 16:48:07.196890 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:07 crc kubenswrapper[4754]: I0126 16:48:07.196906 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:07Z","lastTransitionTime":"2026-01-26T16:48:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:07 crc kubenswrapper[4754]: I0126 16:48:07.299019 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:07 crc kubenswrapper[4754]: I0126 16:48:07.299374 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:07 crc kubenswrapper[4754]: I0126 16:48:07.299476 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:07 crc kubenswrapper[4754]: I0126 16:48:07.299563 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:07 crc kubenswrapper[4754]: I0126 16:48:07.299656 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:07Z","lastTransitionTime":"2026-01-26T16:48:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:07 crc kubenswrapper[4754]: I0126 16:48:07.401933 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:07 crc kubenswrapper[4754]: I0126 16:48:07.402181 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:07 crc kubenswrapper[4754]: I0126 16:48:07.402304 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:07 crc kubenswrapper[4754]: I0126 16:48:07.402387 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:07 crc kubenswrapper[4754]: I0126 16:48:07.402445 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:07Z","lastTransitionTime":"2026-01-26T16:48:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:07 crc kubenswrapper[4754]: I0126 16:48:07.504720 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:07 crc kubenswrapper[4754]: I0126 16:48:07.504760 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:07 crc kubenswrapper[4754]: I0126 16:48:07.504771 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:07 crc kubenswrapper[4754]: I0126 16:48:07.504786 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:07 crc kubenswrapper[4754]: I0126 16:48:07.504797 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:07Z","lastTransitionTime":"2026-01-26T16:48:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:07 crc kubenswrapper[4754]: I0126 16:48:07.607491 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:07 crc kubenswrapper[4754]: I0126 16:48:07.607526 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:07 crc kubenswrapper[4754]: I0126 16:48:07.607536 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:07 crc kubenswrapper[4754]: I0126 16:48:07.607550 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:07 crc kubenswrapper[4754]: I0126 16:48:07.607559 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:07Z","lastTransitionTime":"2026-01-26T16:48:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:07 crc kubenswrapper[4754]: I0126 16:48:07.710820 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:07 crc kubenswrapper[4754]: I0126 16:48:07.710867 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:07 crc kubenswrapper[4754]: I0126 16:48:07.710879 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:07 crc kubenswrapper[4754]: I0126 16:48:07.710897 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:07 crc kubenswrapper[4754]: I0126 16:48:07.710910 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:07Z","lastTransitionTime":"2026-01-26T16:48:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:07 crc kubenswrapper[4754]: I0126 16:48:07.745660 4754 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-01 10:43:50.386070479 +0000 UTC Jan 26 16:48:07 crc kubenswrapper[4754]: I0126 16:48:07.767143 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 16:48:07 crc kubenswrapper[4754]: I0126 16:48:07.767202 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 16:48:07 crc kubenswrapper[4754]: E0126 16:48:07.767265 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 16:48:07 crc kubenswrapper[4754]: I0126 16:48:07.767309 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 16:48:07 crc kubenswrapper[4754]: I0126 16:48:07.767312 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-t4ps6" Jan 26 16:48:07 crc kubenswrapper[4754]: E0126 16:48:07.767353 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 16:48:07 crc kubenswrapper[4754]: E0126 16:48:07.767468 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 16:48:07 crc kubenswrapper[4754]: E0126 16:48:07.767630 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-t4ps6" podUID="d45179f0-9f81-4d77-8805-6b05bdb32f8f" Jan 26 16:48:07 crc kubenswrapper[4754]: I0126 16:48:07.813875 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:07 crc kubenswrapper[4754]: I0126 16:48:07.813921 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:07 crc kubenswrapper[4754]: I0126 16:48:07.813930 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:07 crc kubenswrapper[4754]: I0126 16:48:07.813949 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:07 crc kubenswrapper[4754]: I0126 16:48:07.813961 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:07Z","lastTransitionTime":"2026-01-26T16:48:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:07 crc kubenswrapper[4754]: I0126 16:48:07.916366 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:07 crc kubenswrapper[4754]: I0126 16:48:07.916412 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:07 crc kubenswrapper[4754]: I0126 16:48:07.916421 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:07 crc kubenswrapper[4754]: I0126 16:48:07.916438 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:07 crc kubenswrapper[4754]: I0126 16:48:07.916449 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:07Z","lastTransitionTime":"2026-01-26T16:48:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:08 crc kubenswrapper[4754]: I0126 16:48:08.019057 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:08 crc kubenswrapper[4754]: I0126 16:48:08.019091 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:08 crc kubenswrapper[4754]: I0126 16:48:08.019099 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:08 crc kubenswrapper[4754]: I0126 16:48:08.019111 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:08 crc kubenswrapper[4754]: I0126 16:48:08.019120 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:08Z","lastTransitionTime":"2026-01-26T16:48:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:08 crc kubenswrapper[4754]: I0126 16:48:08.121433 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:08 crc kubenswrapper[4754]: I0126 16:48:08.121482 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:08 crc kubenswrapper[4754]: I0126 16:48:08.121496 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:08 crc kubenswrapper[4754]: I0126 16:48:08.121512 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:08 crc kubenswrapper[4754]: I0126 16:48:08.121525 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:08Z","lastTransitionTime":"2026-01-26T16:48:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:08 crc kubenswrapper[4754]: I0126 16:48:08.214935 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Jan 26 16:48:08 crc kubenswrapper[4754]: I0126 16:48:08.223923 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:08 crc kubenswrapper[4754]: I0126 16:48:08.223969 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:08 crc kubenswrapper[4754]: I0126 16:48:08.223982 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:08 crc kubenswrapper[4754]: I0126 16:48:08.224001 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:08 crc kubenswrapper[4754]: I0126 16:48:08.224012 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:08Z","lastTransitionTime":"2026-01-26T16:48:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:08 crc kubenswrapper[4754]: I0126 16:48:08.227976 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Jan 26 16:48:08 crc kubenswrapper[4754]: I0126 16:48:08.231641 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:08Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:08 crc kubenswrapper[4754]: I0126 16:48:08.243575 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e459fd8c1093d6f3d0a671751b792b26b7df94c21ced1be1a2250821bcc82d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:08Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:08 crc kubenswrapper[4754]: I0126 16:48:08.254194 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2blzw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca65468e-30c6-4666-962e-cc0de05e67fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4603235bb2af495901926b805fe1c0ffc48824a4394ccf928e7fcd53a673ac3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c25z4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:33Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2blzw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:08Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:08 crc kubenswrapper[4754]: I0126 16:48:08.269252 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bwpd6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"351cc6c0-b8e5-4589-ae63-f98e8ab1eca3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1022f101d36873b41bbc1ed8a4a2590347b49f7cce6a172aa1d5bc9f529aed65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3ce0fe080a094316b9456057e347b90a5ff432daa12b9c56d7aea9bee1f72b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3ce0fe080a094316b9456057e347b90a5ff432daa12b9c56d7aea9bee1f72b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1772bb41eb8a36a863d3c691610bcbee5d16b260480b3e9f9d6b083e09e42799\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1772bb41eb8a36a863d3c691610bcbee5d16b260480b3e9f9d6b083e09e42799\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24848d69756024cc664affb9492764ace0638ef611f910f4f2bf86b638cc443d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://24848d69756024cc664affb9492764ace0638ef611f910f4f2bf86b638cc443d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c264fc7c3300a725566acce4464f01d9314779737624401c164d867831fd93bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c264fc7c3300a725566acce4464f01d9314779737624401c164d867831fd93bf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5acfe327eee4aad927ac3a32d24549eff788b14a1c4bdac359b6b56301d884ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5acfe327eee4aad927ac3a32d24549eff788b14a1c4bdac359b6b56301d884ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://63a9b9317494d0e0b33970ef3d461729b10d2a1ad7d3d2bd33f389b184ca56ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://63a9b9317494d0e0b33970ef3d461729b10d2a1ad7d3d2bd33f389b184ca56ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bwpd6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:08Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:08 crc kubenswrapper[4754]: I0126 16:48:08.288486 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jsbxt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"689569c4-93ee-4b82-93fc-9af0f97e6dc3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://033f8b104d4e9ff6d5f87ca998665f12e0c56696b3d30e01ede05f4da540365d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://919572e731813a98c7d8a94879bba931f75c472c7cd681e86d5197ebab43815d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1ac1f49bd564750fcff530a0f76657b0ed41573539acd020a922d19861ef834\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a300a80ad87f46ba1b4fd7ad3cf6c25767fcfe966b93bdd86d729843f53a5cda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31dcab2f9a2d218b0e49dad1df46377ff4521592824a0fbac254d5b8e6d0632e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fe4d9395447ac570eca1338e749dbefea76295d427a12487b63a142a7d6a2e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e9d45c28e67d857b1c50ed78197ed052c1ac2b98f1c8a94addd63698bf0bb23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0e9d45c28e67d857b1c50ed78197ed052c1ac2b98f1c8a94addd63698bf0bb23\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-26T16:47:58Z\\\",\\\"message\\\":\\\":303] Retry object setup: *v1.Pod openshift-machine-config-operator/machine-config-daemon-x65wv\\\\nI0126 16:47:58.625627 6452 obj_retry.go:303] Retry object setup: *v1.Pod openshift-multus/multus-skh5n\\\\nI0126 16:47:58.625718 6452 obj_retry.go:365] Adding new object: *v1.Pod openshift-machine-config-operator/machine-config-daemon-x65wv\\\\nI0126 16:47:58.625617 6452 obj_retry.go:303] Retry object setup: *v1.Pod openshift-multus/multus-additional-cni-plugins-bwpd6\\\\nI0126 16:47:58.625726 6452 ovn.go:134] Ensuring zone local for Pod openshift-machine-config-operator/machine-config-daemon-x65wv in node crc\\\\nI0126 16:47:58.625727 6452 obj_retry.go:365] Adding new object: *v1.Pod openshift-multus/multus-skh5n\\\\nI0126 16:47:58.625732 6452 obj_retry.go:365] Adding new object: *v1.Pod openshift-multus/multus-additional-cni-plugins-bwpd6\\\\nI0126 16:47:58.625706 6452 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {eb8eef51-1a8d-43f9-ae2e-3b2cc00ded60}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0126 16:47:58.625740 6452 ovn.go:134] Ensuring zone local for Pod openshift-multus/multus-additional-cni-plugins-bwpd6 in no\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:57Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-jsbxt_openshift-ovn-kubernetes(689569c4-93ee-4b82-93fc-9af0f97e6dc3)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://07a43ff7860ff65faf51f4a5e8aeb326f06d6d7222677a92560ae2e3a0c81484\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d9a7bafabd44f8b88a6f62cbe45599af8563bc45d9d7ad3a7669f3d8d11eb9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d9a7bafabd44f8b88a6f62cbe45599af8563bc45d9d7ad3a7669f3d8d11eb9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jsbxt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:08Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:08 crc kubenswrapper[4754]: I0126 16:48:08.301735 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"099af702-abf2-4933-b4c9-63ef2326e468\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23017e73f8450839b4c2c02fa61235ffe4604bdce11556def27f25bcadcf89dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a74507c941b9de4d0401c8f1d53bb79ed11359628d3405cdab7f5474a75c862e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4402182b3e2e35a7d5dd4c44ac53e152915b312b7ccb076c4f2b41b925a6e7f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f772a003566198519fe1849fd14f29d4346004b2e655810662507c95b4d5ffa1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:13Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:08Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:08 crc kubenswrapper[4754]: I0126 16:48:08.314607 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:08Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:08 crc kubenswrapper[4754]: I0126 16:48:08.326496 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:08 crc kubenswrapper[4754]: I0126 16:48:08.326539 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:08 crc kubenswrapper[4754]: I0126 16:48:08.326551 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:08 crc kubenswrapper[4754]: I0126 16:48:08.326568 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:08 crc kubenswrapper[4754]: I0126 16:48:08.326581 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:08Z","lastTransitionTime":"2026-01-26T16:48:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:08 crc kubenswrapper[4754]: I0126 16:48:08.330223 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89251fa9c64bc25d6d6f355400d7644c099f22aed1cae812fc991a03595f1e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e90508d8a53be9146f9b388750730fdae9ee200306caf5ad393a356390a69ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:08Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:08 crc kubenswrapper[4754]: I0126 16:48:08.343603 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-84lgq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6763d962-0892-4acc-bab9-7a4e93e39a85\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://953c1917fca896c3def3fc2c69f9e3dbdc9ca26fc2b3d40282be71f71759fd68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8zlw6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:33Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-84lgq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:08Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:08 crc kubenswrapper[4754]: I0126 16:48:08.358061 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-skh5n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b619d233-b592-4b05-a0b7-dc094c88471e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9231a4191565e2e0ef53d8a394bb34615c900b601a0dfec24c2d8e5fa7b0b659\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-skh5n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:08Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:08 crc kubenswrapper[4754]: I0126 16:48:08.372309 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jspvh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5c81aab0-939e-45ce-b19a-89d4818980c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://356dc54150a495dd6f35e0667c82b807b712b07200d8a0adbb1c92fe3f7af8f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8bps\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://05fd57780c44cd127859dc8b163d20eafbd8b06f3348cf545ef41021ce6bc526\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8bps\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-jspvh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:08Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:08 crc kubenswrapper[4754]: I0126 16:48:08.386040 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b89aa081-0a2e-4757-9210-e2a6c228bee9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9c5f40188fda14b14128514fe84bddfe8dcede3d6c560283ccb2ac0c66895fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f45445cbf1d180d6d4dd9588e85bf678f043bfce8cff0812b78ca3510904f9a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://085fa7b8d4fef8b2722fac8c346107699b5a6bd33fbfc74ed35831cb8fdf9dad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c40e018a0979035827963efb9efc48a878a55b9df68ec1a9242d7cff64fe480\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8356f76fbc3a4e636c12d1da5455d3f6767d9b83ce2f2925a11435d7d7261e9f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0126 16:47:26.205208 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0126 16:47:26.206538 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3645934156/tls.crt::/tmp/serving-cert-3645934156/tls.key\\\\\\\"\\\\nI0126 16:47:31.504518 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0126 16:47:31.509366 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0126 16:47:31.509400 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0126 16:47:31.509428 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0126 16:47:31.509436 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0126 16:47:31.515912 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0126 16:47:31.516036 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 16:47:31.516068 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 16:47:31.516096 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0126 16:47:31.516121 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0126 16:47:31.516145 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0126 16:47:31.516170 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0126 16:47:31.516412 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0126 16:47:31.519294 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27b74e2e65a547e0e28fddd4f15feacb17f4d57596eeb8da82df9d39243c85e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22ba7649a1308815bfad52016000987a7a9a5bc214d7ff61e137ba099b20fc6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://22ba7649a1308815bfad52016000987a7a9a5bc214d7ff61e137ba099b20fc6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:13Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:08Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:08 crc kubenswrapper[4754]: I0126 16:48:08.398825 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://205ec256e46a8e08b4b0b0805fb6c5f4966fc3ccef0738df4b97cb404e74da17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:08Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:08 crc kubenswrapper[4754]: I0126 16:48:08.410129 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:08Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:08 crc kubenswrapper[4754]: I0126 16:48:08.425360 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c3718a4-f354-4284-92e0-fdfb45a692bd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b86f46f49163bca7eca9458279029af330dbd4890b03325e06fa3b1bdf030de6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxjcz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fffc03b536bfee3c7bc1bddd500e834e34b391896477a1501e3e06201a19374\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxjcz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-x65wv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:08Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:08 crc kubenswrapper[4754]: I0126 16:48:08.428825 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:08 crc kubenswrapper[4754]: I0126 16:48:08.428857 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:08 crc kubenswrapper[4754]: I0126 16:48:08.428866 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:08 crc kubenswrapper[4754]: I0126 16:48:08.428878 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:08 crc kubenswrapper[4754]: I0126 16:48:08.428886 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:08Z","lastTransitionTime":"2026-01-26T16:48:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:08 crc kubenswrapper[4754]: I0126 16:48:08.439305 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-t4ps6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d45179f0-9f81-4d77-8805-6b05bdb32f8f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vk8tg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vk8tg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:48Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-t4ps6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:08Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:08 crc kubenswrapper[4754]: I0126 16:48:08.530960 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:08 crc kubenswrapper[4754]: I0126 16:48:08.531035 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:08 crc kubenswrapper[4754]: I0126 16:48:08.531052 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:08 crc kubenswrapper[4754]: I0126 16:48:08.531078 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:08 crc kubenswrapper[4754]: I0126 16:48:08.531095 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:08Z","lastTransitionTime":"2026-01-26T16:48:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:08 crc kubenswrapper[4754]: I0126 16:48:08.633538 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:08 crc kubenswrapper[4754]: I0126 16:48:08.633573 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:08 crc kubenswrapper[4754]: I0126 16:48:08.633581 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:08 crc kubenswrapper[4754]: I0126 16:48:08.633593 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:08 crc kubenswrapper[4754]: I0126 16:48:08.633601 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:08Z","lastTransitionTime":"2026-01-26T16:48:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:08 crc kubenswrapper[4754]: I0126 16:48:08.736049 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:08 crc kubenswrapper[4754]: I0126 16:48:08.736090 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:08 crc kubenswrapper[4754]: I0126 16:48:08.736101 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:08 crc kubenswrapper[4754]: I0126 16:48:08.736116 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:08 crc kubenswrapper[4754]: I0126 16:48:08.736126 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:08Z","lastTransitionTime":"2026-01-26T16:48:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:08 crc kubenswrapper[4754]: I0126 16:48:08.746604 4754 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-30 07:30:33.697688683 +0000 UTC Jan 26 16:48:08 crc kubenswrapper[4754]: I0126 16:48:08.838569 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:08 crc kubenswrapper[4754]: I0126 16:48:08.838626 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:08 crc kubenswrapper[4754]: I0126 16:48:08.838638 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:08 crc kubenswrapper[4754]: I0126 16:48:08.838655 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:08 crc kubenswrapper[4754]: I0126 16:48:08.838706 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:08Z","lastTransitionTime":"2026-01-26T16:48:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:08 crc kubenswrapper[4754]: I0126 16:48:08.942090 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:08 crc kubenswrapper[4754]: I0126 16:48:08.942135 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:08 crc kubenswrapper[4754]: I0126 16:48:08.942146 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:08 crc kubenswrapper[4754]: I0126 16:48:08.942161 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:08 crc kubenswrapper[4754]: I0126 16:48:08.942172 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:08Z","lastTransitionTime":"2026-01-26T16:48:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:09 crc kubenswrapper[4754]: I0126 16:48:09.044747 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:09 crc kubenswrapper[4754]: I0126 16:48:09.044831 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:09 crc kubenswrapper[4754]: I0126 16:48:09.044855 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:09 crc kubenswrapper[4754]: I0126 16:48:09.044888 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:09 crc kubenswrapper[4754]: I0126 16:48:09.044912 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:09Z","lastTransitionTime":"2026-01-26T16:48:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:09 crc kubenswrapper[4754]: I0126 16:48:09.147801 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:09 crc kubenswrapper[4754]: I0126 16:48:09.147858 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:09 crc kubenswrapper[4754]: I0126 16:48:09.147876 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:09 crc kubenswrapper[4754]: I0126 16:48:09.147898 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:09 crc kubenswrapper[4754]: I0126 16:48:09.147913 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:09Z","lastTransitionTime":"2026-01-26T16:48:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:09 crc kubenswrapper[4754]: I0126 16:48:09.250412 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:09 crc kubenswrapper[4754]: I0126 16:48:09.250450 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:09 crc kubenswrapper[4754]: I0126 16:48:09.250459 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:09 crc kubenswrapper[4754]: I0126 16:48:09.250472 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:09 crc kubenswrapper[4754]: I0126 16:48:09.250482 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:09Z","lastTransitionTime":"2026-01-26T16:48:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:09 crc kubenswrapper[4754]: I0126 16:48:09.353092 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:09 crc kubenswrapper[4754]: I0126 16:48:09.353141 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:09 crc kubenswrapper[4754]: I0126 16:48:09.353150 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:09 crc kubenswrapper[4754]: I0126 16:48:09.353166 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:09 crc kubenswrapper[4754]: I0126 16:48:09.353176 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:09Z","lastTransitionTime":"2026-01-26T16:48:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:09 crc kubenswrapper[4754]: I0126 16:48:09.455253 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:09 crc kubenswrapper[4754]: I0126 16:48:09.455306 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:09 crc kubenswrapper[4754]: I0126 16:48:09.455323 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:09 crc kubenswrapper[4754]: I0126 16:48:09.455341 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:09 crc kubenswrapper[4754]: I0126 16:48:09.455351 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:09Z","lastTransitionTime":"2026-01-26T16:48:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:09 crc kubenswrapper[4754]: I0126 16:48:09.557509 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:09 crc kubenswrapper[4754]: I0126 16:48:09.557552 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:09 crc kubenswrapper[4754]: I0126 16:48:09.557561 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:09 crc kubenswrapper[4754]: I0126 16:48:09.557575 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:09 crc kubenswrapper[4754]: I0126 16:48:09.557586 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:09Z","lastTransitionTime":"2026-01-26T16:48:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:09 crc kubenswrapper[4754]: I0126 16:48:09.659591 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:09 crc kubenswrapper[4754]: I0126 16:48:09.659632 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:09 crc kubenswrapper[4754]: I0126 16:48:09.659642 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:09 crc kubenswrapper[4754]: I0126 16:48:09.659657 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:09 crc kubenswrapper[4754]: I0126 16:48:09.659686 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:09Z","lastTransitionTime":"2026-01-26T16:48:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:09 crc kubenswrapper[4754]: I0126 16:48:09.746827 4754 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-10 23:31:29.189303229 +0000 UTC Jan 26 16:48:09 crc kubenswrapper[4754]: I0126 16:48:09.762407 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:09 crc kubenswrapper[4754]: I0126 16:48:09.762448 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:09 crc kubenswrapper[4754]: I0126 16:48:09.762456 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:09 crc kubenswrapper[4754]: I0126 16:48:09.762472 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:09 crc kubenswrapper[4754]: I0126 16:48:09.762482 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:09Z","lastTransitionTime":"2026-01-26T16:48:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:09 crc kubenswrapper[4754]: I0126 16:48:09.766824 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 16:48:09 crc kubenswrapper[4754]: I0126 16:48:09.766916 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-t4ps6" Jan 26 16:48:09 crc kubenswrapper[4754]: E0126 16:48:09.766947 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 16:48:09 crc kubenswrapper[4754]: E0126 16:48:09.767056 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-t4ps6" podUID="d45179f0-9f81-4d77-8805-6b05bdb32f8f" Jan 26 16:48:09 crc kubenswrapper[4754]: I0126 16:48:09.767110 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 16:48:09 crc kubenswrapper[4754]: E0126 16:48:09.767171 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 16:48:09 crc kubenswrapper[4754]: I0126 16:48:09.767340 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 16:48:09 crc kubenswrapper[4754]: E0126 16:48:09.767419 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 16:48:09 crc kubenswrapper[4754]: I0126 16:48:09.864628 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:09 crc kubenswrapper[4754]: I0126 16:48:09.864658 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:09 crc kubenswrapper[4754]: I0126 16:48:09.864691 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:09 crc kubenswrapper[4754]: I0126 16:48:09.864710 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:09 crc kubenswrapper[4754]: I0126 16:48:09.864721 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:09Z","lastTransitionTime":"2026-01-26T16:48:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:09 crc kubenswrapper[4754]: I0126 16:48:09.967153 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:09 crc kubenswrapper[4754]: I0126 16:48:09.967195 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:09 crc kubenswrapper[4754]: I0126 16:48:09.967205 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:09 crc kubenswrapper[4754]: I0126 16:48:09.967222 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:09 crc kubenswrapper[4754]: I0126 16:48:09.967239 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:09Z","lastTransitionTime":"2026-01-26T16:48:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:10 crc kubenswrapper[4754]: I0126 16:48:10.069710 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:10 crc kubenswrapper[4754]: I0126 16:48:10.069762 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:10 crc kubenswrapper[4754]: I0126 16:48:10.069778 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:10 crc kubenswrapper[4754]: I0126 16:48:10.069793 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:10 crc kubenswrapper[4754]: I0126 16:48:10.069805 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:10Z","lastTransitionTime":"2026-01-26T16:48:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:10 crc kubenswrapper[4754]: I0126 16:48:10.172500 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:10 crc kubenswrapper[4754]: I0126 16:48:10.172536 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:10 crc kubenswrapper[4754]: I0126 16:48:10.172545 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:10 crc kubenswrapper[4754]: I0126 16:48:10.172558 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:10 crc kubenswrapper[4754]: I0126 16:48:10.172584 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:10Z","lastTransitionTime":"2026-01-26T16:48:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:10 crc kubenswrapper[4754]: I0126 16:48:10.276737 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:10 crc kubenswrapper[4754]: I0126 16:48:10.276792 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:10 crc kubenswrapper[4754]: I0126 16:48:10.276803 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:10 crc kubenswrapper[4754]: I0126 16:48:10.276824 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:10 crc kubenswrapper[4754]: I0126 16:48:10.276846 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:10Z","lastTransitionTime":"2026-01-26T16:48:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:10 crc kubenswrapper[4754]: I0126 16:48:10.378950 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:10 crc kubenswrapper[4754]: I0126 16:48:10.378990 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:10 crc kubenswrapper[4754]: I0126 16:48:10.378998 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:10 crc kubenswrapper[4754]: I0126 16:48:10.379013 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:10 crc kubenswrapper[4754]: I0126 16:48:10.379024 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:10Z","lastTransitionTime":"2026-01-26T16:48:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:10 crc kubenswrapper[4754]: I0126 16:48:10.480837 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:10 crc kubenswrapper[4754]: I0126 16:48:10.480880 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:10 crc kubenswrapper[4754]: I0126 16:48:10.480892 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:10 crc kubenswrapper[4754]: I0126 16:48:10.480906 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:10 crc kubenswrapper[4754]: I0126 16:48:10.480919 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:10Z","lastTransitionTime":"2026-01-26T16:48:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:10 crc kubenswrapper[4754]: I0126 16:48:10.582982 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:10 crc kubenswrapper[4754]: I0126 16:48:10.583036 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:10 crc kubenswrapper[4754]: I0126 16:48:10.583046 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:10 crc kubenswrapper[4754]: I0126 16:48:10.583071 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:10 crc kubenswrapper[4754]: I0126 16:48:10.583088 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:10Z","lastTransitionTime":"2026-01-26T16:48:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:10 crc kubenswrapper[4754]: I0126 16:48:10.685870 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:10 crc kubenswrapper[4754]: I0126 16:48:10.685925 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:10 crc kubenswrapper[4754]: I0126 16:48:10.685934 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:10 crc kubenswrapper[4754]: I0126 16:48:10.685953 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:10 crc kubenswrapper[4754]: I0126 16:48:10.685964 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:10Z","lastTransitionTime":"2026-01-26T16:48:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:10 crc kubenswrapper[4754]: I0126 16:48:10.747238 4754 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-30 23:54:38.047787287 +0000 UTC Jan 26 16:48:10 crc kubenswrapper[4754]: I0126 16:48:10.789391 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:10 crc kubenswrapper[4754]: I0126 16:48:10.789458 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:10 crc kubenswrapper[4754]: I0126 16:48:10.789472 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:10 crc kubenswrapper[4754]: I0126 16:48:10.789494 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:10 crc kubenswrapper[4754]: I0126 16:48:10.789508 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:10Z","lastTransitionTime":"2026-01-26T16:48:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:10 crc kubenswrapper[4754]: I0126 16:48:10.892952 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:10 crc kubenswrapper[4754]: I0126 16:48:10.893019 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:10 crc kubenswrapper[4754]: I0126 16:48:10.893039 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:10 crc kubenswrapper[4754]: I0126 16:48:10.893057 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:10 crc kubenswrapper[4754]: I0126 16:48:10.893076 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:10Z","lastTransitionTime":"2026-01-26T16:48:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:10 crc kubenswrapper[4754]: I0126 16:48:10.996308 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:10 crc kubenswrapper[4754]: I0126 16:48:10.996359 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:10 crc kubenswrapper[4754]: I0126 16:48:10.996370 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:10 crc kubenswrapper[4754]: I0126 16:48:10.996387 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:10 crc kubenswrapper[4754]: I0126 16:48:10.996400 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:10Z","lastTransitionTime":"2026-01-26T16:48:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:11 crc kubenswrapper[4754]: I0126 16:48:11.098753 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:11 crc kubenswrapper[4754]: I0126 16:48:11.098796 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:11 crc kubenswrapper[4754]: I0126 16:48:11.098807 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:11 crc kubenswrapper[4754]: I0126 16:48:11.098823 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:11 crc kubenswrapper[4754]: I0126 16:48:11.098835 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:11Z","lastTransitionTime":"2026-01-26T16:48:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:11 crc kubenswrapper[4754]: I0126 16:48:11.201315 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:11 crc kubenswrapper[4754]: I0126 16:48:11.201364 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:11 crc kubenswrapper[4754]: I0126 16:48:11.201376 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:11 crc kubenswrapper[4754]: I0126 16:48:11.201396 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:11 crc kubenswrapper[4754]: I0126 16:48:11.201408 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:11Z","lastTransitionTime":"2026-01-26T16:48:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:11 crc kubenswrapper[4754]: I0126 16:48:11.303709 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:11 crc kubenswrapper[4754]: I0126 16:48:11.303753 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:11 crc kubenswrapper[4754]: I0126 16:48:11.303764 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:11 crc kubenswrapper[4754]: I0126 16:48:11.303781 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:11 crc kubenswrapper[4754]: I0126 16:48:11.303793 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:11Z","lastTransitionTime":"2026-01-26T16:48:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:11 crc kubenswrapper[4754]: I0126 16:48:11.406169 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:11 crc kubenswrapper[4754]: I0126 16:48:11.406207 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:11 crc kubenswrapper[4754]: I0126 16:48:11.406219 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:11 crc kubenswrapper[4754]: I0126 16:48:11.406234 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:11 crc kubenswrapper[4754]: I0126 16:48:11.406243 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:11Z","lastTransitionTime":"2026-01-26T16:48:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:11 crc kubenswrapper[4754]: I0126 16:48:11.508934 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:11 crc kubenswrapper[4754]: I0126 16:48:11.508971 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:11 crc kubenswrapper[4754]: I0126 16:48:11.508979 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:11 crc kubenswrapper[4754]: I0126 16:48:11.508992 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:11 crc kubenswrapper[4754]: I0126 16:48:11.509001 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:11Z","lastTransitionTime":"2026-01-26T16:48:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:11 crc kubenswrapper[4754]: I0126 16:48:11.611626 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:11 crc kubenswrapper[4754]: I0126 16:48:11.611661 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:11 crc kubenswrapper[4754]: I0126 16:48:11.611698 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:11 crc kubenswrapper[4754]: I0126 16:48:11.611720 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:11 crc kubenswrapper[4754]: I0126 16:48:11.611731 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:11Z","lastTransitionTime":"2026-01-26T16:48:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:11 crc kubenswrapper[4754]: I0126 16:48:11.714295 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:11 crc kubenswrapper[4754]: I0126 16:48:11.714354 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:11 crc kubenswrapper[4754]: I0126 16:48:11.714367 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:11 crc kubenswrapper[4754]: I0126 16:48:11.714390 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:11 crc kubenswrapper[4754]: I0126 16:48:11.714419 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:11Z","lastTransitionTime":"2026-01-26T16:48:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:11 crc kubenswrapper[4754]: I0126 16:48:11.747867 4754 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-07 04:03:00.151344127 +0000 UTC Jan 26 16:48:11 crc kubenswrapper[4754]: I0126 16:48:11.772705 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 16:48:11 crc kubenswrapper[4754]: I0126 16:48:11.772726 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 16:48:11 crc kubenswrapper[4754]: I0126 16:48:11.772785 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 16:48:11 crc kubenswrapper[4754]: I0126 16:48:11.773019 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-t4ps6" Jan 26 16:48:11 crc kubenswrapper[4754]: E0126 16:48:11.773169 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 16:48:11 crc kubenswrapper[4754]: E0126 16:48:11.773216 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-t4ps6" podUID="d45179f0-9f81-4d77-8805-6b05bdb32f8f" Jan 26 16:48:11 crc kubenswrapper[4754]: I0126 16:48:11.773340 4754 scope.go:117] "RemoveContainer" containerID="0e9d45c28e67d857b1c50ed78197ed052c1ac2b98f1c8a94addd63698bf0bb23" Jan 26 16:48:11 crc kubenswrapper[4754]: E0126 16:48:11.773489 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-jsbxt_openshift-ovn-kubernetes(689569c4-93ee-4b82-93fc-9af0f97e6dc3)\"" pod="openshift-ovn-kubernetes/ovnkube-node-jsbxt" podUID="689569c4-93ee-4b82-93fc-9af0f97e6dc3" Jan 26 16:48:11 crc kubenswrapper[4754]: E0126 16:48:11.773822 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 16:48:11 crc kubenswrapper[4754]: E0126 16:48:11.773911 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 16:48:11 crc kubenswrapper[4754]: I0126 16:48:11.816864 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:11 crc kubenswrapper[4754]: I0126 16:48:11.816906 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:11 crc kubenswrapper[4754]: I0126 16:48:11.816918 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:11 crc kubenswrapper[4754]: I0126 16:48:11.816934 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:11 crc kubenswrapper[4754]: I0126 16:48:11.816944 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:11Z","lastTransitionTime":"2026-01-26T16:48:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:11 crc kubenswrapper[4754]: I0126 16:48:11.920857 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:11 crc kubenswrapper[4754]: I0126 16:48:11.920904 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:11 crc kubenswrapper[4754]: I0126 16:48:11.920917 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:11 crc kubenswrapper[4754]: I0126 16:48:11.920935 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:11 crc kubenswrapper[4754]: I0126 16:48:11.920947 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:11Z","lastTransitionTime":"2026-01-26T16:48:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:12 crc kubenswrapper[4754]: I0126 16:48:12.024250 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:12 crc kubenswrapper[4754]: I0126 16:48:12.024300 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:12 crc kubenswrapper[4754]: I0126 16:48:12.024309 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:12 crc kubenswrapper[4754]: I0126 16:48:12.024329 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:12 crc kubenswrapper[4754]: I0126 16:48:12.024347 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:12Z","lastTransitionTime":"2026-01-26T16:48:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:12 crc kubenswrapper[4754]: I0126 16:48:12.127344 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:12 crc kubenswrapper[4754]: I0126 16:48:12.127377 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:12 crc kubenswrapper[4754]: I0126 16:48:12.127389 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:12 crc kubenswrapper[4754]: I0126 16:48:12.127404 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:12 crc kubenswrapper[4754]: I0126 16:48:12.127413 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:12Z","lastTransitionTime":"2026-01-26T16:48:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:12 crc kubenswrapper[4754]: I0126 16:48:12.229745 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:12 crc kubenswrapper[4754]: I0126 16:48:12.229788 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:12 crc kubenswrapper[4754]: I0126 16:48:12.229799 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:12 crc kubenswrapper[4754]: I0126 16:48:12.229814 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:12 crc kubenswrapper[4754]: I0126 16:48:12.229827 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:12Z","lastTransitionTime":"2026-01-26T16:48:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:12 crc kubenswrapper[4754]: I0126 16:48:12.333009 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:12 crc kubenswrapper[4754]: I0126 16:48:12.333051 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:12 crc kubenswrapper[4754]: I0126 16:48:12.333063 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:12 crc kubenswrapper[4754]: I0126 16:48:12.333079 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:12 crc kubenswrapper[4754]: I0126 16:48:12.333090 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:12Z","lastTransitionTime":"2026-01-26T16:48:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:12 crc kubenswrapper[4754]: I0126 16:48:12.435138 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:12 crc kubenswrapper[4754]: I0126 16:48:12.435252 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:12 crc kubenswrapper[4754]: I0126 16:48:12.435320 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:12 crc kubenswrapper[4754]: I0126 16:48:12.435341 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:12 crc kubenswrapper[4754]: I0126 16:48:12.435352 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:12Z","lastTransitionTime":"2026-01-26T16:48:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:12 crc kubenswrapper[4754]: I0126 16:48:12.537403 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:12 crc kubenswrapper[4754]: I0126 16:48:12.537450 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:12 crc kubenswrapper[4754]: I0126 16:48:12.537461 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:12 crc kubenswrapper[4754]: I0126 16:48:12.537478 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:12 crc kubenswrapper[4754]: I0126 16:48:12.537489 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:12Z","lastTransitionTime":"2026-01-26T16:48:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:12 crc kubenswrapper[4754]: I0126 16:48:12.639885 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:12 crc kubenswrapper[4754]: I0126 16:48:12.639938 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:12 crc kubenswrapper[4754]: I0126 16:48:12.639951 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:12 crc kubenswrapper[4754]: I0126 16:48:12.639968 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:12 crc kubenswrapper[4754]: I0126 16:48:12.639980 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:12Z","lastTransitionTime":"2026-01-26T16:48:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:12 crc kubenswrapper[4754]: I0126 16:48:12.743291 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:12 crc kubenswrapper[4754]: I0126 16:48:12.743338 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:12 crc kubenswrapper[4754]: I0126 16:48:12.743351 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:12 crc kubenswrapper[4754]: I0126 16:48:12.743397 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:12 crc kubenswrapper[4754]: I0126 16:48:12.743412 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:12Z","lastTransitionTime":"2026-01-26T16:48:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:12 crc kubenswrapper[4754]: I0126 16:48:12.748138 4754 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-20 04:28:42.286065078 +0000 UTC Jan 26 16:48:12 crc kubenswrapper[4754]: I0126 16:48:12.846534 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:12 crc kubenswrapper[4754]: I0126 16:48:12.846575 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:12 crc kubenswrapper[4754]: I0126 16:48:12.846587 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:12 crc kubenswrapper[4754]: I0126 16:48:12.846602 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:12 crc kubenswrapper[4754]: I0126 16:48:12.846613 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:12Z","lastTransitionTime":"2026-01-26T16:48:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:12 crc kubenswrapper[4754]: I0126 16:48:12.948800 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:12 crc kubenswrapper[4754]: I0126 16:48:12.948871 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:12 crc kubenswrapper[4754]: I0126 16:48:12.948887 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:12 crc kubenswrapper[4754]: I0126 16:48:12.948910 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:12 crc kubenswrapper[4754]: I0126 16:48:12.948930 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:12Z","lastTransitionTime":"2026-01-26T16:48:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:13 crc kubenswrapper[4754]: I0126 16:48:13.051066 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:13 crc kubenswrapper[4754]: I0126 16:48:13.051104 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:13 crc kubenswrapper[4754]: I0126 16:48:13.051113 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:13 crc kubenswrapper[4754]: I0126 16:48:13.051126 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:13 crc kubenswrapper[4754]: I0126 16:48:13.051135 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:13Z","lastTransitionTime":"2026-01-26T16:48:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:13 crc kubenswrapper[4754]: I0126 16:48:13.153318 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:13 crc kubenswrapper[4754]: I0126 16:48:13.153366 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:13 crc kubenswrapper[4754]: I0126 16:48:13.153380 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:13 crc kubenswrapper[4754]: I0126 16:48:13.153410 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:13 crc kubenswrapper[4754]: I0126 16:48:13.153425 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:13Z","lastTransitionTime":"2026-01-26T16:48:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:13 crc kubenswrapper[4754]: I0126 16:48:13.256067 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:13 crc kubenswrapper[4754]: I0126 16:48:13.256137 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:13 crc kubenswrapper[4754]: I0126 16:48:13.256168 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:13 crc kubenswrapper[4754]: I0126 16:48:13.256185 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:13 crc kubenswrapper[4754]: I0126 16:48:13.256200 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:13Z","lastTransitionTime":"2026-01-26T16:48:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:13 crc kubenswrapper[4754]: I0126 16:48:13.359111 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:13 crc kubenswrapper[4754]: I0126 16:48:13.359159 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:13 crc kubenswrapper[4754]: I0126 16:48:13.359167 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:13 crc kubenswrapper[4754]: I0126 16:48:13.359185 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:13 crc kubenswrapper[4754]: I0126 16:48:13.359193 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:13Z","lastTransitionTime":"2026-01-26T16:48:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:13 crc kubenswrapper[4754]: I0126 16:48:13.461103 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:13 crc kubenswrapper[4754]: I0126 16:48:13.461155 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:13 crc kubenswrapper[4754]: I0126 16:48:13.461165 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:13 crc kubenswrapper[4754]: I0126 16:48:13.461179 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:13 crc kubenswrapper[4754]: I0126 16:48:13.461188 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:13Z","lastTransitionTime":"2026-01-26T16:48:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:13 crc kubenswrapper[4754]: I0126 16:48:13.563790 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:13 crc kubenswrapper[4754]: I0126 16:48:13.563856 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:13 crc kubenswrapper[4754]: I0126 16:48:13.563866 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:13 crc kubenswrapper[4754]: I0126 16:48:13.563882 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:13 crc kubenswrapper[4754]: I0126 16:48:13.563891 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:13Z","lastTransitionTime":"2026-01-26T16:48:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:13 crc kubenswrapper[4754]: I0126 16:48:13.669231 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:13 crc kubenswrapper[4754]: I0126 16:48:13.669310 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:13 crc kubenswrapper[4754]: I0126 16:48:13.669365 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:13 crc kubenswrapper[4754]: I0126 16:48:13.669696 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:13 crc kubenswrapper[4754]: I0126 16:48:13.669728 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:13Z","lastTransitionTime":"2026-01-26T16:48:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:13 crc kubenswrapper[4754]: I0126 16:48:13.748990 4754 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-30 17:18:01.8116393 +0000 UTC Jan 26 16:48:13 crc kubenswrapper[4754]: I0126 16:48:13.767000 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 16:48:13 crc kubenswrapper[4754]: I0126 16:48:13.767073 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 16:48:13 crc kubenswrapper[4754]: I0126 16:48:13.767004 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 16:48:13 crc kubenswrapper[4754]: E0126 16:48:13.767128 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 16:48:13 crc kubenswrapper[4754]: E0126 16:48:13.767229 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 16:48:13 crc kubenswrapper[4754]: I0126 16:48:13.767273 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-t4ps6" Jan 26 16:48:13 crc kubenswrapper[4754]: E0126 16:48:13.767325 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 16:48:13 crc kubenswrapper[4754]: E0126 16:48:13.767416 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-t4ps6" podUID="d45179f0-9f81-4d77-8805-6b05bdb32f8f" Jan 26 16:48:13 crc kubenswrapper[4754]: I0126 16:48:13.775240 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:13 crc kubenswrapper[4754]: I0126 16:48:13.775273 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:13 crc kubenswrapper[4754]: I0126 16:48:13.775281 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:13 crc kubenswrapper[4754]: I0126 16:48:13.775295 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:13 crc kubenswrapper[4754]: I0126 16:48:13.775305 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:13Z","lastTransitionTime":"2026-01-26T16:48:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:13 crc kubenswrapper[4754]: I0126 16:48:13.785067 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"099af702-abf2-4933-b4c9-63ef2326e468\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23017e73f8450839b4c2c02fa61235ffe4604bdce11556def27f25bcadcf89dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a74507c941b9de4d0401c8f1d53bb79ed11359628d3405cdab7f5474a75c862e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4402182b3e2e35a7d5dd4c44ac53e152915b312b7ccb076c4f2b41b925a6e7f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f772a003566198519fe1849fd14f29d4346004b2e655810662507c95b4d5ffa1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:13Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:13Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:13 crc kubenswrapper[4754]: I0126 16:48:13.811060 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab1ace4c-137b-4b8a-8032-0b86a164f063\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:48:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:48:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7da43edaa2a81f1b647b96c234d795b1ddf5fa23173b78043123f76dc677c517\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e5fbff42fdebad6a74c21d85931d3c8d983973db82fc7f262ba5383bad76172\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f9bf9e4a705cd849595e59e1123ba70773bfa5f8a0d0b79195e4fc7e447ca26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb7a3b85cca4436068c8ceefc3f2358cfa1a2c398a8acecd367b944528b11489\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb7a3b85cca4436068c8ceefc3f2358cfa1a2c398a8acecd367b944528b11489\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:14Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:13Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:13Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:13 crc kubenswrapper[4754]: I0126 16:48:13.840027 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:13Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:13 crc kubenswrapper[4754]: I0126 16:48:13.854313 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89251fa9c64bc25d6d6f355400d7644c099f22aed1cae812fc991a03595f1e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e90508d8a53be9146f9b388750730fdae9ee200306caf5ad393a356390a69ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:13Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:13 crc kubenswrapper[4754]: I0126 16:48:13.866267 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-84lgq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6763d962-0892-4acc-bab9-7a4e93e39a85\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://953c1917fca896c3def3fc2c69f9e3dbdc9ca26fc2b3d40282be71f71759fd68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8zlw6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:33Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-84lgq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:13Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:13 crc kubenswrapper[4754]: I0126 16:48:13.877493 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:13 crc kubenswrapper[4754]: I0126 16:48:13.877535 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:13 crc kubenswrapper[4754]: I0126 16:48:13.877546 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:13 crc kubenswrapper[4754]: I0126 16:48:13.877563 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:13 crc kubenswrapper[4754]: I0126 16:48:13.877577 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:13Z","lastTransitionTime":"2026-01-26T16:48:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:13 crc kubenswrapper[4754]: I0126 16:48:13.881716 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-skh5n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b619d233-b592-4b05-a0b7-dc094c88471e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9231a4191565e2e0ef53d8a394bb34615c900b601a0dfec24c2d8e5fa7b0b659\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-skh5n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:13Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:13 crc kubenswrapper[4754]: I0126 16:48:13.893022 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jspvh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5c81aab0-939e-45ce-b19a-89d4818980c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://356dc54150a495dd6f35e0667c82b807b712b07200d8a0adbb1c92fe3f7af8f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8bps\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://05fd57780c44cd127859dc8b163d20eafbd8b06f3348cf545ef41021ce6bc526\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8bps\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-jspvh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:13Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:13 crc kubenswrapper[4754]: I0126 16:48:13.907077 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b89aa081-0a2e-4757-9210-e2a6c228bee9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9c5f40188fda14b14128514fe84bddfe8dcede3d6c560283ccb2ac0c66895fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f45445cbf1d180d6d4dd9588e85bf678f043bfce8cff0812b78ca3510904f9a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://085fa7b8d4fef8b2722fac8c346107699b5a6bd33fbfc74ed35831cb8fdf9dad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c40e018a0979035827963efb9efc48a878a55b9df68ec1a9242d7cff64fe480\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8356f76fbc3a4e636c12d1da5455d3f6767d9b83ce2f2925a11435d7d7261e9f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0126 16:47:26.205208 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0126 16:47:26.206538 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3645934156/tls.crt::/tmp/serving-cert-3645934156/tls.key\\\\\\\"\\\\nI0126 16:47:31.504518 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0126 16:47:31.509366 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0126 16:47:31.509400 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0126 16:47:31.509428 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0126 16:47:31.509436 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0126 16:47:31.515912 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0126 16:47:31.516036 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 16:47:31.516068 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 16:47:31.516096 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0126 16:47:31.516121 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0126 16:47:31.516145 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0126 16:47:31.516170 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0126 16:47:31.516412 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0126 16:47:31.519294 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27b74e2e65a547e0e28fddd4f15feacb17f4d57596eeb8da82df9d39243c85e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22ba7649a1308815bfad52016000987a7a9a5bc214d7ff61e137ba099b20fc6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://22ba7649a1308815bfad52016000987a7a9a5bc214d7ff61e137ba099b20fc6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:13Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:13Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:13 crc kubenswrapper[4754]: I0126 16:48:13.919453 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://205ec256e46a8e08b4b0b0805fb6c5f4966fc3ccef0738df4b97cb404e74da17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:13Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:13 crc kubenswrapper[4754]: I0126 16:48:13.931833 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:13Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:13 crc kubenswrapper[4754]: I0126 16:48:13.944145 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c3718a4-f354-4284-92e0-fdfb45a692bd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b86f46f49163bca7eca9458279029af330dbd4890b03325e06fa3b1bdf030de6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxjcz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fffc03b536bfee3c7bc1bddd500e834e34b391896477a1501e3e06201a19374\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxjcz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-x65wv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:13Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:13 crc kubenswrapper[4754]: I0126 16:48:13.958615 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-t4ps6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d45179f0-9f81-4d77-8805-6b05bdb32f8f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vk8tg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vk8tg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:48Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-t4ps6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:13Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:13 crc kubenswrapper[4754]: I0126 16:48:13.970419 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:13Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:13 crc kubenswrapper[4754]: I0126 16:48:13.987797 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e459fd8c1093d6f3d0a671751b792b26b7df94c21ced1be1a2250821bcc82d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:13Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:13 crc kubenswrapper[4754]: I0126 16:48:13.988451 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:13 crc kubenswrapper[4754]: I0126 16:48:13.988527 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:13 crc kubenswrapper[4754]: I0126 16:48:13.988546 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:13 crc kubenswrapper[4754]: I0126 16:48:13.988570 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:13 crc kubenswrapper[4754]: I0126 16:48:13.988586 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:13Z","lastTransitionTime":"2026-01-26T16:48:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:13 crc kubenswrapper[4754]: I0126 16:48:13.999209 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2blzw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca65468e-30c6-4666-962e-cc0de05e67fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4603235bb2af495901926b805fe1c0ffc48824a4394ccf928e7fcd53a673ac3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c25z4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:33Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2blzw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:13Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:14 crc kubenswrapper[4754]: I0126 16:48:14.014517 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bwpd6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"351cc6c0-b8e5-4589-ae63-f98e8ab1eca3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1022f101d36873b41bbc1ed8a4a2590347b49f7cce6a172aa1d5bc9f529aed65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3ce0fe080a094316b9456057e347b90a5ff432daa12b9c56d7aea9bee1f72b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3ce0fe080a094316b9456057e347b90a5ff432daa12b9c56d7aea9bee1f72b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1772bb41eb8a36a863d3c691610bcbee5d16b260480b3e9f9d6b083e09e42799\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1772bb41eb8a36a863d3c691610bcbee5d16b260480b3e9f9d6b083e09e42799\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24848d69756024cc664affb9492764ace0638ef611f910f4f2bf86b638cc443d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://24848d69756024cc664affb9492764ace0638ef611f910f4f2bf86b638cc443d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c264fc7c3300a725566acce4464f01d9314779737624401c164d867831fd93bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c264fc7c3300a725566acce4464f01d9314779737624401c164d867831fd93bf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5acfe327eee4aad927ac3a32d24549eff788b14a1c4bdac359b6b56301d884ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5acfe327eee4aad927ac3a32d24549eff788b14a1c4bdac359b6b56301d884ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://63a9b9317494d0e0b33970ef3d461729b10d2a1ad7d3d2bd33f389b184ca56ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://63a9b9317494d0e0b33970ef3d461729b10d2a1ad7d3d2bd33f389b184ca56ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bwpd6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:14Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:14 crc kubenswrapper[4754]: I0126 16:48:14.035832 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jsbxt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"689569c4-93ee-4b82-93fc-9af0f97e6dc3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://033f8b104d4e9ff6d5f87ca998665f12e0c56696b3d30e01ede05f4da540365d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://919572e731813a98c7d8a94879bba931f75c472c7cd681e86d5197ebab43815d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1ac1f49bd564750fcff530a0f76657b0ed41573539acd020a922d19861ef834\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a300a80ad87f46ba1b4fd7ad3cf6c25767fcfe966b93bdd86d729843f53a5cda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31dcab2f9a2d218b0e49dad1df46377ff4521592824a0fbac254d5b8e6d0632e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fe4d9395447ac570eca1338e749dbefea76295d427a12487b63a142a7d6a2e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e9d45c28e67d857b1c50ed78197ed052c1ac2b98f1c8a94addd63698bf0bb23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0e9d45c28e67d857b1c50ed78197ed052c1ac2b98f1c8a94addd63698bf0bb23\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-26T16:47:58Z\\\",\\\"message\\\":\\\":303] Retry object setup: *v1.Pod openshift-machine-config-operator/machine-config-daemon-x65wv\\\\nI0126 16:47:58.625627 6452 obj_retry.go:303] Retry object setup: *v1.Pod openshift-multus/multus-skh5n\\\\nI0126 16:47:58.625718 6452 obj_retry.go:365] Adding new object: *v1.Pod openshift-machine-config-operator/machine-config-daemon-x65wv\\\\nI0126 16:47:58.625617 6452 obj_retry.go:303] Retry object setup: *v1.Pod openshift-multus/multus-additional-cni-plugins-bwpd6\\\\nI0126 16:47:58.625726 6452 ovn.go:134] Ensuring zone local for Pod openshift-machine-config-operator/machine-config-daemon-x65wv in node crc\\\\nI0126 16:47:58.625727 6452 obj_retry.go:365] Adding new object: *v1.Pod openshift-multus/multus-skh5n\\\\nI0126 16:47:58.625732 6452 obj_retry.go:365] Adding new object: *v1.Pod openshift-multus/multus-additional-cni-plugins-bwpd6\\\\nI0126 16:47:58.625706 6452 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {eb8eef51-1a8d-43f9-ae2e-3b2cc00ded60}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0126 16:47:58.625740 6452 ovn.go:134] Ensuring zone local for Pod openshift-multus/multus-additional-cni-plugins-bwpd6 in no\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:57Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-jsbxt_openshift-ovn-kubernetes(689569c4-93ee-4b82-93fc-9af0f97e6dc3)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://07a43ff7860ff65faf51f4a5e8aeb326f06d6d7222677a92560ae2e3a0c81484\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d9a7bafabd44f8b88a6f62cbe45599af8563bc45d9d7ad3a7669f3d8d11eb9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d9a7bafabd44f8b88a6f62cbe45599af8563bc45d9d7ad3a7669f3d8d11eb9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jsbxt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:14Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:14 crc kubenswrapper[4754]: I0126 16:48:14.090902 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:14 crc kubenswrapper[4754]: I0126 16:48:14.090952 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:14 crc kubenswrapper[4754]: I0126 16:48:14.090962 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:14 crc kubenswrapper[4754]: I0126 16:48:14.090976 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:14 crc kubenswrapper[4754]: I0126 16:48:14.090985 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:14Z","lastTransitionTime":"2026-01-26T16:48:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:14 crc kubenswrapper[4754]: I0126 16:48:14.194509 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:14 crc kubenswrapper[4754]: I0126 16:48:14.194557 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:14 crc kubenswrapper[4754]: I0126 16:48:14.194566 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:14 crc kubenswrapper[4754]: I0126 16:48:14.194586 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:14 crc kubenswrapper[4754]: I0126 16:48:14.194600 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:14Z","lastTransitionTime":"2026-01-26T16:48:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:14 crc kubenswrapper[4754]: I0126 16:48:14.297356 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:14 crc kubenswrapper[4754]: I0126 16:48:14.297402 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:14 crc kubenswrapper[4754]: I0126 16:48:14.297419 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:14 crc kubenswrapper[4754]: I0126 16:48:14.297439 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:14 crc kubenswrapper[4754]: I0126 16:48:14.297452 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:14Z","lastTransitionTime":"2026-01-26T16:48:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:14 crc kubenswrapper[4754]: I0126 16:48:14.400550 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:14 crc kubenswrapper[4754]: I0126 16:48:14.400606 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:14 crc kubenswrapper[4754]: I0126 16:48:14.400622 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:14 crc kubenswrapper[4754]: I0126 16:48:14.400645 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:14 crc kubenswrapper[4754]: I0126 16:48:14.400658 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:14Z","lastTransitionTime":"2026-01-26T16:48:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:14 crc kubenswrapper[4754]: I0126 16:48:14.503597 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:14 crc kubenswrapper[4754]: I0126 16:48:14.503909 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:14 crc kubenswrapper[4754]: I0126 16:48:14.503918 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:14 crc kubenswrapper[4754]: I0126 16:48:14.503932 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:14 crc kubenswrapper[4754]: I0126 16:48:14.503952 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:14Z","lastTransitionTime":"2026-01-26T16:48:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:14 crc kubenswrapper[4754]: I0126 16:48:14.606649 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:14 crc kubenswrapper[4754]: I0126 16:48:14.606721 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:14 crc kubenswrapper[4754]: I0126 16:48:14.606734 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:14 crc kubenswrapper[4754]: I0126 16:48:14.606748 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:14 crc kubenswrapper[4754]: I0126 16:48:14.606759 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:14Z","lastTransitionTime":"2026-01-26T16:48:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:14 crc kubenswrapper[4754]: I0126 16:48:14.710297 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:14 crc kubenswrapper[4754]: I0126 16:48:14.710563 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:14 crc kubenswrapper[4754]: I0126 16:48:14.710714 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:14 crc kubenswrapper[4754]: I0126 16:48:14.710807 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:14 crc kubenswrapper[4754]: I0126 16:48:14.710877 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:14Z","lastTransitionTime":"2026-01-26T16:48:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:14 crc kubenswrapper[4754]: I0126 16:48:14.750066 4754 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-06 21:11:47.336626335 +0000 UTC Jan 26 16:48:14 crc kubenswrapper[4754]: I0126 16:48:14.813758 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:14 crc kubenswrapper[4754]: I0126 16:48:14.813803 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:14 crc kubenswrapper[4754]: I0126 16:48:14.813822 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:14 crc kubenswrapper[4754]: I0126 16:48:14.813841 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:14 crc kubenswrapper[4754]: I0126 16:48:14.813851 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:14Z","lastTransitionTime":"2026-01-26T16:48:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:14 crc kubenswrapper[4754]: I0126 16:48:14.915814 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:14 crc kubenswrapper[4754]: I0126 16:48:14.915845 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:14 crc kubenswrapper[4754]: I0126 16:48:14.915855 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:14 crc kubenswrapper[4754]: I0126 16:48:14.915869 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:14 crc kubenswrapper[4754]: I0126 16:48:14.915878 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:14Z","lastTransitionTime":"2026-01-26T16:48:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:15 crc kubenswrapper[4754]: I0126 16:48:15.018914 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:15 crc kubenswrapper[4754]: I0126 16:48:15.018954 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:15 crc kubenswrapper[4754]: I0126 16:48:15.018962 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:15 crc kubenswrapper[4754]: I0126 16:48:15.018978 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:15 crc kubenswrapper[4754]: I0126 16:48:15.018987 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:15Z","lastTransitionTime":"2026-01-26T16:48:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:15 crc kubenswrapper[4754]: I0126 16:48:15.121227 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:15 crc kubenswrapper[4754]: I0126 16:48:15.121275 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:15 crc kubenswrapper[4754]: I0126 16:48:15.121286 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:15 crc kubenswrapper[4754]: I0126 16:48:15.121305 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:15 crc kubenswrapper[4754]: I0126 16:48:15.121316 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:15Z","lastTransitionTime":"2026-01-26T16:48:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:15 crc kubenswrapper[4754]: I0126 16:48:15.223773 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:15 crc kubenswrapper[4754]: I0126 16:48:15.223816 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:15 crc kubenswrapper[4754]: I0126 16:48:15.223829 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:15 crc kubenswrapper[4754]: I0126 16:48:15.223843 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:15 crc kubenswrapper[4754]: I0126 16:48:15.223852 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:15Z","lastTransitionTime":"2026-01-26T16:48:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:15 crc kubenswrapper[4754]: I0126 16:48:15.296586 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:15 crc kubenswrapper[4754]: I0126 16:48:15.296631 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:15 crc kubenswrapper[4754]: I0126 16:48:15.296644 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:15 crc kubenswrapper[4754]: I0126 16:48:15.296661 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:15 crc kubenswrapper[4754]: I0126 16:48:15.296691 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:15Z","lastTransitionTime":"2026-01-26T16:48:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:15 crc kubenswrapper[4754]: E0126 16:48:15.309963 4754 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:48:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:48:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:48:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:48:15Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:48:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:48:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:48:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:48:15Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d8506764-e7fa-45cb-a13d-6f527164f548\\\",\\\"systemUUID\\\":\\\"bfbb6c33-ce46-4e81-a8a8-b44409b03821\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:15Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:15 crc kubenswrapper[4754]: I0126 16:48:15.316375 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:15 crc kubenswrapper[4754]: I0126 16:48:15.316418 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:15 crc kubenswrapper[4754]: I0126 16:48:15.316429 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:15 crc kubenswrapper[4754]: I0126 16:48:15.316445 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:15 crc kubenswrapper[4754]: I0126 16:48:15.316460 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:15Z","lastTransitionTime":"2026-01-26T16:48:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:15 crc kubenswrapper[4754]: E0126 16:48:15.328908 4754 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:48:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:48:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:48:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:48:15Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:48:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:48:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:48:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:48:15Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d8506764-e7fa-45cb-a13d-6f527164f548\\\",\\\"systemUUID\\\":\\\"bfbb6c33-ce46-4e81-a8a8-b44409b03821\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:15Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:15 crc kubenswrapper[4754]: I0126 16:48:15.333602 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:15 crc kubenswrapper[4754]: I0126 16:48:15.333645 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:15 crc kubenswrapper[4754]: I0126 16:48:15.333658 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:15 crc kubenswrapper[4754]: I0126 16:48:15.333690 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:15 crc kubenswrapper[4754]: I0126 16:48:15.333702 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:15Z","lastTransitionTime":"2026-01-26T16:48:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:15 crc kubenswrapper[4754]: E0126 16:48:15.347014 4754 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:48:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:48:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:48:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:48:15Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:48:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:48:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:48:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:48:15Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d8506764-e7fa-45cb-a13d-6f527164f548\\\",\\\"systemUUID\\\":\\\"bfbb6c33-ce46-4e81-a8a8-b44409b03821\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:15Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:15 crc kubenswrapper[4754]: I0126 16:48:15.351208 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:15 crc kubenswrapper[4754]: I0126 16:48:15.351248 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:15 crc kubenswrapper[4754]: I0126 16:48:15.351259 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:15 crc kubenswrapper[4754]: I0126 16:48:15.351274 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:15 crc kubenswrapper[4754]: I0126 16:48:15.351286 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:15Z","lastTransitionTime":"2026-01-26T16:48:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:15 crc kubenswrapper[4754]: E0126 16:48:15.364891 4754 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:48:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:48:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:48:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:48:15Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:48:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:48:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:48:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:48:15Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d8506764-e7fa-45cb-a13d-6f527164f548\\\",\\\"systemUUID\\\":\\\"bfbb6c33-ce46-4e81-a8a8-b44409b03821\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:15Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:15 crc kubenswrapper[4754]: I0126 16:48:15.367994 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:15 crc kubenswrapper[4754]: I0126 16:48:15.368023 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:15 crc kubenswrapper[4754]: I0126 16:48:15.368032 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:15 crc kubenswrapper[4754]: I0126 16:48:15.368045 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:15 crc kubenswrapper[4754]: I0126 16:48:15.368054 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:15Z","lastTransitionTime":"2026-01-26T16:48:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:15 crc kubenswrapper[4754]: E0126 16:48:15.379507 4754 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:48:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:48:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:48:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:48:15Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:48:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:48:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:48:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:48:15Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d8506764-e7fa-45cb-a13d-6f527164f548\\\",\\\"systemUUID\\\":\\\"bfbb6c33-ce46-4e81-a8a8-b44409b03821\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:15Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:15 crc kubenswrapper[4754]: E0126 16:48:15.379645 4754 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Jan 26 16:48:15 crc kubenswrapper[4754]: I0126 16:48:15.381004 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:15 crc kubenswrapper[4754]: I0126 16:48:15.381066 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:15 crc kubenswrapper[4754]: I0126 16:48:15.381081 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:15 crc kubenswrapper[4754]: I0126 16:48:15.381095 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:15 crc kubenswrapper[4754]: I0126 16:48:15.381106 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:15Z","lastTransitionTime":"2026-01-26T16:48:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:15 crc kubenswrapper[4754]: I0126 16:48:15.483640 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:15 crc kubenswrapper[4754]: I0126 16:48:15.483888 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:15 crc kubenswrapper[4754]: I0126 16:48:15.483908 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:15 crc kubenswrapper[4754]: I0126 16:48:15.483926 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:15 crc kubenswrapper[4754]: I0126 16:48:15.483939 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:15Z","lastTransitionTime":"2026-01-26T16:48:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:15 crc kubenswrapper[4754]: I0126 16:48:15.586207 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:15 crc kubenswrapper[4754]: I0126 16:48:15.586280 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:15 crc kubenswrapper[4754]: I0126 16:48:15.586295 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:15 crc kubenswrapper[4754]: I0126 16:48:15.586311 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:15 crc kubenswrapper[4754]: I0126 16:48:15.586349 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:15Z","lastTransitionTime":"2026-01-26T16:48:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:15 crc kubenswrapper[4754]: I0126 16:48:15.689260 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:15 crc kubenswrapper[4754]: I0126 16:48:15.689348 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:15 crc kubenswrapper[4754]: I0126 16:48:15.689362 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:15 crc kubenswrapper[4754]: I0126 16:48:15.689377 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:15 crc kubenswrapper[4754]: I0126 16:48:15.689386 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:15Z","lastTransitionTime":"2026-01-26T16:48:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:15 crc kubenswrapper[4754]: I0126 16:48:15.750975 4754 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-30 17:55:42.736486779 +0000 UTC Jan 26 16:48:15 crc kubenswrapper[4754]: I0126 16:48:15.766778 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 16:48:15 crc kubenswrapper[4754]: E0126 16:48:15.766927 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 16:48:15 crc kubenswrapper[4754]: I0126 16:48:15.766791 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 16:48:15 crc kubenswrapper[4754]: E0126 16:48:15.767014 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 16:48:15 crc kubenswrapper[4754]: I0126 16:48:15.766785 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 16:48:15 crc kubenswrapper[4754]: I0126 16:48:15.766812 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-t4ps6" Jan 26 16:48:15 crc kubenswrapper[4754]: E0126 16:48:15.767077 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 16:48:15 crc kubenswrapper[4754]: E0126 16:48:15.767256 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-t4ps6" podUID="d45179f0-9f81-4d77-8805-6b05bdb32f8f" Jan 26 16:48:15 crc kubenswrapper[4754]: I0126 16:48:15.792179 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:15 crc kubenswrapper[4754]: I0126 16:48:15.792233 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:15 crc kubenswrapper[4754]: I0126 16:48:15.792249 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:15 crc kubenswrapper[4754]: I0126 16:48:15.792269 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:15 crc kubenswrapper[4754]: I0126 16:48:15.792285 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:15Z","lastTransitionTime":"2026-01-26T16:48:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:15 crc kubenswrapper[4754]: I0126 16:48:15.894562 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:15 crc kubenswrapper[4754]: I0126 16:48:15.894601 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:15 crc kubenswrapper[4754]: I0126 16:48:15.894610 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:15 crc kubenswrapper[4754]: I0126 16:48:15.894623 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:15 crc kubenswrapper[4754]: I0126 16:48:15.894632 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:15Z","lastTransitionTime":"2026-01-26T16:48:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:15 crc kubenswrapper[4754]: I0126 16:48:15.997206 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:15 crc kubenswrapper[4754]: I0126 16:48:15.997245 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:15 crc kubenswrapper[4754]: I0126 16:48:15.997256 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:15 crc kubenswrapper[4754]: I0126 16:48:15.997273 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:15 crc kubenswrapper[4754]: I0126 16:48:15.997285 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:15Z","lastTransitionTime":"2026-01-26T16:48:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:16 crc kubenswrapper[4754]: I0126 16:48:16.098870 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:16 crc kubenswrapper[4754]: I0126 16:48:16.098914 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:16 crc kubenswrapper[4754]: I0126 16:48:16.098924 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:16 crc kubenswrapper[4754]: I0126 16:48:16.098937 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:16 crc kubenswrapper[4754]: I0126 16:48:16.098946 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:16Z","lastTransitionTime":"2026-01-26T16:48:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:16 crc kubenswrapper[4754]: I0126 16:48:16.201210 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:16 crc kubenswrapper[4754]: I0126 16:48:16.201259 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:16 crc kubenswrapper[4754]: I0126 16:48:16.201270 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:16 crc kubenswrapper[4754]: I0126 16:48:16.201285 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:16 crc kubenswrapper[4754]: I0126 16:48:16.201295 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:16Z","lastTransitionTime":"2026-01-26T16:48:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:16 crc kubenswrapper[4754]: I0126 16:48:16.303446 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:16 crc kubenswrapper[4754]: I0126 16:48:16.303523 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:16 crc kubenswrapper[4754]: I0126 16:48:16.303536 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:16 crc kubenswrapper[4754]: I0126 16:48:16.303551 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:16 crc kubenswrapper[4754]: I0126 16:48:16.303561 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:16Z","lastTransitionTime":"2026-01-26T16:48:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:16 crc kubenswrapper[4754]: I0126 16:48:16.405851 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:16 crc kubenswrapper[4754]: I0126 16:48:16.406023 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:16 crc kubenswrapper[4754]: I0126 16:48:16.406042 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:16 crc kubenswrapper[4754]: I0126 16:48:16.406064 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:16 crc kubenswrapper[4754]: I0126 16:48:16.406082 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:16Z","lastTransitionTime":"2026-01-26T16:48:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:16 crc kubenswrapper[4754]: I0126 16:48:16.507889 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:16 crc kubenswrapper[4754]: I0126 16:48:16.507937 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:16 crc kubenswrapper[4754]: I0126 16:48:16.507948 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:16 crc kubenswrapper[4754]: I0126 16:48:16.507962 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:16 crc kubenswrapper[4754]: I0126 16:48:16.507973 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:16Z","lastTransitionTime":"2026-01-26T16:48:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:16 crc kubenswrapper[4754]: I0126 16:48:16.610333 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:16 crc kubenswrapper[4754]: I0126 16:48:16.610370 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:16 crc kubenswrapper[4754]: I0126 16:48:16.610379 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:16 crc kubenswrapper[4754]: I0126 16:48:16.610391 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:16 crc kubenswrapper[4754]: I0126 16:48:16.610400 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:16Z","lastTransitionTime":"2026-01-26T16:48:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:16 crc kubenswrapper[4754]: I0126 16:48:16.712192 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:16 crc kubenswrapper[4754]: I0126 16:48:16.712232 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:16 crc kubenswrapper[4754]: I0126 16:48:16.712241 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:16 crc kubenswrapper[4754]: I0126 16:48:16.712255 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:16 crc kubenswrapper[4754]: I0126 16:48:16.712263 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:16Z","lastTransitionTime":"2026-01-26T16:48:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:16 crc kubenswrapper[4754]: I0126 16:48:16.751621 4754 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-18 16:48:38.377089226 +0000 UTC Jan 26 16:48:16 crc kubenswrapper[4754]: I0126 16:48:16.814858 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:16 crc kubenswrapper[4754]: I0126 16:48:16.814895 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:16 crc kubenswrapper[4754]: I0126 16:48:16.814905 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:16 crc kubenswrapper[4754]: I0126 16:48:16.814921 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:16 crc kubenswrapper[4754]: I0126 16:48:16.814931 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:16Z","lastTransitionTime":"2026-01-26T16:48:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:16 crc kubenswrapper[4754]: I0126 16:48:16.917452 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:16 crc kubenswrapper[4754]: I0126 16:48:16.917496 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:16 crc kubenswrapper[4754]: I0126 16:48:16.917506 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:16 crc kubenswrapper[4754]: I0126 16:48:16.917522 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:16 crc kubenswrapper[4754]: I0126 16:48:16.917538 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:16Z","lastTransitionTime":"2026-01-26T16:48:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:17 crc kubenswrapper[4754]: I0126 16:48:17.020988 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:17 crc kubenswrapper[4754]: I0126 16:48:17.021042 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:17 crc kubenswrapper[4754]: I0126 16:48:17.021058 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:17 crc kubenswrapper[4754]: I0126 16:48:17.021083 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:17 crc kubenswrapper[4754]: I0126 16:48:17.021105 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:17Z","lastTransitionTime":"2026-01-26T16:48:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:17 crc kubenswrapper[4754]: I0126 16:48:17.124765 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:17 crc kubenswrapper[4754]: I0126 16:48:17.124818 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:17 crc kubenswrapper[4754]: I0126 16:48:17.124827 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:17 crc kubenswrapper[4754]: I0126 16:48:17.124846 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:17 crc kubenswrapper[4754]: I0126 16:48:17.124857 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:17Z","lastTransitionTime":"2026-01-26T16:48:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:17 crc kubenswrapper[4754]: I0126 16:48:17.227222 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:17 crc kubenswrapper[4754]: I0126 16:48:17.227287 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:17 crc kubenswrapper[4754]: I0126 16:48:17.227297 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:17 crc kubenswrapper[4754]: I0126 16:48:17.227320 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:17 crc kubenswrapper[4754]: I0126 16:48:17.227330 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:17Z","lastTransitionTime":"2026-01-26T16:48:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:17 crc kubenswrapper[4754]: I0126 16:48:17.329567 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:17 crc kubenswrapper[4754]: I0126 16:48:17.329601 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:17 crc kubenswrapper[4754]: I0126 16:48:17.329609 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:17 crc kubenswrapper[4754]: I0126 16:48:17.329621 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:17 crc kubenswrapper[4754]: I0126 16:48:17.329631 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:17Z","lastTransitionTime":"2026-01-26T16:48:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:17 crc kubenswrapper[4754]: I0126 16:48:17.432230 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:17 crc kubenswrapper[4754]: I0126 16:48:17.432269 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:17 crc kubenswrapper[4754]: I0126 16:48:17.432278 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:17 crc kubenswrapper[4754]: I0126 16:48:17.432291 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:17 crc kubenswrapper[4754]: I0126 16:48:17.432300 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:17Z","lastTransitionTime":"2026-01-26T16:48:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:17 crc kubenswrapper[4754]: I0126 16:48:17.534722 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:17 crc kubenswrapper[4754]: I0126 16:48:17.534776 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:17 crc kubenswrapper[4754]: I0126 16:48:17.534788 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:17 crc kubenswrapper[4754]: I0126 16:48:17.534804 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:17 crc kubenswrapper[4754]: I0126 16:48:17.534814 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:17Z","lastTransitionTime":"2026-01-26T16:48:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:17 crc kubenswrapper[4754]: I0126 16:48:17.637430 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:17 crc kubenswrapper[4754]: I0126 16:48:17.637462 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:17 crc kubenswrapper[4754]: I0126 16:48:17.637470 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:17 crc kubenswrapper[4754]: I0126 16:48:17.637482 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:17 crc kubenswrapper[4754]: I0126 16:48:17.637491 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:17Z","lastTransitionTime":"2026-01-26T16:48:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:17 crc kubenswrapper[4754]: I0126 16:48:17.740628 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:17 crc kubenswrapper[4754]: I0126 16:48:17.740766 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:17 crc kubenswrapper[4754]: I0126 16:48:17.740781 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:17 crc kubenswrapper[4754]: I0126 16:48:17.740807 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:17 crc kubenswrapper[4754]: I0126 16:48:17.740820 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:17Z","lastTransitionTime":"2026-01-26T16:48:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:17 crc kubenswrapper[4754]: I0126 16:48:17.751943 4754 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-15 07:08:18.881682377 +0000 UTC Jan 26 16:48:17 crc kubenswrapper[4754]: I0126 16:48:17.767317 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 16:48:17 crc kubenswrapper[4754]: I0126 16:48:17.767339 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 16:48:17 crc kubenswrapper[4754]: I0126 16:48:17.767396 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 16:48:17 crc kubenswrapper[4754]: E0126 16:48:17.767657 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 16:48:17 crc kubenswrapper[4754]: E0126 16:48:17.767817 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 16:48:17 crc kubenswrapper[4754]: I0126 16:48:17.767749 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-t4ps6" Jan 26 16:48:17 crc kubenswrapper[4754]: E0126 16:48:17.767891 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 16:48:17 crc kubenswrapper[4754]: E0126 16:48:17.768080 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-t4ps6" podUID="d45179f0-9f81-4d77-8805-6b05bdb32f8f" Jan 26 16:48:17 crc kubenswrapper[4754]: I0126 16:48:17.843768 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:17 crc kubenswrapper[4754]: I0126 16:48:17.843813 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:17 crc kubenswrapper[4754]: I0126 16:48:17.843825 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:17 crc kubenswrapper[4754]: I0126 16:48:17.843841 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:17 crc kubenswrapper[4754]: I0126 16:48:17.843854 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:17Z","lastTransitionTime":"2026-01-26T16:48:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:17 crc kubenswrapper[4754]: I0126 16:48:17.946466 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:17 crc kubenswrapper[4754]: I0126 16:48:17.946510 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:17 crc kubenswrapper[4754]: I0126 16:48:17.946522 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:17 crc kubenswrapper[4754]: I0126 16:48:17.946542 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:17 crc kubenswrapper[4754]: I0126 16:48:17.946554 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:17Z","lastTransitionTime":"2026-01-26T16:48:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:18 crc kubenswrapper[4754]: I0126 16:48:18.048760 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:18 crc kubenswrapper[4754]: I0126 16:48:18.048845 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:18 crc kubenswrapper[4754]: I0126 16:48:18.048858 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:18 crc kubenswrapper[4754]: I0126 16:48:18.048881 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:18 crc kubenswrapper[4754]: I0126 16:48:18.048894 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:18Z","lastTransitionTime":"2026-01-26T16:48:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:18 crc kubenswrapper[4754]: I0126 16:48:18.151374 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:18 crc kubenswrapper[4754]: I0126 16:48:18.151424 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:18 crc kubenswrapper[4754]: I0126 16:48:18.151437 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:18 crc kubenswrapper[4754]: I0126 16:48:18.151453 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:18 crc kubenswrapper[4754]: I0126 16:48:18.151464 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:18Z","lastTransitionTime":"2026-01-26T16:48:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:18 crc kubenswrapper[4754]: I0126 16:48:18.253830 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:18 crc kubenswrapper[4754]: I0126 16:48:18.253876 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:18 crc kubenswrapper[4754]: I0126 16:48:18.253887 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:18 crc kubenswrapper[4754]: I0126 16:48:18.253903 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:18 crc kubenswrapper[4754]: I0126 16:48:18.253916 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:18Z","lastTransitionTime":"2026-01-26T16:48:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:18 crc kubenswrapper[4754]: I0126 16:48:18.356216 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:18 crc kubenswrapper[4754]: I0126 16:48:18.356293 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:18 crc kubenswrapper[4754]: I0126 16:48:18.356305 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:18 crc kubenswrapper[4754]: I0126 16:48:18.356347 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:18 crc kubenswrapper[4754]: I0126 16:48:18.356358 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:18Z","lastTransitionTime":"2026-01-26T16:48:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:18 crc kubenswrapper[4754]: I0126 16:48:18.459813 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:18 crc kubenswrapper[4754]: I0126 16:48:18.459870 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:18 crc kubenswrapper[4754]: I0126 16:48:18.459894 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:18 crc kubenswrapper[4754]: I0126 16:48:18.459912 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:18 crc kubenswrapper[4754]: I0126 16:48:18.459923 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:18Z","lastTransitionTime":"2026-01-26T16:48:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:18 crc kubenswrapper[4754]: I0126 16:48:18.563217 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:18 crc kubenswrapper[4754]: I0126 16:48:18.563256 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:18 crc kubenswrapper[4754]: I0126 16:48:18.563264 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:18 crc kubenswrapper[4754]: I0126 16:48:18.563279 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:18 crc kubenswrapper[4754]: I0126 16:48:18.563290 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:18Z","lastTransitionTime":"2026-01-26T16:48:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:18 crc kubenswrapper[4754]: I0126 16:48:18.666128 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:18 crc kubenswrapper[4754]: I0126 16:48:18.666165 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:18 crc kubenswrapper[4754]: I0126 16:48:18.666178 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:18 crc kubenswrapper[4754]: I0126 16:48:18.666196 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:18 crc kubenswrapper[4754]: I0126 16:48:18.666210 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:18Z","lastTransitionTime":"2026-01-26T16:48:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:18 crc kubenswrapper[4754]: I0126 16:48:18.752940 4754 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-25 05:36:32.97998995 +0000 UTC Jan 26 16:48:18 crc kubenswrapper[4754]: I0126 16:48:18.768199 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:18 crc kubenswrapper[4754]: I0126 16:48:18.768240 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:18 crc kubenswrapper[4754]: I0126 16:48:18.768251 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:18 crc kubenswrapper[4754]: I0126 16:48:18.768265 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:18 crc kubenswrapper[4754]: I0126 16:48:18.768277 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:18Z","lastTransitionTime":"2026-01-26T16:48:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:18 crc kubenswrapper[4754]: I0126 16:48:18.870776 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:18 crc kubenswrapper[4754]: I0126 16:48:18.870893 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:18 crc kubenswrapper[4754]: I0126 16:48:18.870908 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:18 crc kubenswrapper[4754]: I0126 16:48:18.870929 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:18 crc kubenswrapper[4754]: I0126 16:48:18.870937 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:18Z","lastTransitionTime":"2026-01-26T16:48:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:18 crc kubenswrapper[4754]: I0126 16:48:18.973787 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:18 crc kubenswrapper[4754]: I0126 16:48:18.973835 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:18 crc kubenswrapper[4754]: I0126 16:48:18.973846 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:18 crc kubenswrapper[4754]: I0126 16:48:18.973861 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:18 crc kubenswrapper[4754]: I0126 16:48:18.973872 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:18Z","lastTransitionTime":"2026-01-26T16:48:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:19 crc kubenswrapper[4754]: I0126 16:48:19.076217 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:19 crc kubenswrapper[4754]: I0126 16:48:19.076258 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:19 crc kubenswrapper[4754]: I0126 16:48:19.076269 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:19 crc kubenswrapper[4754]: I0126 16:48:19.076286 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:19 crc kubenswrapper[4754]: I0126 16:48:19.076299 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:19Z","lastTransitionTime":"2026-01-26T16:48:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:19 crc kubenswrapper[4754]: I0126 16:48:19.178588 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:19 crc kubenswrapper[4754]: I0126 16:48:19.178631 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:19 crc kubenswrapper[4754]: I0126 16:48:19.178640 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:19 crc kubenswrapper[4754]: I0126 16:48:19.178656 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:19 crc kubenswrapper[4754]: I0126 16:48:19.178681 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:19Z","lastTransitionTime":"2026-01-26T16:48:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:19 crc kubenswrapper[4754]: I0126 16:48:19.281078 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:19 crc kubenswrapper[4754]: I0126 16:48:19.281131 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:19 crc kubenswrapper[4754]: I0126 16:48:19.281141 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:19 crc kubenswrapper[4754]: I0126 16:48:19.281166 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:19 crc kubenswrapper[4754]: I0126 16:48:19.281178 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:19Z","lastTransitionTime":"2026-01-26T16:48:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:19 crc kubenswrapper[4754]: I0126 16:48:19.383978 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:19 crc kubenswrapper[4754]: I0126 16:48:19.384018 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:19 crc kubenswrapper[4754]: I0126 16:48:19.384028 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:19 crc kubenswrapper[4754]: I0126 16:48:19.384043 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:19 crc kubenswrapper[4754]: I0126 16:48:19.384054 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:19Z","lastTransitionTime":"2026-01-26T16:48:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:19 crc kubenswrapper[4754]: I0126 16:48:19.486398 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:19 crc kubenswrapper[4754]: I0126 16:48:19.486471 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:19 crc kubenswrapper[4754]: I0126 16:48:19.486507 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:19 crc kubenswrapper[4754]: I0126 16:48:19.486527 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:19 crc kubenswrapper[4754]: I0126 16:48:19.486538 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:19Z","lastTransitionTime":"2026-01-26T16:48:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:19 crc kubenswrapper[4754]: I0126 16:48:19.589474 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:19 crc kubenswrapper[4754]: I0126 16:48:19.589514 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:19 crc kubenswrapper[4754]: I0126 16:48:19.589524 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:19 crc kubenswrapper[4754]: I0126 16:48:19.589538 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:19 crc kubenswrapper[4754]: I0126 16:48:19.589548 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:19Z","lastTransitionTime":"2026-01-26T16:48:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:19 crc kubenswrapper[4754]: I0126 16:48:19.692206 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:19 crc kubenswrapper[4754]: I0126 16:48:19.692250 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:19 crc kubenswrapper[4754]: I0126 16:48:19.692259 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:19 crc kubenswrapper[4754]: I0126 16:48:19.692274 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:19 crc kubenswrapper[4754]: I0126 16:48:19.692287 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:19Z","lastTransitionTime":"2026-01-26T16:48:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:19 crc kubenswrapper[4754]: I0126 16:48:19.754100 4754 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-26 02:12:30.00024549 +0000 UTC Jan 26 16:48:19 crc kubenswrapper[4754]: I0126 16:48:19.766365 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 16:48:19 crc kubenswrapper[4754]: I0126 16:48:19.766439 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 16:48:19 crc kubenswrapper[4754]: I0126 16:48:19.766490 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-t4ps6" Jan 26 16:48:19 crc kubenswrapper[4754]: E0126 16:48:19.766502 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 16:48:19 crc kubenswrapper[4754]: I0126 16:48:19.766545 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 16:48:19 crc kubenswrapper[4754]: E0126 16:48:19.766644 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 16:48:19 crc kubenswrapper[4754]: E0126 16:48:19.766846 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 16:48:19 crc kubenswrapper[4754]: E0126 16:48:19.766966 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-t4ps6" podUID="d45179f0-9f81-4d77-8805-6b05bdb32f8f" Jan 26 16:48:19 crc kubenswrapper[4754]: I0126 16:48:19.794962 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:19 crc kubenswrapper[4754]: I0126 16:48:19.795022 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:19 crc kubenswrapper[4754]: I0126 16:48:19.795038 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:19 crc kubenswrapper[4754]: I0126 16:48:19.795060 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:19 crc kubenswrapper[4754]: I0126 16:48:19.795077 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:19Z","lastTransitionTime":"2026-01-26T16:48:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:19 crc kubenswrapper[4754]: I0126 16:48:19.897764 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:19 crc kubenswrapper[4754]: I0126 16:48:19.897808 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:19 crc kubenswrapper[4754]: I0126 16:48:19.897817 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:19 crc kubenswrapper[4754]: I0126 16:48:19.897830 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:19 crc kubenswrapper[4754]: I0126 16:48:19.897841 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:19Z","lastTransitionTime":"2026-01-26T16:48:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:20 crc kubenswrapper[4754]: I0126 16:48:20.000336 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:20 crc kubenswrapper[4754]: I0126 16:48:20.000386 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:20 crc kubenswrapper[4754]: I0126 16:48:20.000398 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:20 crc kubenswrapper[4754]: I0126 16:48:20.000415 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:20 crc kubenswrapper[4754]: I0126 16:48:20.000427 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:20Z","lastTransitionTime":"2026-01-26T16:48:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:20 crc kubenswrapper[4754]: I0126 16:48:20.102498 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:20 crc kubenswrapper[4754]: I0126 16:48:20.102540 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:20 crc kubenswrapper[4754]: I0126 16:48:20.102549 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:20 crc kubenswrapper[4754]: I0126 16:48:20.102564 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:20 crc kubenswrapper[4754]: I0126 16:48:20.102573 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:20Z","lastTransitionTime":"2026-01-26T16:48:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:20 crc kubenswrapper[4754]: I0126 16:48:20.205227 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:20 crc kubenswrapper[4754]: I0126 16:48:20.205271 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:20 crc kubenswrapper[4754]: I0126 16:48:20.205280 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:20 crc kubenswrapper[4754]: I0126 16:48:20.205331 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:20 crc kubenswrapper[4754]: I0126 16:48:20.205344 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:20Z","lastTransitionTime":"2026-01-26T16:48:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:20 crc kubenswrapper[4754]: I0126 16:48:20.307780 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:20 crc kubenswrapper[4754]: I0126 16:48:20.307833 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:20 crc kubenswrapper[4754]: I0126 16:48:20.307845 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:20 crc kubenswrapper[4754]: I0126 16:48:20.307863 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:20 crc kubenswrapper[4754]: I0126 16:48:20.307876 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:20Z","lastTransitionTime":"2026-01-26T16:48:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:20 crc kubenswrapper[4754]: I0126 16:48:20.339390 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d45179f0-9f81-4d77-8805-6b05bdb32f8f-metrics-certs\") pod \"network-metrics-daemon-t4ps6\" (UID: \"d45179f0-9f81-4d77-8805-6b05bdb32f8f\") " pod="openshift-multus/network-metrics-daemon-t4ps6" Jan 26 16:48:20 crc kubenswrapper[4754]: E0126 16:48:20.339557 4754 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Jan 26 16:48:20 crc kubenswrapper[4754]: E0126 16:48:20.339640 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d45179f0-9f81-4d77-8805-6b05bdb32f8f-metrics-certs podName:d45179f0-9f81-4d77-8805-6b05bdb32f8f nodeName:}" failed. No retries permitted until 2026-01-26 16:48:52.339611658 +0000 UTC m=+98.863792092 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/d45179f0-9f81-4d77-8805-6b05bdb32f8f-metrics-certs") pod "network-metrics-daemon-t4ps6" (UID: "d45179f0-9f81-4d77-8805-6b05bdb32f8f") : object "openshift-multus"/"metrics-daemon-secret" not registered Jan 26 16:48:20 crc kubenswrapper[4754]: I0126 16:48:20.410951 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:20 crc kubenswrapper[4754]: I0126 16:48:20.410986 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:20 crc kubenswrapper[4754]: I0126 16:48:20.410997 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:20 crc kubenswrapper[4754]: I0126 16:48:20.411026 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:20 crc kubenswrapper[4754]: I0126 16:48:20.411036 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:20Z","lastTransitionTime":"2026-01-26T16:48:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:20 crc kubenswrapper[4754]: I0126 16:48:20.512650 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:20 crc kubenswrapper[4754]: I0126 16:48:20.512709 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:20 crc kubenswrapper[4754]: I0126 16:48:20.512723 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:20 crc kubenswrapper[4754]: I0126 16:48:20.512740 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:20 crc kubenswrapper[4754]: I0126 16:48:20.512752 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:20Z","lastTransitionTime":"2026-01-26T16:48:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:20 crc kubenswrapper[4754]: I0126 16:48:20.614972 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:20 crc kubenswrapper[4754]: I0126 16:48:20.615025 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:20 crc kubenswrapper[4754]: I0126 16:48:20.615037 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:20 crc kubenswrapper[4754]: I0126 16:48:20.615051 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:20 crc kubenswrapper[4754]: I0126 16:48:20.615062 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:20Z","lastTransitionTime":"2026-01-26T16:48:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:20 crc kubenswrapper[4754]: I0126 16:48:20.718281 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:20 crc kubenswrapper[4754]: I0126 16:48:20.718319 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:20 crc kubenswrapper[4754]: I0126 16:48:20.718329 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:20 crc kubenswrapper[4754]: I0126 16:48:20.718345 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:20 crc kubenswrapper[4754]: I0126 16:48:20.718356 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:20Z","lastTransitionTime":"2026-01-26T16:48:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:20 crc kubenswrapper[4754]: I0126 16:48:20.820700 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:20 crc kubenswrapper[4754]: I0126 16:48:20.820736 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:20 crc kubenswrapper[4754]: I0126 16:48:20.820752 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:20 crc kubenswrapper[4754]: I0126 16:48:20.820774 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:20 crc kubenswrapper[4754]: I0126 16:48:20.820787 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:20Z","lastTransitionTime":"2026-01-26T16:48:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:20 crc kubenswrapper[4754]: I0126 16:48:20.923262 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:20 crc kubenswrapper[4754]: I0126 16:48:20.923293 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:20 crc kubenswrapper[4754]: I0126 16:48:20.923302 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:20 crc kubenswrapper[4754]: I0126 16:48:20.923318 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:20 crc kubenswrapper[4754]: I0126 16:48:20.923328 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:20Z","lastTransitionTime":"2026-01-26T16:48:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:21 crc kubenswrapper[4754]: I0126 16:48:21.025433 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:21 crc kubenswrapper[4754]: I0126 16:48:21.025470 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:21 crc kubenswrapper[4754]: I0126 16:48:21.025481 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:21 crc kubenswrapper[4754]: I0126 16:48:21.025498 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:21 crc kubenswrapper[4754]: I0126 16:48:21.025510 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:21Z","lastTransitionTime":"2026-01-26T16:48:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:21 crc kubenswrapper[4754]: I0126 16:48:21.127685 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:21 crc kubenswrapper[4754]: I0126 16:48:21.127713 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:21 crc kubenswrapper[4754]: I0126 16:48:21.127723 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:21 crc kubenswrapper[4754]: I0126 16:48:21.127738 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:21 crc kubenswrapper[4754]: I0126 16:48:21.127749 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:21Z","lastTransitionTime":"2026-01-26T16:48:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:21 crc kubenswrapper[4754]: I0126 16:48:21.230466 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:21 crc kubenswrapper[4754]: I0126 16:48:21.230498 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:21 crc kubenswrapper[4754]: I0126 16:48:21.230509 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:21 crc kubenswrapper[4754]: I0126 16:48:21.230524 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:21 crc kubenswrapper[4754]: I0126 16:48:21.230535 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:21Z","lastTransitionTime":"2026-01-26T16:48:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:21 crc kubenswrapper[4754]: I0126 16:48:21.283120 4754 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-11 22:25:33.244740203 +0000 UTC Jan 26 16:48:21 crc kubenswrapper[4754]: I0126 16:48:21.287955 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-skh5n_b619d233-b592-4b05-a0b7-dc094c88471e/kube-multus/0.log" Jan 26 16:48:21 crc kubenswrapper[4754]: I0126 16:48:21.288081 4754 generic.go:334] "Generic (PLEG): container finished" podID="b619d233-b592-4b05-a0b7-dc094c88471e" containerID="9231a4191565e2e0ef53d8a394bb34615c900b601a0dfec24c2d8e5fa7b0b659" exitCode=1 Jan 26 16:48:21 crc kubenswrapper[4754]: I0126 16:48:21.288150 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-skh5n" event={"ID":"b619d233-b592-4b05-a0b7-dc094c88471e","Type":"ContainerDied","Data":"9231a4191565e2e0ef53d8a394bb34615c900b601a0dfec24c2d8e5fa7b0b659"} Jan 26 16:48:21 crc kubenswrapper[4754]: I0126 16:48:21.288529 4754 scope.go:117] "RemoveContainer" containerID="9231a4191565e2e0ef53d8a394bb34615c900b601a0dfec24c2d8e5fa7b0b659" Jan 26 16:48:21 crc kubenswrapper[4754]: I0126 16:48:21.304219 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:21Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:21 crc kubenswrapper[4754]: I0126 16:48:21.318979 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e459fd8c1093d6f3d0a671751b792b26b7df94c21ced1be1a2250821bcc82d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:21Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:21 crc kubenswrapper[4754]: I0126 16:48:21.330540 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2blzw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca65468e-30c6-4666-962e-cc0de05e67fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4603235bb2af495901926b805fe1c0ffc48824a4394ccf928e7fcd53a673ac3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c25z4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:33Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2blzw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:21Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:21 crc kubenswrapper[4754]: I0126 16:48:21.332315 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:21 crc kubenswrapper[4754]: I0126 16:48:21.332432 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:21 crc kubenswrapper[4754]: I0126 16:48:21.332495 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:21 crc kubenswrapper[4754]: I0126 16:48:21.332575 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:21 crc kubenswrapper[4754]: I0126 16:48:21.332647 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:21Z","lastTransitionTime":"2026-01-26T16:48:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:21 crc kubenswrapper[4754]: I0126 16:48:21.346371 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bwpd6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"351cc6c0-b8e5-4589-ae63-f98e8ab1eca3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1022f101d36873b41bbc1ed8a4a2590347b49f7cce6a172aa1d5bc9f529aed65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3ce0fe080a094316b9456057e347b90a5ff432daa12b9c56d7aea9bee1f72b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3ce0fe080a094316b9456057e347b90a5ff432daa12b9c56d7aea9bee1f72b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1772bb41eb8a36a863d3c691610bcbee5d16b260480b3e9f9d6b083e09e42799\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1772bb41eb8a36a863d3c691610bcbee5d16b260480b3e9f9d6b083e09e42799\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24848d69756024cc664affb9492764ace0638ef611f910f4f2bf86b638cc443d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://24848d69756024cc664affb9492764ace0638ef611f910f4f2bf86b638cc443d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c264fc7c3300a725566acce4464f01d9314779737624401c164d867831fd93bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c264fc7c3300a725566acce4464f01d9314779737624401c164d867831fd93bf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5acfe327eee4aad927ac3a32d24549eff788b14a1c4bdac359b6b56301d884ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5acfe327eee4aad927ac3a32d24549eff788b14a1c4bdac359b6b56301d884ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://63a9b9317494d0e0b33970ef3d461729b10d2a1ad7d3d2bd33f389b184ca56ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://63a9b9317494d0e0b33970ef3d461729b10d2a1ad7d3d2bd33f389b184ca56ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bwpd6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:21Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:21 crc kubenswrapper[4754]: I0126 16:48:21.363204 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jsbxt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"689569c4-93ee-4b82-93fc-9af0f97e6dc3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://033f8b104d4e9ff6d5f87ca998665f12e0c56696b3d30e01ede05f4da540365d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://919572e731813a98c7d8a94879bba931f75c472c7cd681e86d5197ebab43815d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1ac1f49bd564750fcff530a0f76657b0ed41573539acd020a922d19861ef834\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a300a80ad87f46ba1b4fd7ad3cf6c25767fcfe966b93bdd86d729843f53a5cda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31dcab2f9a2d218b0e49dad1df46377ff4521592824a0fbac254d5b8e6d0632e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fe4d9395447ac570eca1338e749dbefea76295d427a12487b63a142a7d6a2e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e9d45c28e67d857b1c50ed78197ed052c1ac2b98f1c8a94addd63698bf0bb23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0e9d45c28e67d857b1c50ed78197ed052c1ac2b98f1c8a94addd63698bf0bb23\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-26T16:47:58Z\\\",\\\"message\\\":\\\":303] Retry object setup: *v1.Pod openshift-machine-config-operator/machine-config-daemon-x65wv\\\\nI0126 16:47:58.625627 6452 obj_retry.go:303] Retry object setup: *v1.Pod openshift-multus/multus-skh5n\\\\nI0126 16:47:58.625718 6452 obj_retry.go:365] Adding new object: *v1.Pod openshift-machine-config-operator/machine-config-daemon-x65wv\\\\nI0126 16:47:58.625617 6452 obj_retry.go:303] Retry object setup: *v1.Pod openshift-multus/multus-additional-cni-plugins-bwpd6\\\\nI0126 16:47:58.625726 6452 ovn.go:134] Ensuring zone local for Pod openshift-machine-config-operator/machine-config-daemon-x65wv in node crc\\\\nI0126 16:47:58.625727 6452 obj_retry.go:365] Adding new object: *v1.Pod openshift-multus/multus-skh5n\\\\nI0126 16:47:58.625732 6452 obj_retry.go:365] Adding new object: *v1.Pod openshift-multus/multus-additional-cni-plugins-bwpd6\\\\nI0126 16:47:58.625706 6452 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {eb8eef51-1a8d-43f9-ae2e-3b2cc00ded60}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0126 16:47:58.625740 6452 ovn.go:134] Ensuring zone local for Pod openshift-multus/multus-additional-cni-plugins-bwpd6 in no\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:57Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-jsbxt_openshift-ovn-kubernetes(689569c4-93ee-4b82-93fc-9af0f97e6dc3)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://07a43ff7860ff65faf51f4a5e8aeb326f06d6d7222677a92560ae2e3a0c81484\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d9a7bafabd44f8b88a6f62cbe45599af8563bc45d9d7ad3a7669f3d8d11eb9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d9a7bafabd44f8b88a6f62cbe45599af8563bc45d9d7ad3a7669f3d8d11eb9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jsbxt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:21Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:21 crc kubenswrapper[4754]: I0126 16:48:21.374279 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"099af702-abf2-4933-b4c9-63ef2326e468\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23017e73f8450839b4c2c02fa61235ffe4604bdce11556def27f25bcadcf89dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a74507c941b9de4d0401c8f1d53bb79ed11359628d3405cdab7f5474a75c862e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4402182b3e2e35a7d5dd4c44ac53e152915b312b7ccb076c4f2b41b925a6e7f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f772a003566198519fe1849fd14f29d4346004b2e655810662507c95b4d5ffa1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:13Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:21Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:21 crc kubenswrapper[4754]: I0126 16:48:21.385126 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab1ace4c-137b-4b8a-8032-0b86a164f063\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:48:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:48:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7da43edaa2a81f1b647b96c234d795b1ddf5fa23173b78043123f76dc677c517\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e5fbff42fdebad6a74c21d85931d3c8d983973db82fc7f262ba5383bad76172\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f9bf9e4a705cd849595e59e1123ba70773bfa5f8a0d0b79195e4fc7e447ca26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb7a3b85cca4436068c8ceefc3f2358cfa1a2c398a8acecd367b944528b11489\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb7a3b85cca4436068c8ceefc3f2358cfa1a2c398a8acecd367b944528b11489\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:14Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:13Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:21Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:21 crc kubenswrapper[4754]: I0126 16:48:21.398117 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:21Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:21 crc kubenswrapper[4754]: I0126 16:48:21.413961 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89251fa9c64bc25d6d6f355400d7644c099f22aed1cae812fc991a03595f1e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e90508d8a53be9146f9b388750730fdae9ee200306caf5ad393a356390a69ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:21Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:21 crc kubenswrapper[4754]: I0126 16:48:21.426641 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-84lgq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6763d962-0892-4acc-bab9-7a4e93e39a85\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://953c1917fca896c3def3fc2c69f9e3dbdc9ca26fc2b3d40282be71f71759fd68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8zlw6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:33Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-84lgq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:21Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:21 crc kubenswrapper[4754]: I0126 16:48:21.435343 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:21 crc kubenswrapper[4754]: I0126 16:48:21.435372 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:21 crc kubenswrapper[4754]: I0126 16:48:21.435382 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:21 crc kubenswrapper[4754]: I0126 16:48:21.435397 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:21 crc kubenswrapper[4754]: I0126 16:48:21.435407 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:21Z","lastTransitionTime":"2026-01-26T16:48:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:21 crc kubenswrapper[4754]: I0126 16:48:21.443056 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-skh5n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b619d233-b592-4b05-a0b7-dc094c88471e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:48:21Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:48:21Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9231a4191565e2e0ef53d8a394bb34615c900b601a0dfec24c2d8e5fa7b0b659\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9231a4191565e2e0ef53d8a394bb34615c900b601a0dfec24c2d8e5fa7b0b659\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-26T16:48:20Z\\\",\\\"message\\\":\\\"2026-01-26T16:47:34+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_57895685-a7ab-4a6c-9a6a-49247f349eaf\\\\n2026-01-26T16:47:34+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_57895685-a7ab-4a6c-9a6a-49247f349eaf to /host/opt/cni/bin/\\\\n2026-01-26T16:47:35Z [verbose] multus-daemon started\\\\n2026-01-26T16:47:35Z [verbose] Readiness Indicator file check\\\\n2026-01-26T16:48:20Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-skh5n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:21Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:21 crc kubenswrapper[4754]: I0126 16:48:21.457155 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jspvh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5c81aab0-939e-45ce-b19a-89d4818980c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://356dc54150a495dd6f35e0667c82b807b712b07200d8a0adbb1c92fe3f7af8f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8bps\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://05fd57780c44cd127859dc8b163d20eafbd8b06f3348cf545ef41021ce6bc526\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8bps\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-jspvh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:21Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:21 crc kubenswrapper[4754]: I0126 16:48:21.473128 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b89aa081-0a2e-4757-9210-e2a6c228bee9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9c5f40188fda14b14128514fe84bddfe8dcede3d6c560283ccb2ac0c66895fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f45445cbf1d180d6d4dd9588e85bf678f043bfce8cff0812b78ca3510904f9a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://085fa7b8d4fef8b2722fac8c346107699b5a6bd33fbfc74ed35831cb8fdf9dad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c40e018a0979035827963efb9efc48a878a55b9df68ec1a9242d7cff64fe480\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8356f76fbc3a4e636c12d1da5455d3f6767d9b83ce2f2925a11435d7d7261e9f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0126 16:47:26.205208 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0126 16:47:26.206538 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3645934156/tls.crt::/tmp/serving-cert-3645934156/tls.key\\\\\\\"\\\\nI0126 16:47:31.504518 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0126 16:47:31.509366 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0126 16:47:31.509400 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0126 16:47:31.509428 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0126 16:47:31.509436 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0126 16:47:31.515912 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0126 16:47:31.516036 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 16:47:31.516068 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 16:47:31.516096 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0126 16:47:31.516121 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0126 16:47:31.516145 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0126 16:47:31.516170 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0126 16:47:31.516412 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0126 16:47:31.519294 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27b74e2e65a547e0e28fddd4f15feacb17f4d57596eeb8da82df9d39243c85e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22ba7649a1308815bfad52016000987a7a9a5bc214d7ff61e137ba099b20fc6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://22ba7649a1308815bfad52016000987a7a9a5bc214d7ff61e137ba099b20fc6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:13Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:21Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:21 crc kubenswrapper[4754]: I0126 16:48:21.486800 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://205ec256e46a8e08b4b0b0805fb6c5f4966fc3ccef0738df4b97cb404e74da17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:21Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:21 crc kubenswrapper[4754]: I0126 16:48:21.499470 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:21Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:21 crc kubenswrapper[4754]: I0126 16:48:21.509828 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c3718a4-f354-4284-92e0-fdfb45a692bd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b86f46f49163bca7eca9458279029af330dbd4890b03325e06fa3b1bdf030de6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxjcz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fffc03b536bfee3c7bc1bddd500e834e34b391896477a1501e3e06201a19374\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxjcz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-x65wv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:21Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:21 crc kubenswrapper[4754]: I0126 16:48:21.522938 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-t4ps6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d45179f0-9f81-4d77-8805-6b05bdb32f8f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vk8tg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vk8tg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:48Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-t4ps6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:21Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:21 crc kubenswrapper[4754]: I0126 16:48:21.537819 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:21 crc kubenswrapper[4754]: I0126 16:48:21.537860 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:21 crc kubenswrapper[4754]: I0126 16:48:21.537871 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:21 crc kubenswrapper[4754]: I0126 16:48:21.537976 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:21 crc kubenswrapper[4754]: I0126 16:48:21.538042 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:21Z","lastTransitionTime":"2026-01-26T16:48:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:21 crc kubenswrapper[4754]: I0126 16:48:21.640164 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:21 crc kubenswrapper[4754]: I0126 16:48:21.640383 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:21 crc kubenswrapper[4754]: I0126 16:48:21.640453 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:21 crc kubenswrapper[4754]: I0126 16:48:21.640556 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:21 crc kubenswrapper[4754]: I0126 16:48:21.640638 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:21Z","lastTransitionTime":"2026-01-26T16:48:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:21 crc kubenswrapper[4754]: I0126 16:48:21.743115 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:21 crc kubenswrapper[4754]: I0126 16:48:21.743280 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:21 crc kubenswrapper[4754]: I0126 16:48:21.743397 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:21 crc kubenswrapper[4754]: I0126 16:48:21.743492 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:21 crc kubenswrapper[4754]: I0126 16:48:21.743578 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:21Z","lastTransitionTime":"2026-01-26T16:48:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:21 crc kubenswrapper[4754]: I0126 16:48:21.766531 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 16:48:21 crc kubenswrapper[4754]: I0126 16:48:21.766595 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 16:48:21 crc kubenswrapper[4754]: I0126 16:48:21.766645 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 16:48:21 crc kubenswrapper[4754]: E0126 16:48:21.766704 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 16:48:21 crc kubenswrapper[4754]: I0126 16:48:21.766567 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-t4ps6" Jan 26 16:48:21 crc kubenswrapper[4754]: E0126 16:48:21.766748 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 16:48:21 crc kubenswrapper[4754]: E0126 16:48:21.766835 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-t4ps6" podUID="d45179f0-9f81-4d77-8805-6b05bdb32f8f" Jan 26 16:48:21 crc kubenswrapper[4754]: E0126 16:48:21.766936 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 16:48:21 crc kubenswrapper[4754]: I0126 16:48:21.845979 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:21 crc kubenswrapper[4754]: I0126 16:48:21.846026 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:21 crc kubenswrapper[4754]: I0126 16:48:21.846038 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:21 crc kubenswrapper[4754]: I0126 16:48:21.846053 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:21 crc kubenswrapper[4754]: I0126 16:48:21.846065 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:21Z","lastTransitionTime":"2026-01-26T16:48:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:21 crc kubenswrapper[4754]: I0126 16:48:21.948433 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:21 crc kubenswrapper[4754]: I0126 16:48:21.948490 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:21 crc kubenswrapper[4754]: I0126 16:48:21.948505 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:21 crc kubenswrapper[4754]: I0126 16:48:21.948526 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:21 crc kubenswrapper[4754]: I0126 16:48:21.948539 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:21Z","lastTransitionTime":"2026-01-26T16:48:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:22 crc kubenswrapper[4754]: I0126 16:48:22.051132 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:22 crc kubenswrapper[4754]: I0126 16:48:22.051186 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:22 crc kubenswrapper[4754]: I0126 16:48:22.051198 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:22 crc kubenswrapper[4754]: I0126 16:48:22.051215 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:22 crc kubenswrapper[4754]: I0126 16:48:22.051226 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:22Z","lastTransitionTime":"2026-01-26T16:48:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:22 crc kubenswrapper[4754]: I0126 16:48:22.153173 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:22 crc kubenswrapper[4754]: I0126 16:48:22.153460 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:22 crc kubenswrapper[4754]: I0126 16:48:22.153540 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:22 crc kubenswrapper[4754]: I0126 16:48:22.153647 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:22 crc kubenswrapper[4754]: I0126 16:48:22.153780 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:22Z","lastTransitionTime":"2026-01-26T16:48:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:22 crc kubenswrapper[4754]: I0126 16:48:22.256445 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:22 crc kubenswrapper[4754]: I0126 16:48:22.256499 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:22 crc kubenswrapper[4754]: I0126 16:48:22.256518 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:22 crc kubenswrapper[4754]: I0126 16:48:22.256540 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:22 crc kubenswrapper[4754]: I0126 16:48:22.256551 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:22Z","lastTransitionTime":"2026-01-26T16:48:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:22 crc kubenswrapper[4754]: I0126 16:48:22.283505 4754 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-25 03:22:31.781409248 +0000 UTC Jan 26 16:48:22 crc kubenswrapper[4754]: I0126 16:48:22.293860 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-skh5n_b619d233-b592-4b05-a0b7-dc094c88471e/kube-multus/0.log" Jan 26 16:48:22 crc kubenswrapper[4754]: I0126 16:48:22.293921 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-skh5n" event={"ID":"b619d233-b592-4b05-a0b7-dc094c88471e","Type":"ContainerStarted","Data":"c8d026fdb7ea7dc5c40add18606d8f1a7703865edf93ef19edaf811f598364c5"} Jan 26 16:48:22 crc kubenswrapper[4754]: I0126 16:48:22.311083 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-skh5n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b619d233-b592-4b05-a0b7-dc094c88471e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:48:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:48:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8d026fdb7ea7dc5c40add18606d8f1a7703865edf93ef19edaf811f598364c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9231a4191565e2e0ef53d8a394bb34615c900b601a0dfec24c2d8e5fa7b0b659\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-26T16:48:20Z\\\",\\\"message\\\":\\\"2026-01-26T16:47:34+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_57895685-a7ab-4a6c-9a6a-49247f349eaf\\\\n2026-01-26T16:47:34+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_57895685-a7ab-4a6c-9a6a-49247f349eaf to /host/opt/cni/bin/\\\\n2026-01-26T16:47:35Z [verbose] multus-daemon started\\\\n2026-01-26T16:47:35Z [verbose] Readiness Indicator file check\\\\n2026-01-26T16:48:20Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:48:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-skh5n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:22Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:22 crc kubenswrapper[4754]: I0126 16:48:22.327374 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jspvh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5c81aab0-939e-45ce-b19a-89d4818980c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://356dc54150a495dd6f35e0667c82b807b712b07200d8a0adbb1c92fe3f7af8f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8bps\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://05fd57780c44cd127859dc8b163d20eafbd8b06f3348cf545ef41021ce6bc526\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8bps\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-jspvh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:22Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:22 crc kubenswrapper[4754]: I0126 16:48:22.343493 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"099af702-abf2-4933-b4c9-63ef2326e468\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23017e73f8450839b4c2c02fa61235ffe4604bdce11556def27f25bcadcf89dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a74507c941b9de4d0401c8f1d53bb79ed11359628d3405cdab7f5474a75c862e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4402182b3e2e35a7d5dd4c44ac53e152915b312b7ccb076c4f2b41b925a6e7f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f772a003566198519fe1849fd14f29d4346004b2e655810662507c95b4d5ffa1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:13Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:22Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:22 crc kubenswrapper[4754]: I0126 16:48:22.359231 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab1ace4c-137b-4b8a-8032-0b86a164f063\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:48:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:48:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7da43edaa2a81f1b647b96c234d795b1ddf5fa23173b78043123f76dc677c517\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e5fbff42fdebad6a74c21d85931d3c8d983973db82fc7f262ba5383bad76172\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f9bf9e4a705cd849595e59e1123ba70773bfa5f8a0d0b79195e4fc7e447ca26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb7a3b85cca4436068c8ceefc3f2358cfa1a2c398a8acecd367b944528b11489\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb7a3b85cca4436068c8ceefc3f2358cfa1a2c398a8acecd367b944528b11489\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:14Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:13Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:22Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:22 crc kubenswrapper[4754]: I0126 16:48:22.359575 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:22 crc kubenswrapper[4754]: I0126 16:48:22.359620 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:22 crc kubenswrapper[4754]: I0126 16:48:22.359636 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:22 crc kubenswrapper[4754]: I0126 16:48:22.359657 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:22 crc kubenswrapper[4754]: I0126 16:48:22.359688 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:22Z","lastTransitionTime":"2026-01-26T16:48:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:22 crc kubenswrapper[4754]: I0126 16:48:22.372889 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:22Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:22 crc kubenswrapper[4754]: I0126 16:48:22.385942 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89251fa9c64bc25d6d6f355400d7644c099f22aed1cae812fc991a03595f1e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e90508d8a53be9146f9b388750730fdae9ee200306caf5ad393a356390a69ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:22Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:22 crc kubenswrapper[4754]: I0126 16:48:22.395908 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-84lgq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6763d962-0892-4acc-bab9-7a4e93e39a85\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://953c1917fca896c3def3fc2c69f9e3dbdc9ca26fc2b3d40282be71f71759fd68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8zlw6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:33Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-84lgq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:22Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:22 crc kubenswrapper[4754]: I0126 16:48:22.410744 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b89aa081-0a2e-4757-9210-e2a6c228bee9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9c5f40188fda14b14128514fe84bddfe8dcede3d6c560283ccb2ac0c66895fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f45445cbf1d180d6d4dd9588e85bf678f043bfce8cff0812b78ca3510904f9a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://085fa7b8d4fef8b2722fac8c346107699b5a6bd33fbfc74ed35831cb8fdf9dad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c40e018a0979035827963efb9efc48a878a55b9df68ec1a9242d7cff64fe480\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8356f76fbc3a4e636c12d1da5455d3f6767d9b83ce2f2925a11435d7d7261e9f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0126 16:47:26.205208 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0126 16:47:26.206538 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3645934156/tls.crt::/tmp/serving-cert-3645934156/tls.key\\\\\\\"\\\\nI0126 16:47:31.504518 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0126 16:47:31.509366 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0126 16:47:31.509400 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0126 16:47:31.509428 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0126 16:47:31.509436 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0126 16:47:31.515912 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0126 16:47:31.516036 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 16:47:31.516068 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 16:47:31.516096 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0126 16:47:31.516121 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0126 16:47:31.516145 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0126 16:47:31.516170 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0126 16:47:31.516412 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0126 16:47:31.519294 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27b74e2e65a547e0e28fddd4f15feacb17f4d57596eeb8da82df9d39243c85e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22ba7649a1308815bfad52016000987a7a9a5bc214d7ff61e137ba099b20fc6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://22ba7649a1308815bfad52016000987a7a9a5bc214d7ff61e137ba099b20fc6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:13Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:22Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:22 crc kubenswrapper[4754]: I0126 16:48:22.424990 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://205ec256e46a8e08b4b0b0805fb6c5f4966fc3ccef0738df4b97cb404e74da17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:22Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:22 crc kubenswrapper[4754]: I0126 16:48:22.437618 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:22Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:22 crc kubenswrapper[4754]: I0126 16:48:22.450015 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c3718a4-f354-4284-92e0-fdfb45a692bd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b86f46f49163bca7eca9458279029af330dbd4890b03325e06fa3b1bdf030de6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxjcz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fffc03b536bfee3c7bc1bddd500e834e34b391896477a1501e3e06201a19374\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxjcz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-x65wv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:22Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:22 crc kubenswrapper[4754]: I0126 16:48:22.461623 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-t4ps6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d45179f0-9f81-4d77-8805-6b05bdb32f8f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vk8tg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vk8tg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:48Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-t4ps6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:22Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:22 crc kubenswrapper[4754]: I0126 16:48:22.462131 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:22 crc kubenswrapper[4754]: I0126 16:48:22.462163 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:22 crc kubenswrapper[4754]: I0126 16:48:22.462175 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:22 crc kubenswrapper[4754]: I0126 16:48:22.462193 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:22 crc kubenswrapper[4754]: I0126 16:48:22.462205 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:22Z","lastTransitionTime":"2026-01-26T16:48:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:22 crc kubenswrapper[4754]: I0126 16:48:22.480460 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jsbxt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"689569c4-93ee-4b82-93fc-9af0f97e6dc3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://033f8b104d4e9ff6d5f87ca998665f12e0c56696b3d30e01ede05f4da540365d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://919572e731813a98c7d8a94879bba931f75c472c7cd681e86d5197ebab43815d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1ac1f49bd564750fcff530a0f76657b0ed41573539acd020a922d19861ef834\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a300a80ad87f46ba1b4fd7ad3cf6c25767fcfe966b93bdd86d729843f53a5cda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31dcab2f9a2d218b0e49dad1df46377ff4521592824a0fbac254d5b8e6d0632e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fe4d9395447ac570eca1338e749dbefea76295d427a12487b63a142a7d6a2e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e9d45c28e67d857b1c50ed78197ed052c1ac2b98f1c8a94addd63698bf0bb23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0e9d45c28e67d857b1c50ed78197ed052c1ac2b98f1c8a94addd63698bf0bb23\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-26T16:47:58Z\\\",\\\"message\\\":\\\":303] Retry object setup: *v1.Pod openshift-machine-config-operator/machine-config-daemon-x65wv\\\\nI0126 16:47:58.625627 6452 obj_retry.go:303] Retry object setup: *v1.Pod openshift-multus/multus-skh5n\\\\nI0126 16:47:58.625718 6452 obj_retry.go:365] Adding new object: *v1.Pod openshift-machine-config-operator/machine-config-daemon-x65wv\\\\nI0126 16:47:58.625617 6452 obj_retry.go:303] Retry object setup: *v1.Pod openshift-multus/multus-additional-cni-plugins-bwpd6\\\\nI0126 16:47:58.625726 6452 ovn.go:134] Ensuring zone local for Pod openshift-machine-config-operator/machine-config-daemon-x65wv in node crc\\\\nI0126 16:47:58.625727 6452 obj_retry.go:365] Adding new object: *v1.Pod openshift-multus/multus-skh5n\\\\nI0126 16:47:58.625732 6452 obj_retry.go:365] Adding new object: *v1.Pod openshift-multus/multus-additional-cni-plugins-bwpd6\\\\nI0126 16:47:58.625706 6452 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {eb8eef51-1a8d-43f9-ae2e-3b2cc00ded60}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0126 16:47:58.625740 6452 ovn.go:134] Ensuring zone local for Pod openshift-multus/multus-additional-cni-plugins-bwpd6 in no\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:57Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-jsbxt_openshift-ovn-kubernetes(689569c4-93ee-4b82-93fc-9af0f97e6dc3)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://07a43ff7860ff65faf51f4a5e8aeb326f06d6d7222677a92560ae2e3a0c81484\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d9a7bafabd44f8b88a6f62cbe45599af8563bc45d9d7ad3a7669f3d8d11eb9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d9a7bafabd44f8b88a6f62cbe45599af8563bc45d9d7ad3a7669f3d8d11eb9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jsbxt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:22Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:22 crc kubenswrapper[4754]: I0126 16:48:22.493938 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:22Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:22 crc kubenswrapper[4754]: I0126 16:48:22.506505 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e459fd8c1093d6f3d0a671751b792b26b7df94c21ced1be1a2250821bcc82d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:22Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:22 crc kubenswrapper[4754]: I0126 16:48:22.520284 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2blzw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca65468e-30c6-4666-962e-cc0de05e67fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4603235bb2af495901926b805fe1c0ffc48824a4394ccf928e7fcd53a673ac3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c25z4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:33Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2blzw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:22Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:22 crc kubenswrapper[4754]: I0126 16:48:22.539380 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bwpd6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"351cc6c0-b8e5-4589-ae63-f98e8ab1eca3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1022f101d36873b41bbc1ed8a4a2590347b49f7cce6a172aa1d5bc9f529aed65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3ce0fe080a094316b9456057e347b90a5ff432daa12b9c56d7aea9bee1f72b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3ce0fe080a094316b9456057e347b90a5ff432daa12b9c56d7aea9bee1f72b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1772bb41eb8a36a863d3c691610bcbee5d16b260480b3e9f9d6b083e09e42799\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1772bb41eb8a36a863d3c691610bcbee5d16b260480b3e9f9d6b083e09e42799\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24848d69756024cc664affb9492764ace0638ef611f910f4f2bf86b638cc443d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://24848d69756024cc664affb9492764ace0638ef611f910f4f2bf86b638cc443d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c264fc7c3300a725566acce4464f01d9314779737624401c164d867831fd93bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c264fc7c3300a725566acce4464f01d9314779737624401c164d867831fd93bf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5acfe327eee4aad927ac3a32d24549eff788b14a1c4bdac359b6b56301d884ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5acfe327eee4aad927ac3a32d24549eff788b14a1c4bdac359b6b56301d884ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://63a9b9317494d0e0b33970ef3d461729b10d2a1ad7d3d2bd33f389b184ca56ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://63a9b9317494d0e0b33970ef3d461729b10d2a1ad7d3d2bd33f389b184ca56ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bwpd6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:22Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:22 crc kubenswrapper[4754]: I0126 16:48:22.565076 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:22 crc kubenswrapper[4754]: I0126 16:48:22.565122 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:22 crc kubenswrapper[4754]: I0126 16:48:22.565131 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:22 crc kubenswrapper[4754]: I0126 16:48:22.565146 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:22 crc kubenswrapper[4754]: I0126 16:48:22.565156 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:22Z","lastTransitionTime":"2026-01-26T16:48:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:22 crc kubenswrapper[4754]: I0126 16:48:22.668616 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:22 crc kubenswrapper[4754]: I0126 16:48:22.668701 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:22 crc kubenswrapper[4754]: I0126 16:48:22.668717 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:22 crc kubenswrapper[4754]: I0126 16:48:22.668736 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:22 crc kubenswrapper[4754]: I0126 16:48:22.668747 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:22Z","lastTransitionTime":"2026-01-26T16:48:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:22 crc kubenswrapper[4754]: I0126 16:48:22.767617 4754 scope.go:117] "RemoveContainer" containerID="0e9d45c28e67d857b1c50ed78197ed052c1ac2b98f1c8a94addd63698bf0bb23" Jan 26 16:48:22 crc kubenswrapper[4754]: I0126 16:48:22.771401 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:22 crc kubenswrapper[4754]: I0126 16:48:22.771442 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:22 crc kubenswrapper[4754]: I0126 16:48:22.771453 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:22 crc kubenswrapper[4754]: I0126 16:48:22.771471 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:22 crc kubenswrapper[4754]: I0126 16:48:22.771484 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:22Z","lastTransitionTime":"2026-01-26T16:48:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:22 crc kubenswrapper[4754]: I0126 16:48:22.874645 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:22 crc kubenswrapper[4754]: I0126 16:48:22.874691 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:22 crc kubenswrapper[4754]: I0126 16:48:22.874702 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:22 crc kubenswrapper[4754]: I0126 16:48:22.874719 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:22 crc kubenswrapper[4754]: I0126 16:48:22.874728 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:22Z","lastTransitionTime":"2026-01-26T16:48:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:22 crc kubenswrapper[4754]: I0126 16:48:22.977473 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:22 crc kubenswrapper[4754]: I0126 16:48:22.977519 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:22 crc kubenswrapper[4754]: I0126 16:48:22.977532 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:22 crc kubenswrapper[4754]: I0126 16:48:22.977549 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:22 crc kubenswrapper[4754]: I0126 16:48:22.977559 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:22Z","lastTransitionTime":"2026-01-26T16:48:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:23 crc kubenswrapper[4754]: I0126 16:48:23.080740 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:23 crc kubenswrapper[4754]: I0126 16:48:23.080795 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:23 crc kubenswrapper[4754]: I0126 16:48:23.080806 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:23 crc kubenswrapper[4754]: I0126 16:48:23.080822 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:23 crc kubenswrapper[4754]: I0126 16:48:23.080833 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:23Z","lastTransitionTime":"2026-01-26T16:48:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:23 crc kubenswrapper[4754]: I0126 16:48:23.184219 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:23 crc kubenswrapper[4754]: I0126 16:48:23.184272 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:23 crc kubenswrapper[4754]: I0126 16:48:23.184286 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:23 crc kubenswrapper[4754]: I0126 16:48:23.184307 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:23 crc kubenswrapper[4754]: I0126 16:48:23.184322 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:23Z","lastTransitionTime":"2026-01-26T16:48:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:23 crc kubenswrapper[4754]: I0126 16:48:23.283852 4754 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-02 17:49:41.740348066 +0000 UTC Jan 26 16:48:23 crc kubenswrapper[4754]: I0126 16:48:23.286754 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:23 crc kubenswrapper[4754]: I0126 16:48:23.286795 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:23 crc kubenswrapper[4754]: I0126 16:48:23.286807 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:23 crc kubenswrapper[4754]: I0126 16:48:23.286822 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:23 crc kubenswrapper[4754]: I0126 16:48:23.286834 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:23Z","lastTransitionTime":"2026-01-26T16:48:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:23 crc kubenswrapper[4754]: I0126 16:48:23.299606 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-jsbxt_689569c4-93ee-4b82-93fc-9af0f97e6dc3/ovnkube-controller/2.log" Jan 26 16:48:23 crc kubenswrapper[4754]: I0126 16:48:23.303010 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jsbxt" event={"ID":"689569c4-93ee-4b82-93fc-9af0f97e6dc3","Type":"ContainerStarted","Data":"2c059b7d618442c8e541bce1d2e576de3e9d9f9b3bd82973f0039fc6c8429d6e"} Jan 26 16:48:23 crc kubenswrapper[4754]: I0126 16:48:23.303580 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-jsbxt" Jan 26 16:48:23 crc kubenswrapper[4754]: I0126 16:48:23.320883 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-t4ps6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d45179f0-9f81-4d77-8805-6b05bdb32f8f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vk8tg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vk8tg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:48Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-t4ps6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:23Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:23 crc kubenswrapper[4754]: I0126 16:48:23.337423 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b89aa081-0a2e-4757-9210-e2a6c228bee9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9c5f40188fda14b14128514fe84bddfe8dcede3d6c560283ccb2ac0c66895fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f45445cbf1d180d6d4dd9588e85bf678f043bfce8cff0812b78ca3510904f9a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://085fa7b8d4fef8b2722fac8c346107699b5a6bd33fbfc74ed35831cb8fdf9dad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c40e018a0979035827963efb9efc48a878a55b9df68ec1a9242d7cff64fe480\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8356f76fbc3a4e636c12d1da5455d3f6767d9b83ce2f2925a11435d7d7261e9f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0126 16:47:26.205208 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0126 16:47:26.206538 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3645934156/tls.crt::/tmp/serving-cert-3645934156/tls.key\\\\\\\"\\\\nI0126 16:47:31.504518 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0126 16:47:31.509366 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0126 16:47:31.509400 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0126 16:47:31.509428 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0126 16:47:31.509436 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0126 16:47:31.515912 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0126 16:47:31.516036 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 16:47:31.516068 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 16:47:31.516096 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0126 16:47:31.516121 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0126 16:47:31.516145 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0126 16:47:31.516170 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0126 16:47:31.516412 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0126 16:47:31.519294 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27b74e2e65a547e0e28fddd4f15feacb17f4d57596eeb8da82df9d39243c85e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22ba7649a1308815bfad52016000987a7a9a5bc214d7ff61e137ba099b20fc6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://22ba7649a1308815bfad52016000987a7a9a5bc214d7ff61e137ba099b20fc6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:13Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:23Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:23 crc kubenswrapper[4754]: I0126 16:48:23.354468 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://205ec256e46a8e08b4b0b0805fb6c5f4966fc3ccef0738df4b97cb404e74da17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:23Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:23 crc kubenswrapper[4754]: I0126 16:48:23.368515 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:23Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:23 crc kubenswrapper[4754]: I0126 16:48:23.379842 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c3718a4-f354-4284-92e0-fdfb45a692bd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b86f46f49163bca7eca9458279029af330dbd4890b03325e06fa3b1bdf030de6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxjcz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fffc03b536bfee3c7bc1bddd500e834e34b391896477a1501e3e06201a19374\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxjcz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-x65wv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:23Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:23 crc kubenswrapper[4754]: I0126 16:48:23.389442 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:23 crc kubenswrapper[4754]: I0126 16:48:23.389497 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:23 crc kubenswrapper[4754]: I0126 16:48:23.389511 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:23 crc kubenswrapper[4754]: I0126 16:48:23.389532 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:23 crc kubenswrapper[4754]: I0126 16:48:23.389544 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:23Z","lastTransitionTime":"2026-01-26T16:48:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:23 crc kubenswrapper[4754]: I0126 16:48:23.394954 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bwpd6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"351cc6c0-b8e5-4589-ae63-f98e8ab1eca3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1022f101d36873b41bbc1ed8a4a2590347b49f7cce6a172aa1d5bc9f529aed65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3ce0fe080a094316b9456057e347b90a5ff432daa12b9c56d7aea9bee1f72b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3ce0fe080a094316b9456057e347b90a5ff432daa12b9c56d7aea9bee1f72b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1772bb41eb8a36a863d3c691610bcbee5d16b260480b3e9f9d6b083e09e42799\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1772bb41eb8a36a863d3c691610bcbee5d16b260480b3e9f9d6b083e09e42799\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24848d69756024cc664affb9492764ace0638ef611f910f4f2bf86b638cc443d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://24848d69756024cc664affb9492764ace0638ef611f910f4f2bf86b638cc443d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c264fc7c3300a725566acce4464f01d9314779737624401c164d867831fd93bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c264fc7c3300a725566acce4464f01d9314779737624401c164d867831fd93bf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5acfe327eee4aad927ac3a32d24549eff788b14a1c4bdac359b6b56301d884ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5acfe327eee4aad927ac3a32d24549eff788b14a1c4bdac359b6b56301d884ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://63a9b9317494d0e0b33970ef3d461729b10d2a1ad7d3d2bd33f389b184ca56ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://63a9b9317494d0e0b33970ef3d461729b10d2a1ad7d3d2bd33f389b184ca56ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bwpd6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:23Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:23 crc kubenswrapper[4754]: I0126 16:48:23.416887 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jsbxt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"689569c4-93ee-4b82-93fc-9af0f97e6dc3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://033f8b104d4e9ff6d5f87ca998665f12e0c56696b3d30e01ede05f4da540365d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://919572e731813a98c7d8a94879bba931f75c472c7cd681e86d5197ebab43815d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1ac1f49bd564750fcff530a0f76657b0ed41573539acd020a922d19861ef834\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a300a80ad87f46ba1b4fd7ad3cf6c25767fcfe966b93bdd86d729843f53a5cda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31dcab2f9a2d218b0e49dad1df46377ff4521592824a0fbac254d5b8e6d0632e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fe4d9395447ac570eca1338e749dbefea76295d427a12487b63a142a7d6a2e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c059b7d618442c8e541bce1d2e576de3e9d9f9b3bd82973f0039fc6c8429d6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0e9d45c28e67d857b1c50ed78197ed052c1ac2b98f1c8a94addd63698bf0bb23\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-26T16:47:58Z\\\",\\\"message\\\":\\\":303] Retry object setup: *v1.Pod openshift-machine-config-operator/machine-config-daemon-x65wv\\\\nI0126 16:47:58.625627 6452 obj_retry.go:303] Retry object setup: *v1.Pod openshift-multus/multus-skh5n\\\\nI0126 16:47:58.625718 6452 obj_retry.go:365] Adding new object: *v1.Pod openshift-machine-config-operator/machine-config-daemon-x65wv\\\\nI0126 16:47:58.625617 6452 obj_retry.go:303] Retry object setup: *v1.Pod openshift-multus/multus-additional-cni-plugins-bwpd6\\\\nI0126 16:47:58.625726 6452 ovn.go:134] Ensuring zone local for Pod openshift-machine-config-operator/machine-config-daemon-x65wv in node crc\\\\nI0126 16:47:58.625727 6452 obj_retry.go:365] Adding new object: *v1.Pod openshift-multus/multus-skh5n\\\\nI0126 16:47:58.625732 6452 obj_retry.go:365] Adding new object: *v1.Pod openshift-multus/multus-additional-cni-plugins-bwpd6\\\\nI0126 16:47:58.625706 6452 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {eb8eef51-1a8d-43f9-ae2e-3b2cc00ded60}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0126 16:47:58.625740 6452 ovn.go:134] Ensuring zone local for Pod openshift-multus/multus-additional-cni-plugins-bwpd6 in no\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:57Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:48:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://07a43ff7860ff65faf51f4a5e8aeb326f06d6d7222677a92560ae2e3a0c81484\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d9a7bafabd44f8b88a6f62cbe45599af8563bc45d9d7ad3a7669f3d8d11eb9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d9a7bafabd44f8b88a6f62cbe45599af8563bc45d9d7ad3a7669f3d8d11eb9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jsbxt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:23Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:23 crc kubenswrapper[4754]: I0126 16:48:23.433966 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:23Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:23 crc kubenswrapper[4754]: I0126 16:48:23.450227 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e459fd8c1093d6f3d0a671751b792b26b7df94c21ced1be1a2250821bcc82d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:23Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:23 crc kubenswrapper[4754]: I0126 16:48:23.471627 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2blzw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca65468e-30c6-4666-962e-cc0de05e67fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4603235bb2af495901926b805fe1c0ffc48824a4394ccf928e7fcd53a673ac3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c25z4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:33Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2blzw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:23Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:23 crc kubenswrapper[4754]: I0126 16:48:23.492441 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:23 crc kubenswrapper[4754]: I0126 16:48:23.492476 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:23 crc kubenswrapper[4754]: I0126 16:48:23.492486 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:23 crc kubenswrapper[4754]: I0126 16:48:23.492501 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:23 crc kubenswrapper[4754]: I0126 16:48:23.492484 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-84lgq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6763d962-0892-4acc-bab9-7a4e93e39a85\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://953c1917fca896c3def3fc2c69f9e3dbdc9ca26fc2b3d40282be71f71759fd68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8zlw6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:33Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-84lgq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:23Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:23 crc kubenswrapper[4754]: I0126 16:48:23.492511 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:23Z","lastTransitionTime":"2026-01-26T16:48:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:23 crc kubenswrapper[4754]: I0126 16:48:23.513903 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-skh5n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b619d233-b592-4b05-a0b7-dc094c88471e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:48:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:48:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8d026fdb7ea7dc5c40add18606d8f1a7703865edf93ef19edaf811f598364c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9231a4191565e2e0ef53d8a394bb34615c900b601a0dfec24c2d8e5fa7b0b659\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-26T16:48:20Z\\\",\\\"message\\\":\\\"2026-01-26T16:47:34+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_57895685-a7ab-4a6c-9a6a-49247f349eaf\\\\n2026-01-26T16:47:34+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_57895685-a7ab-4a6c-9a6a-49247f349eaf to /host/opt/cni/bin/\\\\n2026-01-26T16:47:35Z [verbose] multus-daemon started\\\\n2026-01-26T16:47:35Z [verbose] Readiness Indicator file check\\\\n2026-01-26T16:48:20Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:48:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-skh5n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:23Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:23 crc kubenswrapper[4754]: I0126 16:48:23.527252 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jspvh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5c81aab0-939e-45ce-b19a-89d4818980c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://356dc54150a495dd6f35e0667c82b807b712b07200d8a0adbb1c92fe3f7af8f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8bps\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://05fd57780c44cd127859dc8b163d20eafbd8b06f3348cf545ef41021ce6bc526\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8bps\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-jspvh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:23Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:23 crc kubenswrapper[4754]: I0126 16:48:23.543781 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"099af702-abf2-4933-b4c9-63ef2326e468\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23017e73f8450839b4c2c02fa61235ffe4604bdce11556def27f25bcadcf89dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a74507c941b9de4d0401c8f1d53bb79ed11359628d3405cdab7f5474a75c862e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4402182b3e2e35a7d5dd4c44ac53e152915b312b7ccb076c4f2b41b925a6e7f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f772a003566198519fe1849fd14f29d4346004b2e655810662507c95b4d5ffa1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:13Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:23Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:23 crc kubenswrapper[4754]: I0126 16:48:23.560896 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab1ace4c-137b-4b8a-8032-0b86a164f063\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:48:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:48:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7da43edaa2a81f1b647b96c234d795b1ddf5fa23173b78043123f76dc677c517\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e5fbff42fdebad6a74c21d85931d3c8d983973db82fc7f262ba5383bad76172\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f9bf9e4a705cd849595e59e1123ba70773bfa5f8a0d0b79195e4fc7e447ca26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb7a3b85cca4436068c8ceefc3f2358cfa1a2c398a8acecd367b944528b11489\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb7a3b85cca4436068c8ceefc3f2358cfa1a2c398a8acecd367b944528b11489\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:14Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:13Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:23Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:23 crc kubenswrapper[4754]: I0126 16:48:23.580545 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:23Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:23 crc kubenswrapper[4754]: I0126 16:48:23.595261 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:23 crc kubenswrapper[4754]: I0126 16:48:23.595294 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:23 crc kubenswrapper[4754]: I0126 16:48:23.595303 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:23 crc kubenswrapper[4754]: I0126 16:48:23.595319 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:23 crc kubenswrapper[4754]: I0126 16:48:23.595330 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:23Z","lastTransitionTime":"2026-01-26T16:48:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:23 crc kubenswrapper[4754]: I0126 16:48:23.597389 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89251fa9c64bc25d6d6f355400d7644c099f22aed1cae812fc991a03595f1e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e90508d8a53be9146f9b388750730fdae9ee200306caf5ad393a356390a69ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:23Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:23 crc kubenswrapper[4754]: I0126 16:48:23.698649 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:23 crc kubenswrapper[4754]: I0126 16:48:23.698720 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:23 crc kubenswrapper[4754]: I0126 16:48:23.698739 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:23 crc kubenswrapper[4754]: I0126 16:48:23.698761 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:23 crc kubenswrapper[4754]: I0126 16:48:23.698775 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:23Z","lastTransitionTime":"2026-01-26T16:48:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:23 crc kubenswrapper[4754]: I0126 16:48:23.766805 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 16:48:23 crc kubenswrapper[4754]: I0126 16:48:23.766856 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 16:48:23 crc kubenswrapper[4754]: E0126 16:48:23.766931 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 16:48:23 crc kubenswrapper[4754]: I0126 16:48:23.766950 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-t4ps6" Jan 26 16:48:23 crc kubenswrapper[4754]: E0126 16:48:23.767072 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 16:48:23 crc kubenswrapper[4754]: E0126 16:48:23.767200 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-t4ps6" podUID="d45179f0-9f81-4d77-8805-6b05bdb32f8f" Jan 26 16:48:23 crc kubenswrapper[4754]: I0126 16:48:23.767462 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 16:48:23 crc kubenswrapper[4754]: E0126 16:48:23.767548 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 16:48:23 crc kubenswrapper[4754]: I0126 16:48:23.785613 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-skh5n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b619d233-b592-4b05-a0b7-dc094c88471e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:48:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:48:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8d026fdb7ea7dc5c40add18606d8f1a7703865edf93ef19edaf811f598364c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9231a4191565e2e0ef53d8a394bb34615c900b601a0dfec24c2d8e5fa7b0b659\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-26T16:48:20Z\\\",\\\"message\\\":\\\"2026-01-26T16:47:34+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_57895685-a7ab-4a6c-9a6a-49247f349eaf\\\\n2026-01-26T16:47:34+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_57895685-a7ab-4a6c-9a6a-49247f349eaf to /host/opt/cni/bin/\\\\n2026-01-26T16:47:35Z [verbose] multus-daemon started\\\\n2026-01-26T16:47:35Z [verbose] Readiness Indicator file check\\\\n2026-01-26T16:48:20Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:48:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-skh5n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:23Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:23 crc kubenswrapper[4754]: I0126 16:48:23.802403 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:23 crc kubenswrapper[4754]: I0126 16:48:23.802411 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jspvh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5c81aab0-939e-45ce-b19a-89d4818980c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://356dc54150a495dd6f35e0667c82b807b712b07200d8a0adbb1c92fe3f7af8f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8bps\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://05fd57780c44cd127859dc8b163d20eafbd8b06f3348cf545ef41021ce6bc526\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8bps\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-jspvh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:23Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:23 crc kubenswrapper[4754]: I0126 16:48:23.802444 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:23 crc kubenswrapper[4754]: I0126 16:48:23.802572 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:23 crc kubenswrapper[4754]: I0126 16:48:23.802594 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:23 crc kubenswrapper[4754]: I0126 16:48:23.802612 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:23Z","lastTransitionTime":"2026-01-26T16:48:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:23 crc kubenswrapper[4754]: I0126 16:48:23.824185 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"099af702-abf2-4933-b4c9-63ef2326e468\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23017e73f8450839b4c2c02fa61235ffe4604bdce11556def27f25bcadcf89dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a74507c941b9de4d0401c8f1d53bb79ed11359628d3405cdab7f5474a75c862e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4402182b3e2e35a7d5dd4c44ac53e152915b312b7ccb076c4f2b41b925a6e7f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f772a003566198519fe1849fd14f29d4346004b2e655810662507c95b4d5ffa1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:13Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:23Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:23 crc kubenswrapper[4754]: I0126 16:48:23.841016 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab1ace4c-137b-4b8a-8032-0b86a164f063\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:48:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:48:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7da43edaa2a81f1b647b96c234d795b1ddf5fa23173b78043123f76dc677c517\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e5fbff42fdebad6a74c21d85931d3c8d983973db82fc7f262ba5383bad76172\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f9bf9e4a705cd849595e59e1123ba70773bfa5f8a0d0b79195e4fc7e447ca26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb7a3b85cca4436068c8ceefc3f2358cfa1a2c398a8acecd367b944528b11489\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb7a3b85cca4436068c8ceefc3f2358cfa1a2c398a8acecd367b944528b11489\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:14Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:13Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:23Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:23 crc kubenswrapper[4754]: I0126 16:48:23.858685 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:23Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:23 crc kubenswrapper[4754]: I0126 16:48:23.876368 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89251fa9c64bc25d6d6f355400d7644c099f22aed1cae812fc991a03595f1e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e90508d8a53be9146f9b388750730fdae9ee200306caf5ad393a356390a69ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:23Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:23 crc kubenswrapper[4754]: I0126 16:48:23.891552 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-84lgq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6763d962-0892-4acc-bab9-7a4e93e39a85\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://953c1917fca896c3def3fc2c69f9e3dbdc9ca26fc2b3d40282be71f71759fd68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8zlw6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:33Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-84lgq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:23Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:23 crc kubenswrapper[4754]: I0126 16:48:23.904861 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:23 crc kubenswrapper[4754]: I0126 16:48:23.904883 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:23 crc kubenswrapper[4754]: I0126 16:48:23.904893 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:23 crc kubenswrapper[4754]: I0126 16:48:23.904908 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:23 crc kubenswrapper[4754]: I0126 16:48:23.904917 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:23Z","lastTransitionTime":"2026-01-26T16:48:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:23 crc kubenswrapper[4754]: I0126 16:48:23.917652 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b89aa081-0a2e-4757-9210-e2a6c228bee9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9c5f40188fda14b14128514fe84bddfe8dcede3d6c560283ccb2ac0c66895fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f45445cbf1d180d6d4dd9588e85bf678f043bfce8cff0812b78ca3510904f9a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://085fa7b8d4fef8b2722fac8c346107699b5a6bd33fbfc74ed35831cb8fdf9dad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c40e018a0979035827963efb9efc48a878a55b9df68ec1a9242d7cff64fe480\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8356f76fbc3a4e636c12d1da5455d3f6767d9b83ce2f2925a11435d7d7261e9f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0126 16:47:26.205208 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0126 16:47:26.206538 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3645934156/tls.crt::/tmp/serving-cert-3645934156/tls.key\\\\\\\"\\\\nI0126 16:47:31.504518 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0126 16:47:31.509366 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0126 16:47:31.509400 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0126 16:47:31.509428 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0126 16:47:31.509436 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0126 16:47:31.515912 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0126 16:47:31.516036 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 16:47:31.516068 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 16:47:31.516096 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0126 16:47:31.516121 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0126 16:47:31.516145 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0126 16:47:31.516170 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0126 16:47:31.516412 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0126 16:47:31.519294 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27b74e2e65a547e0e28fddd4f15feacb17f4d57596eeb8da82df9d39243c85e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22ba7649a1308815bfad52016000987a7a9a5bc214d7ff61e137ba099b20fc6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://22ba7649a1308815bfad52016000987a7a9a5bc214d7ff61e137ba099b20fc6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:13Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:23Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:23 crc kubenswrapper[4754]: I0126 16:48:23.931225 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://205ec256e46a8e08b4b0b0805fb6c5f4966fc3ccef0738df4b97cb404e74da17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:23Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:23 crc kubenswrapper[4754]: I0126 16:48:23.942828 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:23Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:23 crc kubenswrapper[4754]: I0126 16:48:23.953855 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c3718a4-f354-4284-92e0-fdfb45a692bd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b86f46f49163bca7eca9458279029af330dbd4890b03325e06fa3b1bdf030de6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxjcz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fffc03b536bfee3c7bc1bddd500e834e34b391896477a1501e3e06201a19374\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxjcz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-x65wv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:23Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:23 crc kubenswrapper[4754]: I0126 16:48:23.966827 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-t4ps6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d45179f0-9f81-4d77-8805-6b05bdb32f8f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vk8tg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vk8tg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:48Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-t4ps6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:23Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:23 crc kubenswrapper[4754]: I0126 16:48:23.987405 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jsbxt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"689569c4-93ee-4b82-93fc-9af0f97e6dc3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://033f8b104d4e9ff6d5f87ca998665f12e0c56696b3d30e01ede05f4da540365d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://919572e731813a98c7d8a94879bba931f75c472c7cd681e86d5197ebab43815d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1ac1f49bd564750fcff530a0f76657b0ed41573539acd020a922d19861ef834\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a300a80ad87f46ba1b4fd7ad3cf6c25767fcfe966b93bdd86d729843f53a5cda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31dcab2f9a2d218b0e49dad1df46377ff4521592824a0fbac254d5b8e6d0632e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fe4d9395447ac570eca1338e749dbefea76295d427a12487b63a142a7d6a2e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c059b7d618442c8e541bce1d2e576de3e9d9f9b3bd82973f0039fc6c8429d6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0e9d45c28e67d857b1c50ed78197ed052c1ac2b98f1c8a94addd63698bf0bb23\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-26T16:47:58Z\\\",\\\"message\\\":\\\":303] Retry object setup: *v1.Pod openshift-machine-config-operator/machine-config-daemon-x65wv\\\\nI0126 16:47:58.625627 6452 obj_retry.go:303] Retry object setup: *v1.Pod openshift-multus/multus-skh5n\\\\nI0126 16:47:58.625718 6452 obj_retry.go:365] Adding new object: *v1.Pod openshift-machine-config-operator/machine-config-daemon-x65wv\\\\nI0126 16:47:58.625617 6452 obj_retry.go:303] Retry object setup: *v1.Pod openshift-multus/multus-additional-cni-plugins-bwpd6\\\\nI0126 16:47:58.625726 6452 ovn.go:134] Ensuring zone local for Pod openshift-machine-config-operator/machine-config-daemon-x65wv in node crc\\\\nI0126 16:47:58.625727 6452 obj_retry.go:365] Adding new object: *v1.Pod openshift-multus/multus-skh5n\\\\nI0126 16:47:58.625732 6452 obj_retry.go:365] Adding new object: *v1.Pod openshift-multus/multus-additional-cni-plugins-bwpd6\\\\nI0126 16:47:58.625706 6452 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {eb8eef51-1a8d-43f9-ae2e-3b2cc00ded60}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0126 16:47:58.625740 6452 ovn.go:134] Ensuring zone local for Pod openshift-multus/multus-additional-cni-plugins-bwpd6 in no\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:57Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:48:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://07a43ff7860ff65faf51f4a5e8aeb326f06d6d7222677a92560ae2e3a0c81484\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d9a7bafabd44f8b88a6f62cbe45599af8563bc45d9d7ad3a7669f3d8d11eb9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d9a7bafabd44f8b88a6f62cbe45599af8563bc45d9d7ad3a7669f3d8d11eb9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jsbxt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:23Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:24 crc kubenswrapper[4754]: I0126 16:48:24.000079 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:23Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:24 crc kubenswrapper[4754]: I0126 16:48:24.007269 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:24 crc kubenswrapper[4754]: I0126 16:48:24.007479 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:24 crc kubenswrapper[4754]: I0126 16:48:24.007560 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:24 crc kubenswrapper[4754]: I0126 16:48:24.007650 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:24 crc kubenswrapper[4754]: I0126 16:48:24.007769 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:24Z","lastTransitionTime":"2026-01-26T16:48:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:24 crc kubenswrapper[4754]: I0126 16:48:24.011753 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e459fd8c1093d6f3d0a671751b792b26b7df94c21ced1be1a2250821bcc82d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:24Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:24 crc kubenswrapper[4754]: I0126 16:48:24.024206 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2blzw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca65468e-30c6-4666-962e-cc0de05e67fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4603235bb2af495901926b805fe1c0ffc48824a4394ccf928e7fcd53a673ac3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c25z4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:33Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2blzw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:24Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:24 crc kubenswrapper[4754]: I0126 16:48:24.039997 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bwpd6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"351cc6c0-b8e5-4589-ae63-f98e8ab1eca3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1022f101d36873b41bbc1ed8a4a2590347b49f7cce6a172aa1d5bc9f529aed65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3ce0fe080a094316b9456057e347b90a5ff432daa12b9c56d7aea9bee1f72b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3ce0fe080a094316b9456057e347b90a5ff432daa12b9c56d7aea9bee1f72b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1772bb41eb8a36a863d3c691610bcbee5d16b260480b3e9f9d6b083e09e42799\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1772bb41eb8a36a863d3c691610bcbee5d16b260480b3e9f9d6b083e09e42799\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24848d69756024cc664affb9492764ace0638ef611f910f4f2bf86b638cc443d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://24848d69756024cc664affb9492764ace0638ef611f910f4f2bf86b638cc443d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c264fc7c3300a725566acce4464f01d9314779737624401c164d867831fd93bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c264fc7c3300a725566acce4464f01d9314779737624401c164d867831fd93bf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5acfe327eee4aad927ac3a32d24549eff788b14a1c4bdac359b6b56301d884ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5acfe327eee4aad927ac3a32d24549eff788b14a1c4bdac359b6b56301d884ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://63a9b9317494d0e0b33970ef3d461729b10d2a1ad7d3d2bd33f389b184ca56ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://63a9b9317494d0e0b33970ef3d461729b10d2a1ad7d3d2bd33f389b184ca56ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bwpd6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:24Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:24 crc kubenswrapper[4754]: I0126 16:48:24.111414 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:24 crc kubenswrapper[4754]: I0126 16:48:24.111474 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:24 crc kubenswrapper[4754]: I0126 16:48:24.111486 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:24 crc kubenswrapper[4754]: I0126 16:48:24.111512 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:24 crc kubenswrapper[4754]: I0126 16:48:24.111533 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:24Z","lastTransitionTime":"2026-01-26T16:48:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:24 crc kubenswrapper[4754]: I0126 16:48:24.214956 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:24 crc kubenswrapper[4754]: I0126 16:48:24.215003 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:24 crc kubenswrapper[4754]: I0126 16:48:24.215025 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:24 crc kubenswrapper[4754]: I0126 16:48:24.215046 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:24 crc kubenswrapper[4754]: I0126 16:48:24.215060 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:24Z","lastTransitionTime":"2026-01-26T16:48:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:24 crc kubenswrapper[4754]: I0126 16:48:24.284885 4754 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-28 04:41:59.8526653 +0000 UTC Jan 26 16:48:24 crc kubenswrapper[4754]: I0126 16:48:24.310538 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-jsbxt_689569c4-93ee-4b82-93fc-9af0f97e6dc3/ovnkube-controller/3.log" Jan 26 16:48:24 crc kubenswrapper[4754]: I0126 16:48:24.311396 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-jsbxt_689569c4-93ee-4b82-93fc-9af0f97e6dc3/ovnkube-controller/2.log" Jan 26 16:48:24 crc kubenswrapper[4754]: I0126 16:48:24.315441 4754 generic.go:334] "Generic (PLEG): container finished" podID="689569c4-93ee-4b82-93fc-9af0f97e6dc3" containerID="2c059b7d618442c8e541bce1d2e576de3e9d9f9b3bd82973f0039fc6c8429d6e" exitCode=1 Jan 26 16:48:24 crc kubenswrapper[4754]: I0126 16:48:24.315598 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jsbxt" event={"ID":"689569c4-93ee-4b82-93fc-9af0f97e6dc3","Type":"ContainerDied","Data":"2c059b7d618442c8e541bce1d2e576de3e9d9f9b3bd82973f0039fc6c8429d6e"} Jan 26 16:48:24 crc kubenswrapper[4754]: I0126 16:48:24.315710 4754 scope.go:117] "RemoveContainer" containerID="0e9d45c28e67d857b1c50ed78197ed052c1ac2b98f1c8a94addd63698bf0bb23" Jan 26 16:48:24 crc kubenswrapper[4754]: I0126 16:48:24.316955 4754 scope.go:117] "RemoveContainer" containerID="2c059b7d618442c8e541bce1d2e576de3e9d9f9b3bd82973f0039fc6c8429d6e" Jan 26 16:48:24 crc kubenswrapper[4754]: E0126 16:48:24.317292 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-jsbxt_openshift-ovn-kubernetes(689569c4-93ee-4b82-93fc-9af0f97e6dc3)\"" pod="openshift-ovn-kubernetes/ovnkube-node-jsbxt" podUID="689569c4-93ee-4b82-93fc-9af0f97e6dc3" Jan 26 16:48:24 crc kubenswrapper[4754]: I0126 16:48:24.318079 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:24 crc kubenswrapper[4754]: I0126 16:48:24.318109 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:24 crc kubenswrapper[4754]: I0126 16:48:24.318158 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:24 crc kubenswrapper[4754]: I0126 16:48:24.318179 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:24 crc kubenswrapper[4754]: I0126 16:48:24.318195 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:24Z","lastTransitionTime":"2026-01-26T16:48:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:24 crc kubenswrapper[4754]: I0126 16:48:24.342389 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jsbxt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"689569c4-93ee-4b82-93fc-9af0f97e6dc3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://033f8b104d4e9ff6d5f87ca998665f12e0c56696b3d30e01ede05f4da540365d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://919572e731813a98c7d8a94879bba931f75c472c7cd681e86d5197ebab43815d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1ac1f49bd564750fcff530a0f76657b0ed41573539acd020a922d19861ef834\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a300a80ad87f46ba1b4fd7ad3cf6c25767fcfe966b93bdd86d729843f53a5cda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31dcab2f9a2d218b0e49dad1df46377ff4521592824a0fbac254d5b8e6d0632e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fe4d9395447ac570eca1338e749dbefea76295d427a12487b63a142a7d6a2e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c059b7d618442c8e541bce1d2e576de3e9d9f9b3bd82973f0039fc6c8429d6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0e9d45c28e67d857b1c50ed78197ed052c1ac2b98f1c8a94addd63698bf0bb23\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-26T16:47:58Z\\\",\\\"message\\\":\\\":303] Retry object setup: *v1.Pod openshift-machine-config-operator/machine-config-daemon-x65wv\\\\nI0126 16:47:58.625627 6452 obj_retry.go:303] Retry object setup: *v1.Pod openshift-multus/multus-skh5n\\\\nI0126 16:47:58.625718 6452 obj_retry.go:365] Adding new object: *v1.Pod openshift-machine-config-operator/machine-config-daemon-x65wv\\\\nI0126 16:47:58.625617 6452 obj_retry.go:303] Retry object setup: *v1.Pod openshift-multus/multus-additional-cni-plugins-bwpd6\\\\nI0126 16:47:58.625726 6452 ovn.go:134] Ensuring zone local for Pod openshift-machine-config-operator/machine-config-daemon-x65wv in node crc\\\\nI0126 16:47:58.625727 6452 obj_retry.go:365] Adding new object: *v1.Pod openshift-multus/multus-skh5n\\\\nI0126 16:47:58.625732 6452 obj_retry.go:365] Adding new object: *v1.Pod openshift-multus/multus-additional-cni-plugins-bwpd6\\\\nI0126 16:47:58.625706 6452 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {eb8eef51-1a8d-43f9-ae2e-3b2cc00ded60}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0126 16:47:58.625740 6452 ovn.go:134] Ensuring zone local for Pod openshift-multus/multus-additional-cni-plugins-bwpd6 in no\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:57Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2c059b7d618442c8e541bce1d2e576de3e9d9f9b3bd82973f0039fc6c8429d6e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-26T16:48:23Z\\\",\\\"message\\\":\\\"*v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0126 16:48:23.624369 6831 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI0126 16:48:23.624638 6831 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI0126 16:48:23.625131 6831 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0126 16:48:23.625229 6831 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0126 16:48:23.625927 6831 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0126 16:48:23.626017 6831 factory.go:656] Stopping watch factory\\\\nI0126 16:48:23.626054 6831 handler.go:208] Removed *v1.Node event handler 2\\\\nI0126 16:48:23.680081 6831 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI0126 16:48:23.680139 6831 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI0126 16:48:23.680356 6831 ovnkube.go:599] Stopped ovnkube\\\\nI0126 16:48:23.680442 6831 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0126 16:48:23.680583 6831 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T16:48:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://07a43ff7860ff65faf51f4a5e8aeb326f06d6d7222677a92560ae2e3a0c81484\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d9a7bafabd44f8b88a6f62cbe45599af8563bc45d9d7ad3a7669f3d8d11eb9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d9a7bafabd44f8b88a6f62cbe45599af8563bc45d9d7ad3a7669f3d8d11eb9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jsbxt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:24Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:24 crc kubenswrapper[4754]: I0126 16:48:24.358177 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:24Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:24 crc kubenswrapper[4754]: I0126 16:48:24.373294 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e459fd8c1093d6f3d0a671751b792b26b7df94c21ced1be1a2250821bcc82d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:24Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:24 crc kubenswrapper[4754]: I0126 16:48:24.387434 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2blzw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca65468e-30c6-4666-962e-cc0de05e67fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4603235bb2af495901926b805fe1c0ffc48824a4394ccf928e7fcd53a673ac3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c25z4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:33Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2blzw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:24Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:24 crc kubenswrapper[4754]: I0126 16:48:24.405463 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bwpd6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"351cc6c0-b8e5-4589-ae63-f98e8ab1eca3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1022f101d36873b41bbc1ed8a4a2590347b49f7cce6a172aa1d5bc9f529aed65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3ce0fe080a094316b9456057e347b90a5ff432daa12b9c56d7aea9bee1f72b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3ce0fe080a094316b9456057e347b90a5ff432daa12b9c56d7aea9bee1f72b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1772bb41eb8a36a863d3c691610bcbee5d16b260480b3e9f9d6b083e09e42799\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1772bb41eb8a36a863d3c691610bcbee5d16b260480b3e9f9d6b083e09e42799\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24848d69756024cc664affb9492764ace0638ef611f910f4f2bf86b638cc443d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://24848d69756024cc664affb9492764ace0638ef611f910f4f2bf86b638cc443d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c264fc7c3300a725566acce4464f01d9314779737624401c164d867831fd93bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c264fc7c3300a725566acce4464f01d9314779737624401c164d867831fd93bf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5acfe327eee4aad927ac3a32d24549eff788b14a1c4bdac359b6b56301d884ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5acfe327eee4aad927ac3a32d24549eff788b14a1c4bdac359b6b56301d884ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://63a9b9317494d0e0b33970ef3d461729b10d2a1ad7d3d2bd33f389b184ca56ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://63a9b9317494d0e0b33970ef3d461729b10d2a1ad7d3d2bd33f389b184ca56ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bwpd6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:24Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:24 crc kubenswrapper[4754]: I0126 16:48:24.421456 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:24 crc kubenswrapper[4754]: I0126 16:48:24.421541 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:24 crc kubenswrapper[4754]: I0126 16:48:24.421559 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:24 crc kubenswrapper[4754]: I0126 16:48:24.421586 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:24 crc kubenswrapper[4754]: I0126 16:48:24.421599 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:24Z","lastTransitionTime":"2026-01-26T16:48:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:24 crc kubenswrapper[4754]: I0126 16:48:24.423128 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-skh5n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b619d233-b592-4b05-a0b7-dc094c88471e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:48:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:48:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8d026fdb7ea7dc5c40add18606d8f1a7703865edf93ef19edaf811f598364c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9231a4191565e2e0ef53d8a394bb34615c900b601a0dfec24c2d8e5fa7b0b659\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-26T16:48:20Z\\\",\\\"message\\\":\\\"2026-01-26T16:47:34+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_57895685-a7ab-4a6c-9a6a-49247f349eaf\\\\n2026-01-26T16:47:34+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_57895685-a7ab-4a6c-9a6a-49247f349eaf to /host/opt/cni/bin/\\\\n2026-01-26T16:47:35Z [verbose] multus-daemon started\\\\n2026-01-26T16:47:35Z [verbose] Readiness Indicator file check\\\\n2026-01-26T16:48:20Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:48:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-skh5n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:24Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:24 crc kubenswrapper[4754]: I0126 16:48:24.440328 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jspvh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5c81aab0-939e-45ce-b19a-89d4818980c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://356dc54150a495dd6f35e0667c82b807b712b07200d8a0adbb1c92fe3f7af8f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8bps\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://05fd57780c44cd127859dc8b163d20eafbd8b06f3348cf545ef41021ce6bc526\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8bps\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-jspvh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:24Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:24 crc kubenswrapper[4754]: I0126 16:48:24.457521 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"099af702-abf2-4933-b4c9-63ef2326e468\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23017e73f8450839b4c2c02fa61235ffe4604bdce11556def27f25bcadcf89dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a74507c941b9de4d0401c8f1d53bb79ed11359628d3405cdab7f5474a75c862e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4402182b3e2e35a7d5dd4c44ac53e152915b312b7ccb076c4f2b41b925a6e7f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f772a003566198519fe1849fd14f29d4346004b2e655810662507c95b4d5ffa1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:13Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:24Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:24 crc kubenswrapper[4754]: I0126 16:48:24.473239 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab1ace4c-137b-4b8a-8032-0b86a164f063\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:48:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:48:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7da43edaa2a81f1b647b96c234d795b1ddf5fa23173b78043123f76dc677c517\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e5fbff42fdebad6a74c21d85931d3c8d983973db82fc7f262ba5383bad76172\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f9bf9e4a705cd849595e59e1123ba70773bfa5f8a0d0b79195e4fc7e447ca26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb7a3b85cca4436068c8ceefc3f2358cfa1a2c398a8acecd367b944528b11489\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb7a3b85cca4436068c8ceefc3f2358cfa1a2c398a8acecd367b944528b11489\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:14Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:13Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:24Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:24 crc kubenswrapper[4754]: I0126 16:48:24.487138 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:24Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:24 crc kubenswrapper[4754]: I0126 16:48:24.502124 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89251fa9c64bc25d6d6f355400d7644c099f22aed1cae812fc991a03595f1e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e90508d8a53be9146f9b388750730fdae9ee200306caf5ad393a356390a69ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:24Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:24 crc kubenswrapper[4754]: I0126 16:48:24.513413 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-84lgq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6763d962-0892-4acc-bab9-7a4e93e39a85\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://953c1917fca896c3def3fc2c69f9e3dbdc9ca26fc2b3d40282be71f71759fd68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8zlw6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:33Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-84lgq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:24Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:24 crc kubenswrapper[4754]: I0126 16:48:24.525950 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:24 crc kubenswrapper[4754]: I0126 16:48:24.525993 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:24 crc kubenswrapper[4754]: I0126 16:48:24.526005 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:24 crc kubenswrapper[4754]: I0126 16:48:24.526026 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:24 crc kubenswrapper[4754]: I0126 16:48:24.526040 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:24Z","lastTransitionTime":"2026-01-26T16:48:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:24 crc kubenswrapper[4754]: I0126 16:48:24.530583 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b89aa081-0a2e-4757-9210-e2a6c228bee9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9c5f40188fda14b14128514fe84bddfe8dcede3d6c560283ccb2ac0c66895fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f45445cbf1d180d6d4dd9588e85bf678f043bfce8cff0812b78ca3510904f9a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://085fa7b8d4fef8b2722fac8c346107699b5a6bd33fbfc74ed35831cb8fdf9dad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c40e018a0979035827963efb9efc48a878a55b9df68ec1a9242d7cff64fe480\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8356f76fbc3a4e636c12d1da5455d3f6767d9b83ce2f2925a11435d7d7261e9f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0126 16:47:26.205208 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0126 16:47:26.206538 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3645934156/tls.crt::/tmp/serving-cert-3645934156/tls.key\\\\\\\"\\\\nI0126 16:47:31.504518 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0126 16:47:31.509366 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0126 16:47:31.509400 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0126 16:47:31.509428 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0126 16:47:31.509436 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0126 16:47:31.515912 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0126 16:47:31.516036 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 16:47:31.516068 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 16:47:31.516096 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0126 16:47:31.516121 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0126 16:47:31.516145 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0126 16:47:31.516170 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0126 16:47:31.516412 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0126 16:47:31.519294 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27b74e2e65a547e0e28fddd4f15feacb17f4d57596eeb8da82df9d39243c85e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22ba7649a1308815bfad52016000987a7a9a5bc214d7ff61e137ba099b20fc6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://22ba7649a1308815bfad52016000987a7a9a5bc214d7ff61e137ba099b20fc6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:13Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:24Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:24 crc kubenswrapper[4754]: I0126 16:48:24.548746 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://205ec256e46a8e08b4b0b0805fb6c5f4966fc3ccef0738df4b97cb404e74da17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:24Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:24 crc kubenswrapper[4754]: I0126 16:48:24.566224 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:24Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:24 crc kubenswrapper[4754]: I0126 16:48:24.578558 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c3718a4-f354-4284-92e0-fdfb45a692bd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b86f46f49163bca7eca9458279029af330dbd4890b03325e06fa3b1bdf030de6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxjcz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fffc03b536bfee3c7bc1bddd500e834e34b391896477a1501e3e06201a19374\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxjcz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-x65wv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:24Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:24 crc kubenswrapper[4754]: I0126 16:48:24.589880 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-t4ps6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d45179f0-9f81-4d77-8805-6b05bdb32f8f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vk8tg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vk8tg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:48Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-t4ps6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:24Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:24 crc kubenswrapper[4754]: I0126 16:48:24.629488 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:24 crc kubenswrapper[4754]: I0126 16:48:24.629522 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:24 crc kubenswrapper[4754]: I0126 16:48:24.629534 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:24 crc kubenswrapper[4754]: I0126 16:48:24.629551 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:24 crc kubenswrapper[4754]: I0126 16:48:24.629563 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:24Z","lastTransitionTime":"2026-01-26T16:48:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:24 crc kubenswrapper[4754]: I0126 16:48:24.731820 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:24 crc kubenswrapper[4754]: I0126 16:48:24.731891 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:24 crc kubenswrapper[4754]: I0126 16:48:24.731907 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:24 crc kubenswrapper[4754]: I0126 16:48:24.731936 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:24 crc kubenswrapper[4754]: I0126 16:48:24.731952 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:24Z","lastTransitionTime":"2026-01-26T16:48:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:24 crc kubenswrapper[4754]: I0126 16:48:24.834508 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:24 crc kubenswrapper[4754]: I0126 16:48:24.834540 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:24 crc kubenswrapper[4754]: I0126 16:48:24.834553 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:24 crc kubenswrapper[4754]: I0126 16:48:24.834572 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:24 crc kubenswrapper[4754]: I0126 16:48:24.834585 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:24Z","lastTransitionTime":"2026-01-26T16:48:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:24 crc kubenswrapper[4754]: I0126 16:48:24.936643 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:24 crc kubenswrapper[4754]: I0126 16:48:24.936701 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:24 crc kubenswrapper[4754]: I0126 16:48:24.936713 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:24 crc kubenswrapper[4754]: I0126 16:48:24.936729 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:24 crc kubenswrapper[4754]: I0126 16:48:24.936739 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:24Z","lastTransitionTime":"2026-01-26T16:48:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:25 crc kubenswrapper[4754]: I0126 16:48:25.039518 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:25 crc kubenswrapper[4754]: I0126 16:48:25.039563 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:25 crc kubenswrapper[4754]: I0126 16:48:25.039572 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:25 crc kubenswrapper[4754]: I0126 16:48:25.039587 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:25 crc kubenswrapper[4754]: I0126 16:48:25.039596 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:25Z","lastTransitionTime":"2026-01-26T16:48:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:25 crc kubenswrapper[4754]: I0126 16:48:25.142075 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:25 crc kubenswrapper[4754]: I0126 16:48:25.142112 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:25 crc kubenswrapper[4754]: I0126 16:48:25.142120 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:25 crc kubenswrapper[4754]: I0126 16:48:25.142132 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:25 crc kubenswrapper[4754]: I0126 16:48:25.142140 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:25Z","lastTransitionTime":"2026-01-26T16:48:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:25 crc kubenswrapper[4754]: I0126 16:48:25.244724 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:25 crc kubenswrapper[4754]: I0126 16:48:25.244775 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:25 crc kubenswrapper[4754]: I0126 16:48:25.244812 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:25 crc kubenswrapper[4754]: I0126 16:48:25.244826 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:25 crc kubenswrapper[4754]: I0126 16:48:25.244838 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:25Z","lastTransitionTime":"2026-01-26T16:48:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:25 crc kubenswrapper[4754]: I0126 16:48:25.286038 4754 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-23 03:01:09.770334901 +0000 UTC Jan 26 16:48:25 crc kubenswrapper[4754]: I0126 16:48:25.322093 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-jsbxt_689569c4-93ee-4b82-93fc-9af0f97e6dc3/ovnkube-controller/3.log" Jan 26 16:48:25 crc kubenswrapper[4754]: I0126 16:48:25.326712 4754 scope.go:117] "RemoveContainer" containerID="2c059b7d618442c8e541bce1d2e576de3e9d9f9b3bd82973f0039fc6c8429d6e" Jan 26 16:48:25 crc kubenswrapper[4754]: E0126 16:48:25.326866 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-jsbxt_openshift-ovn-kubernetes(689569c4-93ee-4b82-93fc-9af0f97e6dc3)\"" pod="openshift-ovn-kubernetes/ovnkube-node-jsbxt" podUID="689569c4-93ee-4b82-93fc-9af0f97e6dc3" Jan 26 16:48:25 crc kubenswrapper[4754]: I0126 16:48:25.342546 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://205ec256e46a8e08b4b0b0805fb6c5f4966fc3ccef0738df4b97cb404e74da17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:25Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:25 crc kubenswrapper[4754]: I0126 16:48:25.348113 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:25 crc kubenswrapper[4754]: I0126 16:48:25.348166 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:25 crc kubenswrapper[4754]: I0126 16:48:25.348177 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:25 crc kubenswrapper[4754]: I0126 16:48:25.348197 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:25 crc kubenswrapper[4754]: I0126 16:48:25.348208 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:25Z","lastTransitionTime":"2026-01-26T16:48:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:25 crc kubenswrapper[4754]: I0126 16:48:25.359115 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:25Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:25 crc kubenswrapper[4754]: I0126 16:48:25.374344 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c3718a4-f354-4284-92e0-fdfb45a692bd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b86f46f49163bca7eca9458279029af330dbd4890b03325e06fa3b1bdf030de6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxjcz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fffc03b536bfee3c7bc1bddd500e834e34b391896477a1501e3e06201a19374\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxjcz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-x65wv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:25Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:25 crc kubenswrapper[4754]: I0126 16:48:25.388384 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-t4ps6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d45179f0-9f81-4d77-8805-6b05bdb32f8f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vk8tg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vk8tg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:48Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-t4ps6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:25Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:25 crc kubenswrapper[4754]: I0126 16:48:25.405733 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b89aa081-0a2e-4757-9210-e2a6c228bee9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9c5f40188fda14b14128514fe84bddfe8dcede3d6c560283ccb2ac0c66895fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f45445cbf1d180d6d4dd9588e85bf678f043bfce8cff0812b78ca3510904f9a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://085fa7b8d4fef8b2722fac8c346107699b5a6bd33fbfc74ed35831cb8fdf9dad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c40e018a0979035827963efb9efc48a878a55b9df68ec1a9242d7cff64fe480\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8356f76fbc3a4e636c12d1da5455d3f6767d9b83ce2f2925a11435d7d7261e9f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0126 16:47:26.205208 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0126 16:47:26.206538 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3645934156/tls.crt::/tmp/serving-cert-3645934156/tls.key\\\\\\\"\\\\nI0126 16:47:31.504518 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0126 16:47:31.509366 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0126 16:47:31.509400 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0126 16:47:31.509428 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0126 16:47:31.509436 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0126 16:47:31.515912 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0126 16:47:31.516036 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 16:47:31.516068 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 16:47:31.516096 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0126 16:47:31.516121 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0126 16:47:31.516145 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0126 16:47:31.516170 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0126 16:47:31.516412 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0126 16:47:31.519294 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27b74e2e65a547e0e28fddd4f15feacb17f4d57596eeb8da82df9d39243c85e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22ba7649a1308815bfad52016000987a7a9a5bc214d7ff61e137ba099b20fc6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://22ba7649a1308815bfad52016000987a7a9a5bc214d7ff61e137ba099b20fc6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:13Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:25Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:25 crc kubenswrapper[4754]: I0126 16:48:25.421977 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:25Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:25 crc kubenswrapper[4754]: I0126 16:48:25.440038 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e459fd8c1093d6f3d0a671751b792b26b7df94c21ced1be1a2250821bcc82d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:25Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:25 crc kubenswrapper[4754]: I0126 16:48:25.451382 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:25 crc kubenswrapper[4754]: I0126 16:48:25.451445 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:25 crc kubenswrapper[4754]: I0126 16:48:25.451467 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:25 crc kubenswrapper[4754]: I0126 16:48:25.451493 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:25 crc kubenswrapper[4754]: I0126 16:48:25.451507 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:25Z","lastTransitionTime":"2026-01-26T16:48:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:25 crc kubenswrapper[4754]: I0126 16:48:25.453095 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2blzw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca65468e-30c6-4666-962e-cc0de05e67fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4603235bb2af495901926b805fe1c0ffc48824a4394ccf928e7fcd53a673ac3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c25z4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:33Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2blzw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:25Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:25 crc kubenswrapper[4754]: I0126 16:48:25.470217 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bwpd6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"351cc6c0-b8e5-4589-ae63-f98e8ab1eca3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1022f101d36873b41bbc1ed8a4a2590347b49f7cce6a172aa1d5bc9f529aed65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3ce0fe080a094316b9456057e347b90a5ff432daa12b9c56d7aea9bee1f72b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3ce0fe080a094316b9456057e347b90a5ff432daa12b9c56d7aea9bee1f72b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1772bb41eb8a36a863d3c691610bcbee5d16b260480b3e9f9d6b083e09e42799\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1772bb41eb8a36a863d3c691610bcbee5d16b260480b3e9f9d6b083e09e42799\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24848d69756024cc664affb9492764ace0638ef611f910f4f2bf86b638cc443d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://24848d69756024cc664affb9492764ace0638ef611f910f4f2bf86b638cc443d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c264fc7c3300a725566acce4464f01d9314779737624401c164d867831fd93bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c264fc7c3300a725566acce4464f01d9314779737624401c164d867831fd93bf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5acfe327eee4aad927ac3a32d24549eff788b14a1c4bdac359b6b56301d884ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5acfe327eee4aad927ac3a32d24549eff788b14a1c4bdac359b6b56301d884ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://63a9b9317494d0e0b33970ef3d461729b10d2a1ad7d3d2bd33f389b184ca56ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://63a9b9317494d0e0b33970ef3d461729b10d2a1ad7d3d2bd33f389b184ca56ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bwpd6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:25Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:25 crc kubenswrapper[4754]: I0126 16:48:25.491396 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jsbxt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"689569c4-93ee-4b82-93fc-9af0f97e6dc3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://033f8b104d4e9ff6d5f87ca998665f12e0c56696b3d30e01ede05f4da540365d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://919572e731813a98c7d8a94879bba931f75c472c7cd681e86d5197ebab43815d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1ac1f49bd564750fcff530a0f76657b0ed41573539acd020a922d19861ef834\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a300a80ad87f46ba1b4fd7ad3cf6c25767fcfe966b93bdd86d729843f53a5cda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31dcab2f9a2d218b0e49dad1df46377ff4521592824a0fbac254d5b8e6d0632e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fe4d9395447ac570eca1338e749dbefea76295d427a12487b63a142a7d6a2e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c059b7d618442c8e541bce1d2e576de3e9d9f9b3bd82973f0039fc6c8429d6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2c059b7d618442c8e541bce1d2e576de3e9d9f9b3bd82973f0039fc6c8429d6e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-26T16:48:23Z\\\",\\\"message\\\":\\\"*v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0126 16:48:23.624369 6831 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI0126 16:48:23.624638 6831 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI0126 16:48:23.625131 6831 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0126 16:48:23.625229 6831 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0126 16:48:23.625927 6831 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0126 16:48:23.626017 6831 factory.go:656] Stopping watch factory\\\\nI0126 16:48:23.626054 6831 handler.go:208] Removed *v1.Node event handler 2\\\\nI0126 16:48:23.680081 6831 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI0126 16:48:23.680139 6831 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI0126 16:48:23.680356 6831 ovnkube.go:599] Stopped ovnkube\\\\nI0126 16:48:23.680442 6831 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0126 16:48:23.680583 6831 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T16:48:22Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-jsbxt_openshift-ovn-kubernetes(689569c4-93ee-4b82-93fc-9af0f97e6dc3)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://07a43ff7860ff65faf51f4a5e8aeb326f06d6d7222677a92560ae2e3a0c81484\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d9a7bafabd44f8b88a6f62cbe45599af8563bc45d9d7ad3a7669f3d8d11eb9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d9a7bafabd44f8b88a6f62cbe45599af8563bc45d9d7ad3a7669f3d8d11eb9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jsbxt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:25Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:25 crc kubenswrapper[4754]: I0126 16:48:25.508331 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab1ace4c-137b-4b8a-8032-0b86a164f063\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:48:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:48:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7da43edaa2a81f1b647b96c234d795b1ddf5fa23173b78043123f76dc677c517\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e5fbff42fdebad6a74c21d85931d3c8d983973db82fc7f262ba5383bad76172\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f9bf9e4a705cd849595e59e1123ba70773bfa5f8a0d0b79195e4fc7e447ca26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb7a3b85cca4436068c8ceefc3f2358cfa1a2c398a8acecd367b944528b11489\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb7a3b85cca4436068c8ceefc3f2358cfa1a2c398a8acecd367b944528b11489\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:14Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:13Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:25Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:25 crc kubenswrapper[4754]: I0126 16:48:25.531108 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:25Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:25 crc kubenswrapper[4754]: I0126 16:48:25.546217 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89251fa9c64bc25d6d6f355400d7644c099f22aed1cae812fc991a03595f1e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e90508d8a53be9146f9b388750730fdae9ee200306caf5ad393a356390a69ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:25Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:25 crc kubenswrapper[4754]: I0126 16:48:25.554521 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:25 crc kubenswrapper[4754]: I0126 16:48:25.554550 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:25 crc kubenswrapper[4754]: I0126 16:48:25.554559 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:25 crc kubenswrapper[4754]: I0126 16:48:25.554574 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:25 crc kubenswrapper[4754]: I0126 16:48:25.554584 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:25Z","lastTransitionTime":"2026-01-26T16:48:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:25 crc kubenswrapper[4754]: I0126 16:48:25.556935 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-84lgq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6763d962-0892-4acc-bab9-7a4e93e39a85\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://953c1917fca896c3def3fc2c69f9e3dbdc9ca26fc2b3d40282be71f71759fd68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8zlw6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:33Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-84lgq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:25Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:25 crc kubenswrapper[4754]: I0126 16:48:25.570242 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-skh5n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b619d233-b592-4b05-a0b7-dc094c88471e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:48:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:48:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8d026fdb7ea7dc5c40add18606d8f1a7703865edf93ef19edaf811f598364c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9231a4191565e2e0ef53d8a394bb34615c900b601a0dfec24c2d8e5fa7b0b659\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-26T16:48:20Z\\\",\\\"message\\\":\\\"2026-01-26T16:47:34+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_57895685-a7ab-4a6c-9a6a-49247f349eaf\\\\n2026-01-26T16:47:34+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_57895685-a7ab-4a6c-9a6a-49247f349eaf to /host/opt/cni/bin/\\\\n2026-01-26T16:47:35Z [verbose] multus-daemon started\\\\n2026-01-26T16:47:35Z [verbose] Readiness Indicator file check\\\\n2026-01-26T16:48:20Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:48:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-skh5n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:25Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:25 crc kubenswrapper[4754]: I0126 16:48:25.582965 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jspvh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5c81aab0-939e-45ce-b19a-89d4818980c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://356dc54150a495dd6f35e0667c82b807b712b07200d8a0adbb1c92fe3f7af8f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8bps\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://05fd57780c44cd127859dc8b163d20eafbd8b06f3348cf545ef41021ce6bc526\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8bps\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-jspvh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:25Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:25 crc kubenswrapper[4754]: I0126 16:48:25.598032 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"099af702-abf2-4933-b4c9-63ef2326e468\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23017e73f8450839b4c2c02fa61235ffe4604bdce11556def27f25bcadcf89dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a74507c941b9de4d0401c8f1d53bb79ed11359628d3405cdab7f5474a75c862e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4402182b3e2e35a7d5dd4c44ac53e152915b312b7ccb076c4f2b41b925a6e7f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f772a003566198519fe1849fd14f29d4346004b2e655810662507c95b4d5ffa1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:13Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:25Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:25 crc kubenswrapper[4754]: I0126 16:48:25.657556 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:25 crc kubenswrapper[4754]: I0126 16:48:25.657597 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:25 crc kubenswrapper[4754]: I0126 16:48:25.657609 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:25 crc kubenswrapper[4754]: I0126 16:48:25.657629 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:25 crc kubenswrapper[4754]: I0126 16:48:25.657639 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:25Z","lastTransitionTime":"2026-01-26T16:48:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:25 crc kubenswrapper[4754]: I0126 16:48:25.698414 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:25 crc kubenswrapper[4754]: I0126 16:48:25.698443 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:25 crc kubenswrapper[4754]: I0126 16:48:25.698452 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:25 crc kubenswrapper[4754]: I0126 16:48:25.698466 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:25 crc kubenswrapper[4754]: I0126 16:48:25.698475 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:25Z","lastTransitionTime":"2026-01-26T16:48:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:25 crc kubenswrapper[4754]: E0126 16:48:25.710633 4754 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:48:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:48:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:48:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:48:25Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:48:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:48:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:48:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:48:25Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d8506764-e7fa-45cb-a13d-6f527164f548\\\",\\\"systemUUID\\\":\\\"bfbb6c33-ce46-4e81-a8a8-b44409b03821\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:25Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:25 crc kubenswrapper[4754]: I0126 16:48:25.715913 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:25 crc kubenswrapper[4754]: I0126 16:48:25.715959 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:25 crc kubenswrapper[4754]: I0126 16:48:25.715974 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:25 crc kubenswrapper[4754]: I0126 16:48:25.715994 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:25 crc kubenswrapper[4754]: I0126 16:48:25.716009 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:25Z","lastTransitionTime":"2026-01-26T16:48:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:25 crc kubenswrapper[4754]: E0126 16:48:25.728620 4754 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:48:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:48:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:48:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:48:25Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:48:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:48:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:48:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:48:25Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d8506764-e7fa-45cb-a13d-6f527164f548\\\",\\\"systemUUID\\\":\\\"bfbb6c33-ce46-4e81-a8a8-b44409b03821\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:25Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:25 crc kubenswrapper[4754]: I0126 16:48:25.733629 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:25 crc kubenswrapper[4754]: I0126 16:48:25.733687 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:25 crc kubenswrapper[4754]: I0126 16:48:25.733697 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:25 crc kubenswrapper[4754]: I0126 16:48:25.733713 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:25 crc kubenswrapper[4754]: I0126 16:48:25.733722 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:25Z","lastTransitionTime":"2026-01-26T16:48:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:25 crc kubenswrapper[4754]: E0126 16:48:25.745700 4754 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:48:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:48:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:48:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:48:25Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:48:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:48:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:48:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:48:25Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d8506764-e7fa-45cb-a13d-6f527164f548\\\",\\\"systemUUID\\\":\\\"bfbb6c33-ce46-4e81-a8a8-b44409b03821\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:25Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:25 crc kubenswrapper[4754]: I0126 16:48:25.750693 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:25 crc kubenswrapper[4754]: I0126 16:48:25.750741 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:25 crc kubenswrapper[4754]: I0126 16:48:25.750755 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:25 crc kubenswrapper[4754]: I0126 16:48:25.750774 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:25 crc kubenswrapper[4754]: I0126 16:48:25.750786 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:25Z","lastTransitionTime":"2026-01-26T16:48:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:25 crc kubenswrapper[4754]: E0126 16:48:25.763771 4754 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:48:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:48:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:48:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:48:25Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:48:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:48:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:48:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:48:25Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d8506764-e7fa-45cb-a13d-6f527164f548\\\",\\\"systemUUID\\\":\\\"bfbb6c33-ce46-4e81-a8a8-b44409b03821\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:25Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:25 crc kubenswrapper[4754]: I0126 16:48:25.766865 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 16:48:25 crc kubenswrapper[4754]: I0126 16:48:25.766891 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 16:48:25 crc kubenswrapper[4754]: I0126 16:48:25.766999 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-t4ps6" Jan 26 16:48:25 crc kubenswrapper[4754]: E0126 16:48:25.767034 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 16:48:25 crc kubenswrapper[4754]: E0126 16:48:25.767183 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 16:48:25 crc kubenswrapper[4754]: I0126 16:48:25.767267 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 16:48:25 crc kubenswrapper[4754]: E0126 16:48:25.767371 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-t4ps6" podUID="d45179f0-9f81-4d77-8805-6b05bdb32f8f" Jan 26 16:48:25 crc kubenswrapper[4754]: E0126 16:48:25.767609 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 16:48:25 crc kubenswrapper[4754]: I0126 16:48:25.768602 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:25 crc kubenswrapper[4754]: I0126 16:48:25.768659 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:25 crc kubenswrapper[4754]: I0126 16:48:25.768696 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:25 crc kubenswrapper[4754]: I0126 16:48:25.768717 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:25 crc kubenswrapper[4754]: I0126 16:48:25.768732 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:25Z","lastTransitionTime":"2026-01-26T16:48:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:25 crc kubenswrapper[4754]: E0126 16:48:25.785377 4754 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:48:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:48:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:48:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:48:25Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:48:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:48:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:48:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:48:25Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d8506764-e7fa-45cb-a13d-6f527164f548\\\",\\\"systemUUID\\\":\\\"bfbb6c33-ce46-4e81-a8a8-b44409b03821\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:25Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:25 crc kubenswrapper[4754]: E0126 16:48:25.785576 4754 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Jan 26 16:48:25 crc kubenswrapper[4754]: I0126 16:48:25.789033 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:25 crc kubenswrapper[4754]: I0126 16:48:25.789078 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:25 crc kubenswrapper[4754]: I0126 16:48:25.789093 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:25 crc kubenswrapper[4754]: I0126 16:48:25.789129 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:25 crc kubenswrapper[4754]: I0126 16:48:25.789142 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:25Z","lastTransitionTime":"2026-01-26T16:48:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:25 crc kubenswrapper[4754]: I0126 16:48:25.892164 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:25 crc kubenswrapper[4754]: I0126 16:48:25.892215 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:25 crc kubenswrapper[4754]: I0126 16:48:25.892228 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:25 crc kubenswrapper[4754]: I0126 16:48:25.892247 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:25 crc kubenswrapper[4754]: I0126 16:48:25.892260 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:25Z","lastTransitionTime":"2026-01-26T16:48:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:25 crc kubenswrapper[4754]: I0126 16:48:25.994416 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:25 crc kubenswrapper[4754]: I0126 16:48:25.994475 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:25 crc kubenswrapper[4754]: I0126 16:48:25.994493 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:25 crc kubenswrapper[4754]: I0126 16:48:25.994516 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:25 crc kubenswrapper[4754]: I0126 16:48:25.994533 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:25Z","lastTransitionTime":"2026-01-26T16:48:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:26 crc kubenswrapper[4754]: I0126 16:48:26.097355 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:26 crc kubenswrapper[4754]: I0126 16:48:26.097402 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:26 crc kubenswrapper[4754]: I0126 16:48:26.097412 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:26 crc kubenswrapper[4754]: I0126 16:48:26.097429 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:26 crc kubenswrapper[4754]: I0126 16:48:26.097442 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:26Z","lastTransitionTime":"2026-01-26T16:48:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:26 crc kubenswrapper[4754]: I0126 16:48:26.200502 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:26 crc kubenswrapper[4754]: I0126 16:48:26.200571 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:26 crc kubenswrapper[4754]: I0126 16:48:26.200581 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:26 crc kubenswrapper[4754]: I0126 16:48:26.200601 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:26 crc kubenswrapper[4754]: I0126 16:48:26.200614 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:26Z","lastTransitionTime":"2026-01-26T16:48:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:26 crc kubenswrapper[4754]: I0126 16:48:26.286866 4754 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-28 16:55:56.265813167 +0000 UTC Jan 26 16:48:26 crc kubenswrapper[4754]: I0126 16:48:26.302786 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:26 crc kubenswrapper[4754]: I0126 16:48:26.302824 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:26 crc kubenswrapper[4754]: I0126 16:48:26.302836 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:26 crc kubenswrapper[4754]: I0126 16:48:26.302852 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:26 crc kubenswrapper[4754]: I0126 16:48:26.302864 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:26Z","lastTransitionTime":"2026-01-26T16:48:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:26 crc kubenswrapper[4754]: I0126 16:48:26.405234 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:26 crc kubenswrapper[4754]: I0126 16:48:26.405273 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:26 crc kubenswrapper[4754]: I0126 16:48:26.405283 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:26 crc kubenswrapper[4754]: I0126 16:48:26.405297 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:26 crc kubenswrapper[4754]: I0126 16:48:26.405309 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:26Z","lastTransitionTime":"2026-01-26T16:48:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:26 crc kubenswrapper[4754]: I0126 16:48:26.508048 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:26 crc kubenswrapper[4754]: I0126 16:48:26.508309 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:26 crc kubenswrapper[4754]: I0126 16:48:26.508411 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:26 crc kubenswrapper[4754]: I0126 16:48:26.508483 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:26 crc kubenswrapper[4754]: I0126 16:48:26.508562 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:26Z","lastTransitionTime":"2026-01-26T16:48:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:26 crc kubenswrapper[4754]: I0126 16:48:26.611763 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:26 crc kubenswrapper[4754]: I0126 16:48:26.611808 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:26 crc kubenswrapper[4754]: I0126 16:48:26.611817 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:26 crc kubenswrapper[4754]: I0126 16:48:26.611833 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:26 crc kubenswrapper[4754]: I0126 16:48:26.611842 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:26Z","lastTransitionTime":"2026-01-26T16:48:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:26 crc kubenswrapper[4754]: I0126 16:48:26.714505 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:26 crc kubenswrapper[4754]: I0126 16:48:26.714545 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:26 crc kubenswrapper[4754]: I0126 16:48:26.714554 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:26 crc kubenswrapper[4754]: I0126 16:48:26.714568 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:26 crc kubenswrapper[4754]: I0126 16:48:26.714577 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:26Z","lastTransitionTime":"2026-01-26T16:48:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:26 crc kubenswrapper[4754]: I0126 16:48:26.821505 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:26 crc kubenswrapper[4754]: I0126 16:48:26.821564 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:26 crc kubenswrapper[4754]: I0126 16:48:26.821579 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:26 crc kubenswrapper[4754]: I0126 16:48:26.821601 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:26 crc kubenswrapper[4754]: I0126 16:48:26.821619 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:26Z","lastTransitionTime":"2026-01-26T16:48:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:26 crc kubenswrapper[4754]: I0126 16:48:26.925075 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:26 crc kubenswrapper[4754]: I0126 16:48:26.925111 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:26 crc kubenswrapper[4754]: I0126 16:48:26.925120 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:26 crc kubenswrapper[4754]: I0126 16:48:26.925134 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:26 crc kubenswrapper[4754]: I0126 16:48:26.925143 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:26Z","lastTransitionTime":"2026-01-26T16:48:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:27 crc kubenswrapper[4754]: I0126 16:48:27.026928 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:27 crc kubenswrapper[4754]: I0126 16:48:27.026976 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:27 crc kubenswrapper[4754]: I0126 16:48:27.026989 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:27 crc kubenswrapper[4754]: I0126 16:48:27.027006 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:27 crc kubenswrapper[4754]: I0126 16:48:27.027018 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:27Z","lastTransitionTime":"2026-01-26T16:48:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:27 crc kubenswrapper[4754]: I0126 16:48:27.130256 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:27 crc kubenswrapper[4754]: I0126 16:48:27.130318 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:27 crc kubenswrapper[4754]: I0126 16:48:27.130330 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:27 crc kubenswrapper[4754]: I0126 16:48:27.130353 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:27 crc kubenswrapper[4754]: I0126 16:48:27.130367 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:27Z","lastTransitionTime":"2026-01-26T16:48:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:27 crc kubenswrapper[4754]: I0126 16:48:27.233501 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:27 crc kubenswrapper[4754]: I0126 16:48:27.233557 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:27 crc kubenswrapper[4754]: I0126 16:48:27.233571 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:27 crc kubenswrapper[4754]: I0126 16:48:27.233591 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:27 crc kubenswrapper[4754]: I0126 16:48:27.233609 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:27Z","lastTransitionTime":"2026-01-26T16:48:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:27 crc kubenswrapper[4754]: I0126 16:48:27.287427 4754 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-01 22:38:53.101206074 +0000 UTC Jan 26 16:48:27 crc kubenswrapper[4754]: I0126 16:48:27.336244 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:27 crc kubenswrapper[4754]: I0126 16:48:27.336304 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:27 crc kubenswrapper[4754]: I0126 16:48:27.336318 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:27 crc kubenswrapper[4754]: I0126 16:48:27.336334 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:27 crc kubenswrapper[4754]: I0126 16:48:27.336346 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:27Z","lastTransitionTime":"2026-01-26T16:48:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:27 crc kubenswrapper[4754]: I0126 16:48:27.438464 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:27 crc kubenswrapper[4754]: I0126 16:48:27.438511 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:27 crc kubenswrapper[4754]: I0126 16:48:27.438522 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:27 crc kubenswrapper[4754]: I0126 16:48:27.438540 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:27 crc kubenswrapper[4754]: I0126 16:48:27.438551 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:27Z","lastTransitionTime":"2026-01-26T16:48:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:27 crc kubenswrapper[4754]: I0126 16:48:27.540644 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:27 crc kubenswrapper[4754]: I0126 16:48:27.540713 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:27 crc kubenswrapper[4754]: I0126 16:48:27.540725 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:27 crc kubenswrapper[4754]: I0126 16:48:27.540740 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:27 crc kubenswrapper[4754]: I0126 16:48:27.540750 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:27Z","lastTransitionTime":"2026-01-26T16:48:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:27 crc kubenswrapper[4754]: I0126 16:48:27.643469 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:27 crc kubenswrapper[4754]: I0126 16:48:27.643524 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:27 crc kubenswrapper[4754]: I0126 16:48:27.643533 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:27 crc kubenswrapper[4754]: I0126 16:48:27.643548 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:27 crc kubenswrapper[4754]: I0126 16:48:27.643558 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:27Z","lastTransitionTime":"2026-01-26T16:48:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:27 crc kubenswrapper[4754]: I0126 16:48:27.746579 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:27 crc kubenswrapper[4754]: I0126 16:48:27.746609 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:27 crc kubenswrapper[4754]: I0126 16:48:27.746619 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:27 crc kubenswrapper[4754]: I0126 16:48:27.746633 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:27 crc kubenswrapper[4754]: I0126 16:48:27.746643 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:27Z","lastTransitionTime":"2026-01-26T16:48:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:27 crc kubenswrapper[4754]: I0126 16:48:27.770496 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 16:48:27 crc kubenswrapper[4754]: E0126 16:48:27.770623 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 16:48:27 crc kubenswrapper[4754]: I0126 16:48:27.770837 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 16:48:27 crc kubenswrapper[4754]: E0126 16:48:27.770908 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 16:48:27 crc kubenswrapper[4754]: I0126 16:48:27.771047 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-t4ps6" Jan 26 16:48:27 crc kubenswrapper[4754]: E0126 16:48:27.771131 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-t4ps6" podUID="d45179f0-9f81-4d77-8805-6b05bdb32f8f" Jan 26 16:48:27 crc kubenswrapper[4754]: I0126 16:48:27.771265 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 16:48:27 crc kubenswrapper[4754]: E0126 16:48:27.771343 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 16:48:27 crc kubenswrapper[4754]: I0126 16:48:27.848416 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:27 crc kubenswrapper[4754]: I0126 16:48:27.848512 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:27 crc kubenswrapper[4754]: I0126 16:48:27.848534 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:27 crc kubenswrapper[4754]: I0126 16:48:27.848562 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:27 crc kubenswrapper[4754]: I0126 16:48:27.848584 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:27Z","lastTransitionTime":"2026-01-26T16:48:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:27 crc kubenswrapper[4754]: I0126 16:48:27.950931 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:27 crc kubenswrapper[4754]: I0126 16:48:27.950977 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:27 crc kubenswrapper[4754]: I0126 16:48:27.950988 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:27 crc kubenswrapper[4754]: I0126 16:48:27.951003 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:27 crc kubenswrapper[4754]: I0126 16:48:27.951013 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:27Z","lastTransitionTime":"2026-01-26T16:48:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:28 crc kubenswrapper[4754]: I0126 16:48:28.052794 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:28 crc kubenswrapper[4754]: I0126 16:48:28.052841 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:28 crc kubenswrapper[4754]: I0126 16:48:28.052854 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:28 crc kubenswrapper[4754]: I0126 16:48:28.052870 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:28 crc kubenswrapper[4754]: I0126 16:48:28.052881 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:28Z","lastTransitionTime":"2026-01-26T16:48:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:28 crc kubenswrapper[4754]: I0126 16:48:28.155051 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:28 crc kubenswrapper[4754]: I0126 16:48:28.155086 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:28 crc kubenswrapper[4754]: I0126 16:48:28.155097 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:28 crc kubenswrapper[4754]: I0126 16:48:28.155112 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:28 crc kubenswrapper[4754]: I0126 16:48:28.155125 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:28Z","lastTransitionTime":"2026-01-26T16:48:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:28 crc kubenswrapper[4754]: I0126 16:48:28.257096 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:28 crc kubenswrapper[4754]: I0126 16:48:28.257154 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:28 crc kubenswrapper[4754]: I0126 16:48:28.257167 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:28 crc kubenswrapper[4754]: I0126 16:48:28.257183 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:28 crc kubenswrapper[4754]: I0126 16:48:28.257195 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:28Z","lastTransitionTime":"2026-01-26T16:48:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:28 crc kubenswrapper[4754]: I0126 16:48:28.288027 4754 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-29 09:27:17.393462677 +0000 UTC Jan 26 16:48:28 crc kubenswrapper[4754]: I0126 16:48:28.359178 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:28 crc kubenswrapper[4754]: I0126 16:48:28.359249 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:28 crc kubenswrapper[4754]: I0126 16:48:28.359261 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:28 crc kubenswrapper[4754]: I0126 16:48:28.359278 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:28 crc kubenswrapper[4754]: I0126 16:48:28.359292 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:28Z","lastTransitionTime":"2026-01-26T16:48:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:28 crc kubenswrapper[4754]: I0126 16:48:28.462799 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:28 crc kubenswrapper[4754]: I0126 16:48:28.462851 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:28 crc kubenswrapper[4754]: I0126 16:48:28.462861 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:28 crc kubenswrapper[4754]: I0126 16:48:28.462880 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:28 crc kubenswrapper[4754]: I0126 16:48:28.462889 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:28Z","lastTransitionTime":"2026-01-26T16:48:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:28 crc kubenswrapper[4754]: I0126 16:48:28.565598 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:28 crc kubenswrapper[4754]: I0126 16:48:28.565715 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:28 crc kubenswrapper[4754]: I0126 16:48:28.565728 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:28 crc kubenswrapper[4754]: I0126 16:48:28.565762 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:28 crc kubenswrapper[4754]: I0126 16:48:28.565773 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:28Z","lastTransitionTime":"2026-01-26T16:48:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:28 crc kubenswrapper[4754]: I0126 16:48:28.667597 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:28 crc kubenswrapper[4754]: I0126 16:48:28.667641 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:28 crc kubenswrapper[4754]: I0126 16:48:28.667655 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:28 crc kubenswrapper[4754]: I0126 16:48:28.667698 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:28 crc kubenswrapper[4754]: I0126 16:48:28.667712 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:28Z","lastTransitionTime":"2026-01-26T16:48:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:28 crc kubenswrapper[4754]: I0126 16:48:28.769919 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:28 crc kubenswrapper[4754]: I0126 16:48:28.769961 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:28 crc kubenswrapper[4754]: I0126 16:48:28.769969 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:28 crc kubenswrapper[4754]: I0126 16:48:28.769981 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:28 crc kubenswrapper[4754]: I0126 16:48:28.769991 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:28Z","lastTransitionTime":"2026-01-26T16:48:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:28 crc kubenswrapper[4754]: I0126 16:48:28.778045 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Jan 26 16:48:28 crc kubenswrapper[4754]: I0126 16:48:28.872816 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:28 crc kubenswrapper[4754]: I0126 16:48:28.872883 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:28 crc kubenswrapper[4754]: I0126 16:48:28.872895 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:28 crc kubenswrapper[4754]: I0126 16:48:28.872911 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:28 crc kubenswrapper[4754]: I0126 16:48:28.872922 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:28Z","lastTransitionTime":"2026-01-26T16:48:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:28 crc kubenswrapper[4754]: I0126 16:48:28.976496 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:28 crc kubenswrapper[4754]: I0126 16:48:28.976537 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:28 crc kubenswrapper[4754]: I0126 16:48:28.976549 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:28 crc kubenswrapper[4754]: I0126 16:48:28.976566 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:28 crc kubenswrapper[4754]: I0126 16:48:28.976577 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:28Z","lastTransitionTime":"2026-01-26T16:48:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:29 crc kubenswrapper[4754]: I0126 16:48:29.078789 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:29 crc kubenswrapper[4754]: I0126 16:48:29.078857 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:29 crc kubenswrapper[4754]: I0126 16:48:29.078872 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:29 crc kubenswrapper[4754]: I0126 16:48:29.078888 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:29 crc kubenswrapper[4754]: I0126 16:48:29.078897 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:29Z","lastTransitionTime":"2026-01-26T16:48:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:29 crc kubenswrapper[4754]: I0126 16:48:29.181743 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:29 crc kubenswrapper[4754]: I0126 16:48:29.181819 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:29 crc kubenswrapper[4754]: I0126 16:48:29.181842 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:29 crc kubenswrapper[4754]: I0126 16:48:29.181874 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:29 crc kubenswrapper[4754]: I0126 16:48:29.181894 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:29Z","lastTransitionTime":"2026-01-26T16:48:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:29 crc kubenswrapper[4754]: I0126 16:48:29.284637 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:29 crc kubenswrapper[4754]: I0126 16:48:29.284784 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:29 crc kubenswrapper[4754]: I0126 16:48:29.284798 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:29 crc kubenswrapper[4754]: I0126 16:48:29.284817 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:29 crc kubenswrapper[4754]: I0126 16:48:29.284829 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:29Z","lastTransitionTime":"2026-01-26T16:48:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:29 crc kubenswrapper[4754]: I0126 16:48:29.289199 4754 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-11 04:18:36.125916566 +0000 UTC Jan 26 16:48:29 crc kubenswrapper[4754]: I0126 16:48:29.387235 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:29 crc kubenswrapper[4754]: I0126 16:48:29.387292 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:29 crc kubenswrapper[4754]: I0126 16:48:29.387305 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:29 crc kubenswrapper[4754]: I0126 16:48:29.387321 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:29 crc kubenswrapper[4754]: I0126 16:48:29.387333 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:29Z","lastTransitionTime":"2026-01-26T16:48:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:29 crc kubenswrapper[4754]: I0126 16:48:29.489636 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:29 crc kubenswrapper[4754]: I0126 16:48:29.489720 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:29 crc kubenswrapper[4754]: I0126 16:48:29.489737 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:29 crc kubenswrapper[4754]: I0126 16:48:29.489761 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:29 crc kubenswrapper[4754]: I0126 16:48:29.489778 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:29Z","lastTransitionTime":"2026-01-26T16:48:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:29 crc kubenswrapper[4754]: I0126 16:48:29.592305 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:29 crc kubenswrapper[4754]: I0126 16:48:29.592358 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:29 crc kubenswrapper[4754]: I0126 16:48:29.592374 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:29 crc kubenswrapper[4754]: I0126 16:48:29.592396 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:29 crc kubenswrapper[4754]: I0126 16:48:29.592411 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:29Z","lastTransitionTime":"2026-01-26T16:48:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:29 crc kubenswrapper[4754]: I0126 16:48:29.694489 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:29 crc kubenswrapper[4754]: I0126 16:48:29.694534 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:29 crc kubenswrapper[4754]: I0126 16:48:29.694544 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:29 crc kubenswrapper[4754]: I0126 16:48:29.694559 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:29 crc kubenswrapper[4754]: I0126 16:48:29.694570 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:29Z","lastTransitionTime":"2026-01-26T16:48:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:29 crc kubenswrapper[4754]: I0126 16:48:29.767230 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 16:48:29 crc kubenswrapper[4754]: I0126 16:48:29.767260 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 16:48:29 crc kubenswrapper[4754]: I0126 16:48:29.767335 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 16:48:29 crc kubenswrapper[4754]: E0126 16:48:29.767443 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 16:48:29 crc kubenswrapper[4754]: I0126 16:48:29.767504 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-t4ps6" Jan 26 16:48:29 crc kubenswrapper[4754]: E0126 16:48:29.767646 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 16:48:29 crc kubenswrapper[4754]: E0126 16:48:29.767769 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-t4ps6" podUID="d45179f0-9f81-4d77-8805-6b05bdb32f8f" Jan 26 16:48:29 crc kubenswrapper[4754]: E0126 16:48:29.767849 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 16:48:29 crc kubenswrapper[4754]: I0126 16:48:29.797189 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:29 crc kubenswrapper[4754]: I0126 16:48:29.797241 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:29 crc kubenswrapper[4754]: I0126 16:48:29.797254 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:29 crc kubenswrapper[4754]: I0126 16:48:29.797272 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:29 crc kubenswrapper[4754]: I0126 16:48:29.797284 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:29Z","lastTransitionTime":"2026-01-26T16:48:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:29 crc kubenswrapper[4754]: I0126 16:48:29.899925 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:29 crc kubenswrapper[4754]: I0126 16:48:29.899964 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:29 crc kubenswrapper[4754]: I0126 16:48:29.899973 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:29 crc kubenswrapper[4754]: I0126 16:48:29.899986 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:29 crc kubenswrapper[4754]: I0126 16:48:29.899995 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:29Z","lastTransitionTime":"2026-01-26T16:48:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:30 crc kubenswrapper[4754]: I0126 16:48:30.002160 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:30 crc kubenswrapper[4754]: I0126 16:48:30.002222 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:30 crc kubenswrapper[4754]: I0126 16:48:30.002232 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:30 crc kubenswrapper[4754]: I0126 16:48:30.002247 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:30 crc kubenswrapper[4754]: I0126 16:48:30.002279 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:30Z","lastTransitionTime":"2026-01-26T16:48:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:30 crc kubenswrapper[4754]: I0126 16:48:30.104563 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:30 crc kubenswrapper[4754]: I0126 16:48:30.104613 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:30 crc kubenswrapper[4754]: I0126 16:48:30.104623 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:30 crc kubenswrapper[4754]: I0126 16:48:30.104638 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:30 crc kubenswrapper[4754]: I0126 16:48:30.104648 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:30Z","lastTransitionTime":"2026-01-26T16:48:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:30 crc kubenswrapper[4754]: I0126 16:48:30.207194 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:30 crc kubenswrapper[4754]: I0126 16:48:30.207255 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:30 crc kubenswrapper[4754]: I0126 16:48:30.207271 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:30 crc kubenswrapper[4754]: I0126 16:48:30.207290 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:30 crc kubenswrapper[4754]: I0126 16:48:30.207305 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:30Z","lastTransitionTime":"2026-01-26T16:48:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:30 crc kubenswrapper[4754]: I0126 16:48:30.289633 4754 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-12 09:12:16.45622438 +0000 UTC Jan 26 16:48:30 crc kubenswrapper[4754]: I0126 16:48:30.310132 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:30 crc kubenswrapper[4754]: I0126 16:48:30.310208 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:30 crc kubenswrapper[4754]: I0126 16:48:30.310231 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:30 crc kubenswrapper[4754]: I0126 16:48:30.310262 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:30 crc kubenswrapper[4754]: I0126 16:48:30.310277 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:30Z","lastTransitionTime":"2026-01-26T16:48:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:30 crc kubenswrapper[4754]: I0126 16:48:30.413008 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:30 crc kubenswrapper[4754]: I0126 16:48:30.413043 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:30 crc kubenswrapper[4754]: I0126 16:48:30.413051 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:30 crc kubenswrapper[4754]: I0126 16:48:30.413066 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:30 crc kubenswrapper[4754]: I0126 16:48:30.413077 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:30Z","lastTransitionTime":"2026-01-26T16:48:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:30 crc kubenswrapper[4754]: I0126 16:48:30.515290 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:30 crc kubenswrapper[4754]: I0126 16:48:30.515319 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:30 crc kubenswrapper[4754]: I0126 16:48:30.515328 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:30 crc kubenswrapper[4754]: I0126 16:48:30.515343 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:30 crc kubenswrapper[4754]: I0126 16:48:30.515352 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:30Z","lastTransitionTime":"2026-01-26T16:48:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:30 crc kubenswrapper[4754]: I0126 16:48:30.617682 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:30 crc kubenswrapper[4754]: I0126 16:48:30.617724 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:30 crc kubenswrapper[4754]: I0126 16:48:30.617733 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:30 crc kubenswrapper[4754]: I0126 16:48:30.617751 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:30 crc kubenswrapper[4754]: I0126 16:48:30.617762 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:30Z","lastTransitionTime":"2026-01-26T16:48:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:30 crc kubenswrapper[4754]: I0126 16:48:30.720003 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:30 crc kubenswrapper[4754]: I0126 16:48:30.720041 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:30 crc kubenswrapper[4754]: I0126 16:48:30.720050 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:30 crc kubenswrapper[4754]: I0126 16:48:30.720066 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:30 crc kubenswrapper[4754]: I0126 16:48:30.720075 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:30Z","lastTransitionTime":"2026-01-26T16:48:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:30 crc kubenswrapper[4754]: I0126 16:48:30.823039 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:30 crc kubenswrapper[4754]: I0126 16:48:30.823099 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:30 crc kubenswrapper[4754]: I0126 16:48:30.823108 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:30 crc kubenswrapper[4754]: I0126 16:48:30.823135 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:30 crc kubenswrapper[4754]: I0126 16:48:30.823145 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:30Z","lastTransitionTime":"2026-01-26T16:48:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:30 crc kubenswrapper[4754]: I0126 16:48:30.925713 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:30 crc kubenswrapper[4754]: I0126 16:48:30.925746 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:30 crc kubenswrapper[4754]: I0126 16:48:30.925754 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:30 crc kubenswrapper[4754]: I0126 16:48:30.925769 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:30 crc kubenswrapper[4754]: I0126 16:48:30.925786 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:30Z","lastTransitionTime":"2026-01-26T16:48:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:31 crc kubenswrapper[4754]: I0126 16:48:31.028204 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:31 crc kubenswrapper[4754]: I0126 16:48:31.028538 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:31 crc kubenswrapper[4754]: I0126 16:48:31.028681 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:31 crc kubenswrapper[4754]: I0126 16:48:31.028770 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:31 crc kubenswrapper[4754]: I0126 16:48:31.028844 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:31Z","lastTransitionTime":"2026-01-26T16:48:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:31 crc kubenswrapper[4754]: I0126 16:48:31.131694 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:31 crc kubenswrapper[4754]: I0126 16:48:31.131751 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:31 crc kubenswrapper[4754]: I0126 16:48:31.131764 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:31 crc kubenswrapper[4754]: I0126 16:48:31.131782 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:31 crc kubenswrapper[4754]: I0126 16:48:31.131793 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:31Z","lastTransitionTime":"2026-01-26T16:48:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:31 crc kubenswrapper[4754]: I0126 16:48:31.234719 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:31 crc kubenswrapper[4754]: I0126 16:48:31.234759 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:31 crc kubenswrapper[4754]: I0126 16:48:31.234770 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:31 crc kubenswrapper[4754]: I0126 16:48:31.234785 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:31 crc kubenswrapper[4754]: I0126 16:48:31.234793 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:31Z","lastTransitionTime":"2026-01-26T16:48:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:31 crc kubenswrapper[4754]: I0126 16:48:31.289801 4754 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-10 15:50:43.969344236 +0000 UTC Jan 26 16:48:31 crc kubenswrapper[4754]: I0126 16:48:31.337852 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:31 crc kubenswrapper[4754]: I0126 16:48:31.337935 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:31 crc kubenswrapper[4754]: I0126 16:48:31.337947 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:31 crc kubenswrapper[4754]: I0126 16:48:31.337964 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:31 crc kubenswrapper[4754]: I0126 16:48:31.337977 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:31Z","lastTransitionTime":"2026-01-26T16:48:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:31 crc kubenswrapper[4754]: I0126 16:48:31.440826 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:31 crc kubenswrapper[4754]: I0126 16:48:31.440887 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:31 crc kubenswrapper[4754]: I0126 16:48:31.440901 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:31 crc kubenswrapper[4754]: I0126 16:48:31.440921 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:31 crc kubenswrapper[4754]: I0126 16:48:31.440935 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:31Z","lastTransitionTime":"2026-01-26T16:48:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:31 crc kubenswrapper[4754]: I0126 16:48:31.543187 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:31 crc kubenswrapper[4754]: I0126 16:48:31.543225 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:31 crc kubenswrapper[4754]: I0126 16:48:31.543234 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:31 crc kubenswrapper[4754]: I0126 16:48:31.543248 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:31 crc kubenswrapper[4754]: I0126 16:48:31.543258 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:31Z","lastTransitionTime":"2026-01-26T16:48:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:31 crc kubenswrapper[4754]: I0126 16:48:31.645505 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:31 crc kubenswrapper[4754]: I0126 16:48:31.645544 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:31 crc kubenswrapper[4754]: I0126 16:48:31.645553 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:31 crc kubenswrapper[4754]: I0126 16:48:31.645567 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:31 crc kubenswrapper[4754]: I0126 16:48:31.645578 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:31Z","lastTransitionTime":"2026-01-26T16:48:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:31 crc kubenswrapper[4754]: I0126 16:48:31.748061 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:31 crc kubenswrapper[4754]: I0126 16:48:31.748107 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:31 crc kubenswrapper[4754]: I0126 16:48:31.748118 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:31 crc kubenswrapper[4754]: I0126 16:48:31.748134 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:31 crc kubenswrapper[4754]: I0126 16:48:31.748146 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:31Z","lastTransitionTime":"2026-01-26T16:48:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:31 crc kubenswrapper[4754]: I0126 16:48:31.767436 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-t4ps6" Jan 26 16:48:31 crc kubenswrapper[4754]: I0126 16:48:31.767492 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 16:48:31 crc kubenswrapper[4754]: I0126 16:48:31.767523 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 16:48:31 crc kubenswrapper[4754]: I0126 16:48:31.767548 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 16:48:31 crc kubenswrapper[4754]: E0126 16:48:31.767706 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-t4ps6" podUID="d45179f0-9f81-4d77-8805-6b05bdb32f8f" Jan 26 16:48:31 crc kubenswrapper[4754]: E0126 16:48:31.767776 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 16:48:31 crc kubenswrapper[4754]: E0126 16:48:31.767843 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 16:48:31 crc kubenswrapper[4754]: E0126 16:48:31.767916 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 16:48:31 crc kubenswrapper[4754]: I0126 16:48:31.850381 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:31 crc kubenswrapper[4754]: I0126 16:48:31.850430 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:31 crc kubenswrapper[4754]: I0126 16:48:31.850444 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:31 crc kubenswrapper[4754]: I0126 16:48:31.850477 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:31 crc kubenswrapper[4754]: I0126 16:48:31.850494 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:31Z","lastTransitionTime":"2026-01-26T16:48:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:31 crc kubenswrapper[4754]: I0126 16:48:31.952913 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:31 crc kubenswrapper[4754]: I0126 16:48:31.952959 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:31 crc kubenswrapper[4754]: I0126 16:48:31.952969 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:31 crc kubenswrapper[4754]: I0126 16:48:31.952982 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:31 crc kubenswrapper[4754]: I0126 16:48:31.952991 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:31Z","lastTransitionTime":"2026-01-26T16:48:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:32 crc kubenswrapper[4754]: I0126 16:48:32.056515 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:32 crc kubenswrapper[4754]: I0126 16:48:32.056582 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:32 crc kubenswrapper[4754]: I0126 16:48:32.056604 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:32 crc kubenswrapper[4754]: I0126 16:48:32.056632 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:32 crc kubenswrapper[4754]: I0126 16:48:32.056654 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:32Z","lastTransitionTime":"2026-01-26T16:48:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:32 crc kubenswrapper[4754]: I0126 16:48:32.159123 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:32 crc kubenswrapper[4754]: I0126 16:48:32.159168 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:32 crc kubenswrapper[4754]: I0126 16:48:32.159178 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:32 crc kubenswrapper[4754]: I0126 16:48:32.159193 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:32 crc kubenswrapper[4754]: I0126 16:48:32.159203 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:32Z","lastTransitionTime":"2026-01-26T16:48:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:32 crc kubenswrapper[4754]: I0126 16:48:32.261858 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:32 crc kubenswrapper[4754]: I0126 16:48:32.261902 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:32 crc kubenswrapper[4754]: I0126 16:48:32.261915 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:32 crc kubenswrapper[4754]: I0126 16:48:32.261930 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:32 crc kubenswrapper[4754]: I0126 16:48:32.261942 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:32Z","lastTransitionTime":"2026-01-26T16:48:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:32 crc kubenswrapper[4754]: I0126 16:48:32.290939 4754 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-26 08:37:21.600842102 +0000 UTC Jan 26 16:48:32 crc kubenswrapper[4754]: I0126 16:48:32.363923 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:32 crc kubenswrapper[4754]: I0126 16:48:32.363955 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:32 crc kubenswrapper[4754]: I0126 16:48:32.363965 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:32 crc kubenswrapper[4754]: I0126 16:48:32.363977 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:32 crc kubenswrapper[4754]: I0126 16:48:32.363986 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:32Z","lastTransitionTime":"2026-01-26T16:48:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:32 crc kubenswrapper[4754]: I0126 16:48:32.467404 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:32 crc kubenswrapper[4754]: I0126 16:48:32.467445 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:32 crc kubenswrapper[4754]: I0126 16:48:32.467456 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:32 crc kubenswrapper[4754]: I0126 16:48:32.467471 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:32 crc kubenswrapper[4754]: I0126 16:48:32.467481 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:32Z","lastTransitionTime":"2026-01-26T16:48:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:32 crc kubenswrapper[4754]: I0126 16:48:32.570782 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:32 crc kubenswrapper[4754]: I0126 16:48:32.570850 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:32 crc kubenswrapper[4754]: I0126 16:48:32.570867 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:32 crc kubenswrapper[4754]: I0126 16:48:32.570893 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:32 crc kubenswrapper[4754]: I0126 16:48:32.570910 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:32Z","lastTransitionTime":"2026-01-26T16:48:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:32 crc kubenswrapper[4754]: I0126 16:48:32.672952 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:32 crc kubenswrapper[4754]: I0126 16:48:32.673033 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:32 crc kubenswrapper[4754]: I0126 16:48:32.673056 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:32 crc kubenswrapper[4754]: I0126 16:48:32.673083 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:32 crc kubenswrapper[4754]: I0126 16:48:32.673102 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:32Z","lastTransitionTime":"2026-01-26T16:48:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:32 crc kubenswrapper[4754]: I0126 16:48:32.775341 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:32 crc kubenswrapper[4754]: I0126 16:48:32.775389 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:32 crc kubenswrapper[4754]: I0126 16:48:32.775401 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:32 crc kubenswrapper[4754]: I0126 16:48:32.775419 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:32 crc kubenswrapper[4754]: I0126 16:48:32.775432 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:32Z","lastTransitionTime":"2026-01-26T16:48:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:32 crc kubenswrapper[4754]: I0126 16:48:32.878759 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:32 crc kubenswrapper[4754]: I0126 16:48:32.878811 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:32 crc kubenswrapper[4754]: I0126 16:48:32.878822 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:32 crc kubenswrapper[4754]: I0126 16:48:32.878842 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:32 crc kubenswrapper[4754]: I0126 16:48:32.878854 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:32Z","lastTransitionTime":"2026-01-26T16:48:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:32 crc kubenswrapper[4754]: I0126 16:48:32.981386 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:32 crc kubenswrapper[4754]: I0126 16:48:32.981445 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:32 crc kubenswrapper[4754]: I0126 16:48:32.981464 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:32 crc kubenswrapper[4754]: I0126 16:48:32.981484 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:32 crc kubenswrapper[4754]: I0126 16:48:32.981495 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:32Z","lastTransitionTime":"2026-01-26T16:48:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:33 crc kubenswrapper[4754]: I0126 16:48:33.083968 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:33 crc kubenswrapper[4754]: I0126 16:48:33.084036 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:33 crc kubenswrapper[4754]: I0126 16:48:33.084046 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:33 crc kubenswrapper[4754]: I0126 16:48:33.084063 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:33 crc kubenswrapper[4754]: I0126 16:48:33.084074 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:33Z","lastTransitionTime":"2026-01-26T16:48:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:33 crc kubenswrapper[4754]: I0126 16:48:33.187080 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:33 crc kubenswrapper[4754]: I0126 16:48:33.187139 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:33 crc kubenswrapper[4754]: I0126 16:48:33.187151 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:33 crc kubenswrapper[4754]: I0126 16:48:33.187168 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:33 crc kubenswrapper[4754]: I0126 16:48:33.187181 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:33Z","lastTransitionTime":"2026-01-26T16:48:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:33 crc kubenswrapper[4754]: I0126 16:48:33.289300 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:33 crc kubenswrapper[4754]: I0126 16:48:33.289370 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:33 crc kubenswrapper[4754]: I0126 16:48:33.289387 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:33 crc kubenswrapper[4754]: I0126 16:48:33.289412 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:33 crc kubenswrapper[4754]: I0126 16:48:33.289427 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:33Z","lastTransitionTime":"2026-01-26T16:48:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:33 crc kubenswrapper[4754]: I0126 16:48:33.291738 4754 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-15 10:28:44.559546086 +0000 UTC Jan 26 16:48:33 crc kubenswrapper[4754]: I0126 16:48:33.393133 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:33 crc kubenswrapper[4754]: I0126 16:48:33.393178 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:33 crc kubenswrapper[4754]: I0126 16:48:33.393189 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:33 crc kubenswrapper[4754]: I0126 16:48:33.393210 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:33 crc kubenswrapper[4754]: I0126 16:48:33.393223 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:33Z","lastTransitionTime":"2026-01-26T16:48:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:33 crc kubenswrapper[4754]: I0126 16:48:33.495820 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:33 crc kubenswrapper[4754]: I0126 16:48:33.495859 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:33 crc kubenswrapper[4754]: I0126 16:48:33.495869 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:33 crc kubenswrapper[4754]: I0126 16:48:33.495885 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:33 crc kubenswrapper[4754]: I0126 16:48:33.495894 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:33Z","lastTransitionTime":"2026-01-26T16:48:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:33 crc kubenswrapper[4754]: I0126 16:48:33.598104 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:33 crc kubenswrapper[4754]: I0126 16:48:33.598157 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:33 crc kubenswrapper[4754]: I0126 16:48:33.598171 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:33 crc kubenswrapper[4754]: I0126 16:48:33.598190 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:33 crc kubenswrapper[4754]: I0126 16:48:33.598204 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:33Z","lastTransitionTime":"2026-01-26T16:48:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:33 crc kubenswrapper[4754]: I0126 16:48:33.700596 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:33 crc kubenswrapper[4754]: I0126 16:48:33.700646 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:33 crc kubenswrapper[4754]: I0126 16:48:33.700656 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:33 crc kubenswrapper[4754]: I0126 16:48:33.700687 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:33 crc kubenswrapper[4754]: I0126 16:48:33.700697 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:33Z","lastTransitionTime":"2026-01-26T16:48:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:33 crc kubenswrapper[4754]: I0126 16:48:33.766390 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 16:48:33 crc kubenswrapper[4754]: E0126 16:48:33.766566 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 16:48:33 crc kubenswrapper[4754]: I0126 16:48:33.766608 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 16:48:33 crc kubenswrapper[4754]: I0126 16:48:33.766657 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-t4ps6" Jan 26 16:48:33 crc kubenswrapper[4754]: E0126 16:48:33.766761 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 16:48:33 crc kubenswrapper[4754]: I0126 16:48:33.766795 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 16:48:33 crc kubenswrapper[4754]: E0126 16:48:33.766831 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 16:48:33 crc kubenswrapper[4754]: E0126 16:48:33.766885 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-t4ps6" podUID="d45179f0-9f81-4d77-8805-6b05bdb32f8f" Jan 26 16:48:33 crc kubenswrapper[4754]: I0126 16:48:33.782036 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab1ace4c-137b-4b8a-8032-0b86a164f063\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:48:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:48:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7da43edaa2a81f1b647b96c234d795b1ddf5fa23173b78043123f76dc677c517\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e5fbff42fdebad6a74c21d85931d3c8d983973db82fc7f262ba5383bad76172\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f9bf9e4a705cd849595e59e1123ba70773bfa5f8a0d0b79195e4fc7e447ca26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb7a3b85cca4436068c8ceefc3f2358cfa1a2c398a8acecd367b944528b11489\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb7a3b85cca4436068c8ceefc3f2358cfa1a2c398a8acecd367b944528b11489\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:14Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:13Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:33Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:33 crc kubenswrapper[4754]: I0126 16:48:33.797372 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:33Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:33 crc kubenswrapper[4754]: I0126 16:48:33.807840 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:33 crc kubenswrapper[4754]: I0126 16:48:33.807890 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:33 crc kubenswrapper[4754]: I0126 16:48:33.807903 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:33 crc kubenswrapper[4754]: I0126 16:48:33.807924 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:33 crc kubenswrapper[4754]: I0126 16:48:33.807936 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:33Z","lastTransitionTime":"2026-01-26T16:48:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:33 crc kubenswrapper[4754]: I0126 16:48:33.817928 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89251fa9c64bc25d6d6f355400d7644c099f22aed1cae812fc991a03595f1e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e90508d8a53be9146f9b388750730fdae9ee200306caf5ad393a356390a69ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:33Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:33 crc kubenswrapper[4754]: I0126 16:48:33.831577 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-84lgq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6763d962-0892-4acc-bab9-7a4e93e39a85\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://953c1917fca896c3def3fc2c69f9e3dbdc9ca26fc2b3d40282be71f71759fd68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8zlw6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:33Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-84lgq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:33Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:33 crc kubenswrapper[4754]: I0126 16:48:33.847127 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-skh5n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b619d233-b592-4b05-a0b7-dc094c88471e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:48:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:48:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8d026fdb7ea7dc5c40add18606d8f1a7703865edf93ef19edaf811f598364c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9231a4191565e2e0ef53d8a394bb34615c900b601a0dfec24c2d8e5fa7b0b659\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-26T16:48:20Z\\\",\\\"message\\\":\\\"2026-01-26T16:47:34+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_57895685-a7ab-4a6c-9a6a-49247f349eaf\\\\n2026-01-26T16:47:34+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_57895685-a7ab-4a6c-9a6a-49247f349eaf to /host/opt/cni/bin/\\\\n2026-01-26T16:47:35Z [verbose] multus-daemon started\\\\n2026-01-26T16:47:35Z [verbose] Readiness Indicator file check\\\\n2026-01-26T16:48:20Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:48:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tj78\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-skh5n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:33Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:33 crc kubenswrapper[4754]: I0126 16:48:33.861389 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jspvh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5c81aab0-939e-45ce-b19a-89d4818980c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://356dc54150a495dd6f35e0667c82b807b712b07200d8a0adbb1c92fe3f7af8f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8bps\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://05fd57780c44cd127859dc8b163d20eafbd8b06f3348cf545ef41021ce6bc526\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8bps\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-jspvh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:33Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:33 crc kubenswrapper[4754]: I0126 16:48:33.877315 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"099af702-abf2-4933-b4c9-63ef2326e468\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23017e73f8450839b4c2c02fa61235ffe4604bdce11556def27f25bcadcf89dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a74507c941b9de4d0401c8f1d53bb79ed11359628d3405cdab7f5474a75c862e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4402182b3e2e35a7d5dd4c44ac53e152915b312b7ccb076c4f2b41b925a6e7f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f772a003566198519fe1849fd14f29d4346004b2e655810662507c95b4d5ffa1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:13Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:33Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:33 crc kubenswrapper[4754]: I0126 16:48:33.893599 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://205ec256e46a8e08b4b0b0805fb6c5f4966fc3ccef0738df4b97cb404e74da17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:33Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:33 crc kubenswrapper[4754]: I0126 16:48:33.909578 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:33Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:33 crc kubenswrapper[4754]: I0126 16:48:33.911029 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:33 crc kubenswrapper[4754]: I0126 16:48:33.911079 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:33 crc kubenswrapper[4754]: I0126 16:48:33.911089 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:33 crc kubenswrapper[4754]: I0126 16:48:33.911118 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:33 crc kubenswrapper[4754]: I0126 16:48:33.911130 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:33Z","lastTransitionTime":"2026-01-26T16:48:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:33 crc kubenswrapper[4754]: I0126 16:48:33.924136 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c3718a4-f354-4284-92e0-fdfb45a692bd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b86f46f49163bca7eca9458279029af330dbd4890b03325e06fa3b1bdf030de6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxjcz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fffc03b536bfee3c7bc1bddd500e834e34b391896477a1501e3e06201a19374\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxjcz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-x65wv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:33Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:33 crc kubenswrapper[4754]: I0126 16:48:33.938800 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-t4ps6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d45179f0-9f81-4d77-8805-6b05bdb32f8f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vk8tg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vk8tg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:48Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-t4ps6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:33Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:33 crc kubenswrapper[4754]: I0126 16:48:33.959166 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b89aa081-0a2e-4757-9210-e2a6c228bee9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9c5f40188fda14b14128514fe84bddfe8dcede3d6c560283ccb2ac0c66895fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f45445cbf1d180d6d4dd9588e85bf678f043bfce8cff0812b78ca3510904f9a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://085fa7b8d4fef8b2722fac8c346107699b5a6bd33fbfc74ed35831cb8fdf9dad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c40e018a0979035827963efb9efc48a878a55b9df68ec1a9242d7cff64fe480\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8356f76fbc3a4e636c12d1da5455d3f6767d9b83ce2f2925a11435d7d7261e9f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0126 16:47:26.205208 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0126 16:47:26.206538 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3645934156/tls.crt::/tmp/serving-cert-3645934156/tls.key\\\\\\\"\\\\nI0126 16:47:31.504518 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0126 16:47:31.509366 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0126 16:47:31.509400 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0126 16:47:31.509428 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0126 16:47:31.509436 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0126 16:47:31.515912 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0126 16:47:31.516036 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 16:47:31.516068 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 16:47:31.516096 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0126 16:47:31.516121 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0126 16:47:31.516145 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0126 16:47:31.516170 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0126 16:47:31.516412 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0126 16:47:31.519294 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27b74e2e65a547e0e28fddd4f15feacb17f4d57596eeb8da82df9d39243c85e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:15Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22ba7649a1308815bfad52016000987a7a9a5bc214d7ff61e137ba099b20fc6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://22ba7649a1308815bfad52016000987a7a9a5bc214d7ff61e137ba099b20fc6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:13Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:33Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:33 crc kubenswrapper[4754]: I0126 16:48:33.971586 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17efad68-c5fe-48a8-b7fa-5638dd3a7af4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://750d9397eeb93fcc4e6bfb4e1417933e459f8e5541fb80b58306dda8b12a5a8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91ba0a944e0cac147aac51868c09fedf144859e7db28a129acd17895172391b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91ba0a944e0cac147aac51868c09fedf144859e7db28a129acd17895172391b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:13Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:33Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:33 crc kubenswrapper[4754]: I0126 16:48:33.988714 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:33Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:33 crc kubenswrapper[4754]: I0126 16:48:33.999145 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e459fd8c1093d6f3d0a671751b792b26b7df94c21ced1be1a2250821bcc82d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:33Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:34 crc kubenswrapper[4754]: I0126 16:48:34.008156 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2blzw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca65468e-30c6-4666-962e-cc0de05e67fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4603235bb2af495901926b805fe1c0ffc48824a4394ccf928e7fcd53a673ac3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c25z4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:33Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2blzw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:34Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:34 crc kubenswrapper[4754]: I0126 16:48:34.013790 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:34 crc kubenswrapper[4754]: I0126 16:48:34.013853 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:34 crc kubenswrapper[4754]: I0126 16:48:34.013866 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:34 crc kubenswrapper[4754]: I0126 16:48:34.013887 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:34 crc kubenswrapper[4754]: I0126 16:48:34.013900 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:34Z","lastTransitionTime":"2026-01-26T16:48:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:34 crc kubenswrapper[4754]: I0126 16:48:34.028136 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bwpd6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"351cc6c0-b8e5-4589-ae63-f98e8ab1eca3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1022f101d36873b41bbc1ed8a4a2590347b49f7cce6a172aa1d5bc9f529aed65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3ce0fe080a094316b9456057e347b90a5ff432daa12b9c56d7aea9bee1f72b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3ce0fe080a094316b9456057e347b90a5ff432daa12b9c56d7aea9bee1f72b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1772bb41eb8a36a863d3c691610bcbee5d16b260480b3e9f9d6b083e09e42799\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1772bb41eb8a36a863d3c691610bcbee5d16b260480b3e9f9d6b083e09e42799\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24848d69756024cc664affb9492764ace0638ef611f910f4f2bf86b638cc443d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://24848d69756024cc664affb9492764ace0638ef611f910f4f2bf86b638cc443d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c264fc7c3300a725566acce4464f01d9314779737624401c164d867831fd93bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c264fc7c3300a725566acce4464f01d9314779737624401c164d867831fd93bf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5acfe327eee4aad927ac3a32d24549eff788b14a1c4bdac359b6b56301d884ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5acfe327eee4aad927ac3a32d24549eff788b14a1c4bdac359b6b56301d884ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://63a9b9317494d0e0b33970ef3d461729b10d2a1ad7d3d2bd33f389b184ca56ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://63a9b9317494d0e0b33970ef3d461729b10d2a1ad7d3d2bd33f389b184ca56ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wm9cg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bwpd6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:34Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:34 crc kubenswrapper[4754]: I0126 16:48:34.052492 4754 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jsbxt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"689569c4-93ee-4b82-93fc-9af0f97e6dc3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://033f8b104d4e9ff6d5f87ca998665f12e0c56696b3d30e01ede05f4da540365d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://919572e731813a98c7d8a94879bba931f75c472c7cd681e86d5197ebab43815d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1ac1f49bd564750fcff530a0f76657b0ed41573539acd020a922d19861ef834\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a300a80ad87f46ba1b4fd7ad3cf6c25767fcfe966b93bdd86d729843f53a5cda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31dcab2f9a2d218b0e49dad1df46377ff4521592824a0fbac254d5b8e6d0632e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fe4d9395447ac570eca1338e749dbefea76295d427a12487b63a142a7d6a2e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c059b7d618442c8e541bce1d2e576de3e9d9f9b3bd82973f0039fc6c8429d6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2c059b7d618442c8e541bce1d2e576de3e9d9f9b3bd82973f0039fc6c8429d6e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-26T16:48:23Z\\\",\\\"message\\\":\\\"*v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0126 16:48:23.624369 6831 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI0126 16:48:23.624638 6831 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI0126 16:48:23.625131 6831 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0126 16:48:23.625229 6831 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0126 16:48:23.625927 6831 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0126 16:48:23.626017 6831 factory.go:656] Stopping watch factory\\\\nI0126 16:48:23.626054 6831 handler.go:208] Removed *v1.Node event handler 2\\\\nI0126 16:48:23.680081 6831 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI0126 16:48:23.680139 6831 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI0126 16:48:23.680356 6831 ovnkube.go:599] Stopped ovnkube\\\\nI0126 16:48:23.680442 6831 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0126 16:48:23.680583 6831 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T16:48:22Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-jsbxt_openshift-ovn-kubernetes(689569c4-93ee-4b82-93fc-9af0f97e6dc3)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://07a43ff7860ff65faf51f4a5e8aeb326f06d6d7222677a92560ae2e3a0c81484\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d9a7bafabd44f8b88a6f62cbe45599af8563bc45d9d7ad3a7669f3d8d11eb9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d9a7bafabd44f8b88a6f62cbe45599af8563bc45d9d7ad3a7669f3d8d11eb9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:47:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:47:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpxwt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:47:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jsbxt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:34Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:34 crc kubenswrapper[4754]: I0126 16:48:34.116661 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:34 crc kubenswrapper[4754]: I0126 16:48:34.116741 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:34 crc kubenswrapper[4754]: I0126 16:48:34.116759 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:34 crc kubenswrapper[4754]: I0126 16:48:34.116779 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:34 crc kubenswrapper[4754]: I0126 16:48:34.116794 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:34Z","lastTransitionTime":"2026-01-26T16:48:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:34 crc kubenswrapper[4754]: I0126 16:48:34.219287 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:34 crc kubenswrapper[4754]: I0126 16:48:34.219321 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:34 crc kubenswrapper[4754]: I0126 16:48:34.219332 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:34 crc kubenswrapper[4754]: I0126 16:48:34.219350 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:34 crc kubenswrapper[4754]: I0126 16:48:34.219361 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:34Z","lastTransitionTime":"2026-01-26T16:48:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:34 crc kubenswrapper[4754]: I0126 16:48:34.292478 4754 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-26 20:11:01.799805958 +0000 UTC Jan 26 16:48:34 crc kubenswrapper[4754]: I0126 16:48:34.322811 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:34 crc kubenswrapper[4754]: I0126 16:48:34.322889 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:34 crc kubenswrapper[4754]: I0126 16:48:34.322903 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:34 crc kubenswrapper[4754]: I0126 16:48:34.322922 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:34 crc kubenswrapper[4754]: I0126 16:48:34.322936 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:34Z","lastTransitionTime":"2026-01-26T16:48:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:34 crc kubenswrapper[4754]: I0126 16:48:34.425528 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:34 crc kubenswrapper[4754]: I0126 16:48:34.425579 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:34 crc kubenswrapper[4754]: I0126 16:48:34.425589 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:34 crc kubenswrapper[4754]: I0126 16:48:34.425609 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:34 crc kubenswrapper[4754]: I0126 16:48:34.425618 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:34Z","lastTransitionTime":"2026-01-26T16:48:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:34 crc kubenswrapper[4754]: I0126 16:48:34.528818 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:34 crc kubenswrapper[4754]: I0126 16:48:34.528864 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:34 crc kubenswrapper[4754]: I0126 16:48:34.528878 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:34 crc kubenswrapper[4754]: I0126 16:48:34.528898 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:34 crc kubenswrapper[4754]: I0126 16:48:34.528911 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:34Z","lastTransitionTime":"2026-01-26T16:48:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:34 crc kubenswrapper[4754]: I0126 16:48:34.630770 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:34 crc kubenswrapper[4754]: I0126 16:48:34.630822 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:34 crc kubenswrapper[4754]: I0126 16:48:34.630836 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:34 crc kubenswrapper[4754]: I0126 16:48:34.630851 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:34 crc kubenswrapper[4754]: I0126 16:48:34.630865 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:34Z","lastTransitionTime":"2026-01-26T16:48:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:34 crc kubenswrapper[4754]: I0126 16:48:34.733613 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:34 crc kubenswrapper[4754]: I0126 16:48:34.733722 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:34 crc kubenswrapper[4754]: I0126 16:48:34.733751 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:34 crc kubenswrapper[4754]: I0126 16:48:34.733780 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:34 crc kubenswrapper[4754]: I0126 16:48:34.733801 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:34Z","lastTransitionTime":"2026-01-26T16:48:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:34 crc kubenswrapper[4754]: I0126 16:48:34.836167 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:34 crc kubenswrapper[4754]: I0126 16:48:34.836226 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:34 crc kubenswrapper[4754]: I0126 16:48:34.836271 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:34 crc kubenswrapper[4754]: I0126 16:48:34.836299 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:34 crc kubenswrapper[4754]: I0126 16:48:34.836315 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:34Z","lastTransitionTime":"2026-01-26T16:48:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:34 crc kubenswrapper[4754]: I0126 16:48:34.938726 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:34 crc kubenswrapper[4754]: I0126 16:48:34.938775 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:34 crc kubenswrapper[4754]: I0126 16:48:34.938795 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:34 crc kubenswrapper[4754]: I0126 16:48:34.938812 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:34 crc kubenswrapper[4754]: I0126 16:48:34.938823 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:34Z","lastTransitionTime":"2026-01-26T16:48:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:35 crc kubenswrapper[4754]: I0126 16:48:35.040864 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:35 crc kubenswrapper[4754]: I0126 16:48:35.040908 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:35 crc kubenswrapper[4754]: I0126 16:48:35.040923 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:35 crc kubenswrapper[4754]: I0126 16:48:35.040943 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:35 crc kubenswrapper[4754]: I0126 16:48:35.040956 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:35Z","lastTransitionTime":"2026-01-26T16:48:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:35 crc kubenswrapper[4754]: I0126 16:48:35.144711 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:35 crc kubenswrapper[4754]: I0126 16:48:35.147732 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:35 crc kubenswrapper[4754]: I0126 16:48:35.147841 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:35 crc kubenswrapper[4754]: I0126 16:48:35.147927 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:35 crc kubenswrapper[4754]: I0126 16:48:35.148180 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:35Z","lastTransitionTime":"2026-01-26T16:48:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:35 crc kubenswrapper[4754]: I0126 16:48:35.252349 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:35 crc kubenswrapper[4754]: I0126 16:48:35.252623 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:35 crc kubenswrapper[4754]: I0126 16:48:35.252631 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:35 crc kubenswrapper[4754]: I0126 16:48:35.252645 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:35 crc kubenswrapper[4754]: I0126 16:48:35.252655 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:35Z","lastTransitionTime":"2026-01-26T16:48:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:35 crc kubenswrapper[4754]: I0126 16:48:35.293528 4754 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-22 23:29:30.779737493 +0000 UTC Jan 26 16:48:35 crc kubenswrapper[4754]: I0126 16:48:35.355326 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:35 crc kubenswrapper[4754]: I0126 16:48:35.355375 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:35 crc kubenswrapper[4754]: I0126 16:48:35.355385 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:35 crc kubenswrapper[4754]: I0126 16:48:35.355401 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:35 crc kubenswrapper[4754]: I0126 16:48:35.355411 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:35Z","lastTransitionTime":"2026-01-26T16:48:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:35 crc kubenswrapper[4754]: I0126 16:48:35.458925 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:35 crc kubenswrapper[4754]: I0126 16:48:35.459029 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:35 crc kubenswrapper[4754]: I0126 16:48:35.459041 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:35 crc kubenswrapper[4754]: I0126 16:48:35.459066 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:35 crc kubenswrapper[4754]: I0126 16:48:35.459082 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:35Z","lastTransitionTime":"2026-01-26T16:48:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:35 crc kubenswrapper[4754]: I0126 16:48:35.561857 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:35 crc kubenswrapper[4754]: I0126 16:48:35.561922 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:35 crc kubenswrapper[4754]: I0126 16:48:35.561939 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:35 crc kubenswrapper[4754]: I0126 16:48:35.561965 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:35 crc kubenswrapper[4754]: I0126 16:48:35.561982 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:35Z","lastTransitionTime":"2026-01-26T16:48:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:35 crc kubenswrapper[4754]: I0126 16:48:35.664388 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:35 crc kubenswrapper[4754]: I0126 16:48:35.664438 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:35 crc kubenswrapper[4754]: I0126 16:48:35.664449 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:35 crc kubenswrapper[4754]: I0126 16:48:35.664464 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:35 crc kubenswrapper[4754]: I0126 16:48:35.664474 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:35Z","lastTransitionTime":"2026-01-26T16:48:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:35 crc kubenswrapper[4754]: I0126 16:48:35.767180 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 16:48:35 crc kubenswrapper[4754]: I0126 16:48:35.767380 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 16:48:35 crc kubenswrapper[4754]: E0126 16:48:35.767487 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 16:48:35 crc kubenswrapper[4754]: I0126 16:48:35.767335 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-t4ps6" Jan 26 16:48:35 crc kubenswrapper[4754]: I0126 16:48:35.767724 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 16:48:35 crc kubenswrapper[4754]: E0126 16:48:35.767725 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 16:48:35 crc kubenswrapper[4754]: E0126 16:48:35.767862 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-t4ps6" podUID="d45179f0-9f81-4d77-8805-6b05bdb32f8f" Jan 26 16:48:35 crc kubenswrapper[4754]: E0126 16:48:35.767925 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 16:48:35 crc kubenswrapper[4754]: I0126 16:48:35.767938 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:35 crc kubenswrapper[4754]: I0126 16:48:35.768120 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:35 crc kubenswrapper[4754]: I0126 16:48:35.768203 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:35 crc kubenswrapper[4754]: I0126 16:48:35.768290 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:35 crc kubenswrapper[4754]: I0126 16:48:35.768363 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:35Z","lastTransitionTime":"2026-01-26T16:48:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:35 crc kubenswrapper[4754]: I0126 16:48:35.836362 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 16:48:35 crc kubenswrapper[4754]: I0126 16:48:35.836544 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 16:48:35 crc kubenswrapper[4754]: E0126 16:48:35.836593 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 16:49:39.836558462 +0000 UTC m=+146.360738906 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 16:48:35 crc kubenswrapper[4754]: I0126 16:48:35.836659 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 16:48:35 crc kubenswrapper[4754]: E0126 16:48:35.836777 4754 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 26 16:48:35 crc kubenswrapper[4754]: E0126 16:48:35.836808 4754 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 26 16:48:35 crc kubenswrapper[4754]: E0126 16:48:35.836824 4754 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 26 16:48:35 crc kubenswrapper[4754]: E0126 16:48:35.836902 4754 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Jan 26 16:48:35 crc kubenswrapper[4754]: I0126 16:48:35.836780 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 16:48:35 crc kubenswrapper[4754]: E0126 16:48:35.837031 4754 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 26 16:48:35 crc kubenswrapper[4754]: E0126 16:48:35.836905 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-01-26 16:49:39.83687741 +0000 UTC m=+146.361058055 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 26 16:48:35 crc kubenswrapper[4754]: E0126 16:48:35.837213 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-26 16:49:39.837193689 +0000 UTC m=+146.361374273 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Jan 26 16:48:35 crc kubenswrapper[4754]: I0126 16:48:35.837236 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 16:48:35 crc kubenswrapper[4754]: E0126 16:48:35.837333 4754 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 26 16:48:35 crc kubenswrapper[4754]: E0126 16:48:35.837353 4754 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 26 16:48:35 crc kubenswrapper[4754]: E0126 16:48:35.837367 4754 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 26 16:48:35 crc kubenswrapper[4754]: E0126 16:48:35.837361 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-26 16:49:39.837322422 +0000 UTC m=+146.361503026 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 26 16:48:35 crc kubenswrapper[4754]: E0126 16:48:35.837419 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-01-26 16:49:39.837408424 +0000 UTC m=+146.361588858 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 26 16:48:35 crc kubenswrapper[4754]: I0126 16:48:35.870222 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:35 crc kubenswrapper[4754]: I0126 16:48:35.870279 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:35 crc kubenswrapper[4754]: I0126 16:48:35.870288 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:35 crc kubenswrapper[4754]: I0126 16:48:35.870309 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:35 crc kubenswrapper[4754]: I0126 16:48:35.870319 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:35Z","lastTransitionTime":"2026-01-26T16:48:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:35 crc kubenswrapper[4754]: I0126 16:48:35.973223 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:35 crc kubenswrapper[4754]: I0126 16:48:35.973263 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:35 crc kubenswrapper[4754]: I0126 16:48:35.973273 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:35 crc kubenswrapper[4754]: I0126 16:48:35.973287 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:35 crc kubenswrapper[4754]: I0126 16:48:35.973297 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:35Z","lastTransitionTime":"2026-01-26T16:48:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:36 crc kubenswrapper[4754]: I0126 16:48:36.076284 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:36 crc kubenswrapper[4754]: I0126 16:48:36.076359 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:36 crc kubenswrapper[4754]: I0126 16:48:36.076380 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:36 crc kubenswrapper[4754]: I0126 16:48:36.076406 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:36 crc kubenswrapper[4754]: I0126 16:48:36.076426 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:36Z","lastTransitionTime":"2026-01-26T16:48:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:36 crc kubenswrapper[4754]: I0126 16:48:36.161347 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:36 crc kubenswrapper[4754]: I0126 16:48:36.161388 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:36 crc kubenswrapper[4754]: I0126 16:48:36.161403 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:36 crc kubenswrapper[4754]: I0126 16:48:36.161423 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:36 crc kubenswrapper[4754]: I0126 16:48:36.161436 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:36Z","lastTransitionTime":"2026-01-26T16:48:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:36 crc kubenswrapper[4754]: E0126 16:48:36.175863 4754 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:48:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:48:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:48:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:48:36Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:48:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:48:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:48:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:48:36Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d8506764-e7fa-45cb-a13d-6f527164f548\\\",\\\"systemUUID\\\":\\\"bfbb6c33-ce46-4e81-a8a8-b44409b03821\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:36Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:36 crc kubenswrapper[4754]: I0126 16:48:36.179258 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:36 crc kubenswrapper[4754]: I0126 16:48:36.179298 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:36 crc kubenswrapper[4754]: I0126 16:48:36.179308 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:36 crc kubenswrapper[4754]: I0126 16:48:36.179324 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:36 crc kubenswrapper[4754]: I0126 16:48:36.179335 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:36Z","lastTransitionTime":"2026-01-26T16:48:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:36 crc kubenswrapper[4754]: E0126 16:48:36.193190 4754 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:48:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:48:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:48:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:48:36Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:48:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:48:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:48:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:48:36Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d8506764-e7fa-45cb-a13d-6f527164f548\\\",\\\"systemUUID\\\":\\\"bfbb6c33-ce46-4e81-a8a8-b44409b03821\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:36Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:36 crc kubenswrapper[4754]: I0126 16:48:36.198554 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:36 crc kubenswrapper[4754]: I0126 16:48:36.198600 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:36 crc kubenswrapper[4754]: I0126 16:48:36.198612 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:36 crc kubenswrapper[4754]: I0126 16:48:36.198629 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:36 crc kubenswrapper[4754]: I0126 16:48:36.198642 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:36Z","lastTransitionTime":"2026-01-26T16:48:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:36 crc kubenswrapper[4754]: E0126 16:48:36.211435 4754 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:48:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:48:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:48:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:48:36Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:48:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:48:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:48:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:48:36Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d8506764-e7fa-45cb-a13d-6f527164f548\\\",\\\"systemUUID\\\":\\\"bfbb6c33-ce46-4e81-a8a8-b44409b03821\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:36Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:36 crc kubenswrapper[4754]: I0126 16:48:36.215692 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:36 crc kubenswrapper[4754]: I0126 16:48:36.215759 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:36 crc kubenswrapper[4754]: I0126 16:48:36.215771 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:36 crc kubenswrapper[4754]: I0126 16:48:36.215790 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:36 crc kubenswrapper[4754]: I0126 16:48:36.215803 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:36Z","lastTransitionTime":"2026-01-26T16:48:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:36 crc kubenswrapper[4754]: E0126 16:48:36.228792 4754 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:48:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:48:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:48:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:48:36Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:48:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:48:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:48:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:48:36Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d8506764-e7fa-45cb-a13d-6f527164f548\\\",\\\"systemUUID\\\":\\\"bfbb6c33-ce46-4e81-a8a8-b44409b03821\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:36Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:36 crc kubenswrapper[4754]: I0126 16:48:36.231861 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:36 crc kubenswrapper[4754]: I0126 16:48:36.231914 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:36 crc kubenswrapper[4754]: I0126 16:48:36.231925 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:36 crc kubenswrapper[4754]: I0126 16:48:36.231941 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:36 crc kubenswrapper[4754]: I0126 16:48:36.231951 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:36Z","lastTransitionTime":"2026-01-26T16:48:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:36 crc kubenswrapper[4754]: E0126 16:48:36.243154 4754 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:48:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:48:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:48:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:48:36Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:48:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:48:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:48:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:48:36Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d8506764-e7fa-45cb-a13d-6f527164f548\\\",\\\"systemUUID\\\":\\\"bfbb6c33-ce46-4e81-a8a8-b44409b03821\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:48:36Z is after 2025-08-24T17:21:41Z" Jan 26 16:48:36 crc kubenswrapper[4754]: E0126 16:48:36.243285 4754 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Jan 26 16:48:36 crc kubenswrapper[4754]: I0126 16:48:36.244916 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:36 crc kubenswrapper[4754]: I0126 16:48:36.244980 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:36 crc kubenswrapper[4754]: I0126 16:48:36.244991 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:36 crc kubenswrapper[4754]: I0126 16:48:36.245013 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:36 crc kubenswrapper[4754]: I0126 16:48:36.245025 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:36Z","lastTransitionTime":"2026-01-26T16:48:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:36 crc kubenswrapper[4754]: I0126 16:48:36.294366 4754 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-25 08:12:18.665122665 +0000 UTC Jan 26 16:48:36 crc kubenswrapper[4754]: I0126 16:48:36.347473 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:36 crc kubenswrapper[4754]: I0126 16:48:36.347508 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:36 crc kubenswrapper[4754]: I0126 16:48:36.347517 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:36 crc kubenswrapper[4754]: I0126 16:48:36.347565 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:36 crc kubenswrapper[4754]: I0126 16:48:36.347575 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:36Z","lastTransitionTime":"2026-01-26T16:48:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:36 crc kubenswrapper[4754]: I0126 16:48:36.449973 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:36 crc kubenswrapper[4754]: I0126 16:48:36.450031 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:36 crc kubenswrapper[4754]: I0126 16:48:36.450046 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:36 crc kubenswrapper[4754]: I0126 16:48:36.450064 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:36 crc kubenswrapper[4754]: I0126 16:48:36.450076 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:36Z","lastTransitionTime":"2026-01-26T16:48:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:36 crc kubenswrapper[4754]: I0126 16:48:36.552493 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:36 crc kubenswrapper[4754]: I0126 16:48:36.552517 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:36 crc kubenswrapper[4754]: I0126 16:48:36.552525 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:36 crc kubenswrapper[4754]: I0126 16:48:36.552539 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:36 crc kubenswrapper[4754]: I0126 16:48:36.552547 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:36Z","lastTransitionTime":"2026-01-26T16:48:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:36 crc kubenswrapper[4754]: I0126 16:48:36.654822 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:36 crc kubenswrapper[4754]: I0126 16:48:36.654863 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:36 crc kubenswrapper[4754]: I0126 16:48:36.654872 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:36 crc kubenswrapper[4754]: I0126 16:48:36.654887 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:36 crc kubenswrapper[4754]: I0126 16:48:36.654897 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:36Z","lastTransitionTime":"2026-01-26T16:48:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:36 crc kubenswrapper[4754]: I0126 16:48:36.758103 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:36 crc kubenswrapper[4754]: I0126 16:48:36.758184 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:36 crc kubenswrapper[4754]: I0126 16:48:36.758210 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:36 crc kubenswrapper[4754]: I0126 16:48:36.758235 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:36 crc kubenswrapper[4754]: I0126 16:48:36.758254 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:36Z","lastTransitionTime":"2026-01-26T16:48:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:36 crc kubenswrapper[4754]: I0126 16:48:36.771931 4754 scope.go:117] "RemoveContainer" containerID="2c059b7d618442c8e541bce1d2e576de3e9d9f9b3bd82973f0039fc6c8429d6e" Jan 26 16:48:36 crc kubenswrapper[4754]: E0126 16:48:36.772271 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-jsbxt_openshift-ovn-kubernetes(689569c4-93ee-4b82-93fc-9af0f97e6dc3)\"" pod="openshift-ovn-kubernetes/ovnkube-node-jsbxt" podUID="689569c4-93ee-4b82-93fc-9af0f97e6dc3" Jan 26 16:48:36 crc kubenswrapper[4754]: I0126 16:48:36.860703 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:36 crc kubenswrapper[4754]: I0126 16:48:36.860774 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:36 crc kubenswrapper[4754]: I0126 16:48:36.860796 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:36 crc kubenswrapper[4754]: I0126 16:48:36.860822 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:36 crc kubenswrapper[4754]: I0126 16:48:36.860843 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:36Z","lastTransitionTime":"2026-01-26T16:48:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:36 crc kubenswrapper[4754]: I0126 16:48:36.963175 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:36 crc kubenswrapper[4754]: I0126 16:48:36.963218 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:36 crc kubenswrapper[4754]: I0126 16:48:36.963250 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:36 crc kubenswrapper[4754]: I0126 16:48:36.963287 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:36 crc kubenswrapper[4754]: I0126 16:48:36.963301 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:36Z","lastTransitionTime":"2026-01-26T16:48:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:37 crc kubenswrapper[4754]: I0126 16:48:37.065886 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:37 crc kubenswrapper[4754]: I0126 16:48:37.065944 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:37 crc kubenswrapper[4754]: I0126 16:48:37.065953 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:37 crc kubenswrapper[4754]: I0126 16:48:37.065967 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:37 crc kubenswrapper[4754]: I0126 16:48:37.065976 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:37Z","lastTransitionTime":"2026-01-26T16:48:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:37 crc kubenswrapper[4754]: I0126 16:48:37.167900 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:37 crc kubenswrapper[4754]: I0126 16:48:37.167929 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:37 crc kubenswrapper[4754]: I0126 16:48:37.167938 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:37 crc kubenswrapper[4754]: I0126 16:48:37.167951 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:37 crc kubenswrapper[4754]: I0126 16:48:37.167961 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:37Z","lastTransitionTime":"2026-01-26T16:48:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:37 crc kubenswrapper[4754]: I0126 16:48:37.270894 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:37 crc kubenswrapper[4754]: I0126 16:48:37.270952 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:37 crc kubenswrapper[4754]: I0126 16:48:37.270962 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:37 crc kubenswrapper[4754]: I0126 16:48:37.270977 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:37 crc kubenswrapper[4754]: I0126 16:48:37.270985 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:37Z","lastTransitionTime":"2026-01-26T16:48:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:37 crc kubenswrapper[4754]: I0126 16:48:37.294519 4754 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-19 12:00:49.315308313 +0000 UTC Jan 26 16:48:37 crc kubenswrapper[4754]: I0126 16:48:37.373285 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:37 crc kubenswrapper[4754]: I0126 16:48:37.373328 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:37 crc kubenswrapper[4754]: I0126 16:48:37.373340 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:37 crc kubenswrapper[4754]: I0126 16:48:37.373355 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:37 crc kubenswrapper[4754]: I0126 16:48:37.373365 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:37Z","lastTransitionTime":"2026-01-26T16:48:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:37 crc kubenswrapper[4754]: I0126 16:48:37.475938 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:37 crc kubenswrapper[4754]: I0126 16:48:37.475976 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:37 crc kubenswrapper[4754]: I0126 16:48:37.475987 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:37 crc kubenswrapper[4754]: I0126 16:48:37.476001 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:37 crc kubenswrapper[4754]: I0126 16:48:37.476010 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:37Z","lastTransitionTime":"2026-01-26T16:48:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:37 crc kubenswrapper[4754]: I0126 16:48:37.578985 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:37 crc kubenswrapper[4754]: I0126 16:48:37.579058 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:37 crc kubenswrapper[4754]: I0126 16:48:37.579080 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:37 crc kubenswrapper[4754]: I0126 16:48:37.579109 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:37 crc kubenswrapper[4754]: I0126 16:48:37.579129 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:37Z","lastTransitionTime":"2026-01-26T16:48:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:37 crc kubenswrapper[4754]: I0126 16:48:37.681431 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:37 crc kubenswrapper[4754]: I0126 16:48:37.681465 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:37 crc kubenswrapper[4754]: I0126 16:48:37.681473 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:37 crc kubenswrapper[4754]: I0126 16:48:37.681487 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:37 crc kubenswrapper[4754]: I0126 16:48:37.681497 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:37Z","lastTransitionTime":"2026-01-26T16:48:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:37 crc kubenswrapper[4754]: I0126 16:48:37.770002 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 16:48:37 crc kubenswrapper[4754]: E0126 16:48:37.770153 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 16:48:37 crc kubenswrapper[4754]: I0126 16:48:37.770228 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 16:48:37 crc kubenswrapper[4754]: E0126 16:48:37.770304 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 16:48:37 crc kubenswrapper[4754]: I0126 16:48:37.770369 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 16:48:37 crc kubenswrapper[4754]: E0126 16:48:37.770443 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 16:48:37 crc kubenswrapper[4754]: I0126 16:48:37.770593 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-t4ps6" Jan 26 16:48:37 crc kubenswrapper[4754]: E0126 16:48:37.770731 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-t4ps6" podUID="d45179f0-9f81-4d77-8805-6b05bdb32f8f" Jan 26 16:48:37 crc kubenswrapper[4754]: I0126 16:48:37.783324 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:37 crc kubenswrapper[4754]: I0126 16:48:37.783367 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:37 crc kubenswrapper[4754]: I0126 16:48:37.783378 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:37 crc kubenswrapper[4754]: I0126 16:48:37.783393 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:37 crc kubenswrapper[4754]: I0126 16:48:37.783405 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:37Z","lastTransitionTime":"2026-01-26T16:48:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:37 crc kubenswrapper[4754]: I0126 16:48:37.886178 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:37 crc kubenswrapper[4754]: I0126 16:48:37.886248 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:37 crc kubenswrapper[4754]: I0126 16:48:37.886267 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:37 crc kubenswrapper[4754]: I0126 16:48:37.886294 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:37 crc kubenswrapper[4754]: I0126 16:48:37.886312 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:37Z","lastTransitionTime":"2026-01-26T16:48:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:37 crc kubenswrapper[4754]: I0126 16:48:37.989068 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:37 crc kubenswrapper[4754]: I0126 16:48:37.989139 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:37 crc kubenswrapper[4754]: I0126 16:48:37.989149 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:37 crc kubenswrapper[4754]: I0126 16:48:37.989162 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:37 crc kubenswrapper[4754]: I0126 16:48:37.989171 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:37Z","lastTransitionTime":"2026-01-26T16:48:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:38 crc kubenswrapper[4754]: I0126 16:48:38.091351 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:38 crc kubenswrapper[4754]: I0126 16:48:38.091412 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:38 crc kubenswrapper[4754]: I0126 16:48:38.091422 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:38 crc kubenswrapper[4754]: I0126 16:48:38.091437 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:38 crc kubenswrapper[4754]: I0126 16:48:38.091446 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:38Z","lastTransitionTime":"2026-01-26T16:48:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:38 crc kubenswrapper[4754]: I0126 16:48:38.194303 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:38 crc kubenswrapper[4754]: I0126 16:48:38.194365 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:38 crc kubenswrapper[4754]: I0126 16:48:38.194376 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:38 crc kubenswrapper[4754]: I0126 16:48:38.194391 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:38 crc kubenswrapper[4754]: I0126 16:48:38.194401 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:38Z","lastTransitionTime":"2026-01-26T16:48:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:38 crc kubenswrapper[4754]: I0126 16:48:38.295225 4754 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-06 05:36:23.155257393 +0000 UTC Jan 26 16:48:38 crc kubenswrapper[4754]: I0126 16:48:38.297226 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:38 crc kubenswrapper[4754]: I0126 16:48:38.297279 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:38 crc kubenswrapper[4754]: I0126 16:48:38.297288 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:38 crc kubenswrapper[4754]: I0126 16:48:38.297303 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:38 crc kubenswrapper[4754]: I0126 16:48:38.297316 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:38Z","lastTransitionTime":"2026-01-26T16:48:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:38 crc kubenswrapper[4754]: I0126 16:48:38.399840 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:38 crc kubenswrapper[4754]: I0126 16:48:38.399898 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:38 crc kubenswrapper[4754]: I0126 16:48:38.399909 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:38 crc kubenswrapper[4754]: I0126 16:48:38.399922 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:38 crc kubenswrapper[4754]: I0126 16:48:38.399931 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:38Z","lastTransitionTime":"2026-01-26T16:48:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:38 crc kubenswrapper[4754]: I0126 16:48:38.502528 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:38 crc kubenswrapper[4754]: I0126 16:48:38.502582 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:38 crc kubenswrapper[4754]: I0126 16:48:38.502595 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:38 crc kubenswrapper[4754]: I0126 16:48:38.502611 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:38 crc kubenswrapper[4754]: I0126 16:48:38.502624 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:38Z","lastTransitionTime":"2026-01-26T16:48:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:38 crc kubenswrapper[4754]: I0126 16:48:38.604992 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:38 crc kubenswrapper[4754]: I0126 16:48:38.605028 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:38 crc kubenswrapper[4754]: I0126 16:48:38.605038 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:38 crc kubenswrapper[4754]: I0126 16:48:38.605053 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:38 crc kubenswrapper[4754]: I0126 16:48:38.605062 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:38Z","lastTransitionTime":"2026-01-26T16:48:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:38 crc kubenswrapper[4754]: I0126 16:48:38.707081 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:38 crc kubenswrapper[4754]: I0126 16:48:38.707143 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:38 crc kubenswrapper[4754]: I0126 16:48:38.707158 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:38 crc kubenswrapper[4754]: I0126 16:48:38.707181 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:38 crc kubenswrapper[4754]: I0126 16:48:38.707196 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:38Z","lastTransitionTime":"2026-01-26T16:48:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:38 crc kubenswrapper[4754]: I0126 16:48:38.809703 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:38 crc kubenswrapper[4754]: I0126 16:48:38.809750 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:38 crc kubenswrapper[4754]: I0126 16:48:38.809778 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:38 crc kubenswrapper[4754]: I0126 16:48:38.809792 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:38 crc kubenswrapper[4754]: I0126 16:48:38.809803 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:38Z","lastTransitionTime":"2026-01-26T16:48:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:38 crc kubenswrapper[4754]: I0126 16:48:38.911772 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:38 crc kubenswrapper[4754]: I0126 16:48:38.911808 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:38 crc kubenswrapper[4754]: I0126 16:48:38.911818 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:38 crc kubenswrapper[4754]: I0126 16:48:38.911832 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:38 crc kubenswrapper[4754]: I0126 16:48:38.911842 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:38Z","lastTransitionTime":"2026-01-26T16:48:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:39 crc kubenswrapper[4754]: I0126 16:48:39.014587 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:39 crc kubenswrapper[4754]: I0126 16:48:39.014651 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:39 crc kubenswrapper[4754]: I0126 16:48:39.014662 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:39 crc kubenswrapper[4754]: I0126 16:48:39.014715 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:39 crc kubenswrapper[4754]: I0126 16:48:39.014725 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:39Z","lastTransitionTime":"2026-01-26T16:48:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:39 crc kubenswrapper[4754]: I0126 16:48:39.117140 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:39 crc kubenswrapper[4754]: I0126 16:48:39.117179 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:39 crc kubenswrapper[4754]: I0126 16:48:39.117187 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:39 crc kubenswrapper[4754]: I0126 16:48:39.117199 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:39 crc kubenswrapper[4754]: I0126 16:48:39.117208 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:39Z","lastTransitionTime":"2026-01-26T16:48:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:39 crc kubenswrapper[4754]: I0126 16:48:39.219455 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:39 crc kubenswrapper[4754]: I0126 16:48:39.219519 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:39 crc kubenswrapper[4754]: I0126 16:48:39.219539 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:39 crc kubenswrapper[4754]: I0126 16:48:39.219563 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:39 crc kubenswrapper[4754]: I0126 16:48:39.219581 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:39Z","lastTransitionTime":"2026-01-26T16:48:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:39 crc kubenswrapper[4754]: I0126 16:48:39.295334 4754 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-20 18:27:15.803136728 +0000 UTC Jan 26 16:48:39 crc kubenswrapper[4754]: I0126 16:48:39.322957 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:39 crc kubenswrapper[4754]: I0126 16:48:39.322997 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:39 crc kubenswrapper[4754]: I0126 16:48:39.323007 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:39 crc kubenswrapper[4754]: I0126 16:48:39.323020 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:39 crc kubenswrapper[4754]: I0126 16:48:39.323030 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:39Z","lastTransitionTime":"2026-01-26T16:48:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:39 crc kubenswrapper[4754]: I0126 16:48:39.425364 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:39 crc kubenswrapper[4754]: I0126 16:48:39.425407 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:39 crc kubenswrapper[4754]: I0126 16:48:39.425422 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:39 crc kubenswrapper[4754]: I0126 16:48:39.425438 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:39 crc kubenswrapper[4754]: I0126 16:48:39.425448 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:39Z","lastTransitionTime":"2026-01-26T16:48:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:39 crc kubenswrapper[4754]: I0126 16:48:39.528330 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:39 crc kubenswrapper[4754]: I0126 16:48:39.528385 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:39 crc kubenswrapper[4754]: I0126 16:48:39.528397 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:39 crc kubenswrapper[4754]: I0126 16:48:39.528413 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:39 crc kubenswrapper[4754]: I0126 16:48:39.528425 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:39Z","lastTransitionTime":"2026-01-26T16:48:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:39 crc kubenswrapper[4754]: I0126 16:48:39.631258 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:39 crc kubenswrapper[4754]: I0126 16:48:39.631320 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:39 crc kubenswrapper[4754]: I0126 16:48:39.631332 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:39 crc kubenswrapper[4754]: I0126 16:48:39.631350 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:39 crc kubenswrapper[4754]: I0126 16:48:39.631362 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:39Z","lastTransitionTime":"2026-01-26T16:48:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:39 crc kubenswrapper[4754]: I0126 16:48:39.734820 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:39 crc kubenswrapper[4754]: I0126 16:48:39.734869 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:39 crc kubenswrapper[4754]: I0126 16:48:39.734883 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:39 crc kubenswrapper[4754]: I0126 16:48:39.734899 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:39 crc kubenswrapper[4754]: I0126 16:48:39.734911 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:39Z","lastTransitionTime":"2026-01-26T16:48:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:39 crc kubenswrapper[4754]: I0126 16:48:39.767333 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 16:48:39 crc kubenswrapper[4754]: I0126 16:48:39.767336 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 16:48:39 crc kubenswrapper[4754]: I0126 16:48:39.767472 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 16:48:39 crc kubenswrapper[4754]: I0126 16:48:39.767493 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-t4ps6" Jan 26 16:48:39 crc kubenswrapper[4754]: E0126 16:48:39.767566 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 16:48:39 crc kubenswrapper[4754]: E0126 16:48:39.767779 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-t4ps6" podUID="d45179f0-9f81-4d77-8805-6b05bdb32f8f" Jan 26 16:48:39 crc kubenswrapper[4754]: E0126 16:48:39.767894 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 16:48:39 crc kubenswrapper[4754]: E0126 16:48:39.768112 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 16:48:39 crc kubenswrapper[4754]: I0126 16:48:39.837286 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:39 crc kubenswrapper[4754]: I0126 16:48:39.837374 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:39 crc kubenswrapper[4754]: I0126 16:48:39.837395 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:39 crc kubenswrapper[4754]: I0126 16:48:39.837418 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:39 crc kubenswrapper[4754]: I0126 16:48:39.837438 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:39Z","lastTransitionTime":"2026-01-26T16:48:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:39 crc kubenswrapper[4754]: I0126 16:48:39.940187 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:39 crc kubenswrapper[4754]: I0126 16:48:39.940247 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:39 crc kubenswrapper[4754]: I0126 16:48:39.940257 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:39 crc kubenswrapper[4754]: I0126 16:48:39.940272 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:39 crc kubenswrapper[4754]: I0126 16:48:39.940282 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:39Z","lastTransitionTime":"2026-01-26T16:48:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:40 crc kubenswrapper[4754]: I0126 16:48:40.042444 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:40 crc kubenswrapper[4754]: I0126 16:48:40.042489 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:40 crc kubenswrapper[4754]: I0126 16:48:40.042519 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:40 crc kubenswrapper[4754]: I0126 16:48:40.042536 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:40 crc kubenswrapper[4754]: I0126 16:48:40.042546 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:40Z","lastTransitionTime":"2026-01-26T16:48:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:40 crc kubenswrapper[4754]: I0126 16:48:40.144467 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:40 crc kubenswrapper[4754]: I0126 16:48:40.144521 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:40 crc kubenswrapper[4754]: I0126 16:48:40.144533 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:40 crc kubenswrapper[4754]: I0126 16:48:40.144550 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:40 crc kubenswrapper[4754]: I0126 16:48:40.144563 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:40Z","lastTransitionTime":"2026-01-26T16:48:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:40 crc kubenswrapper[4754]: I0126 16:48:40.246733 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:40 crc kubenswrapper[4754]: I0126 16:48:40.246771 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:40 crc kubenswrapper[4754]: I0126 16:48:40.246782 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:40 crc kubenswrapper[4754]: I0126 16:48:40.246797 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:40 crc kubenswrapper[4754]: I0126 16:48:40.246808 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:40Z","lastTransitionTime":"2026-01-26T16:48:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:40 crc kubenswrapper[4754]: I0126 16:48:40.296457 4754 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-08 06:10:19.544247211 +0000 UTC Jan 26 16:48:40 crc kubenswrapper[4754]: I0126 16:48:40.349478 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:40 crc kubenswrapper[4754]: I0126 16:48:40.349519 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:40 crc kubenswrapper[4754]: I0126 16:48:40.349527 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:40 crc kubenswrapper[4754]: I0126 16:48:40.349540 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:40 crc kubenswrapper[4754]: I0126 16:48:40.349548 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:40Z","lastTransitionTime":"2026-01-26T16:48:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:40 crc kubenswrapper[4754]: I0126 16:48:40.452197 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:40 crc kubenswrapper[4754]: I0126 16:48:40.452235 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:40 crc kubenswrapper[4754]: I0126 16:48:40.452266 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:40 crc kubenswrapper[4754]: I0126 16:48:40.452281 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:40 crc kubenswrapper[4754]: I0126 16:48:40.452289 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:40Z","lastTransitionTime":"2026-01-26T16:48:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:40 crc kubenswrapper[4754]: I0126 16:48:40.555535 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:40 crc kubenswrapper[4754]: I0126 16:48:40.555601 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:40 crc kubenswrapper[4754]: I0126 16:48:40.555611 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:40 crc kubenswrapper[4754]: I0126 16:48:40.555650 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:40 crc kubenswrapper[4754]: I0126 16:48:40.555660 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:40Z","lastTransitionTime":"2026-01-26T16:48:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:40 crc kubenswrapper[4754]: I0126 16:48:40.657699 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:40 crc kubenswrapper[4754]: I0126 16:48:40.657765 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:40 crc kubenswrapper[4754]: I0126 16:48:40.657775 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:40 crc kubenswrapper[4754]: I0126 16:48:40.657793 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:40 crc kubenswrapper[4754]: I0126 16:48:40.657803 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:40Z","lastTransitionTime":"2026-01-26T16:48:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:40 crc kubenswrapper[4754]: I0126 16:48:40.760312 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:40 crc kubenswrapper[4754]: I0126 16:48:40.760358 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:40 crc kubenswrapper[4754]: I0126 16:48:40.760374 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:40 crc kubenswrapper[4754]: I0126 16:48:40.760403 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:40 crc kubenswrapper[4754]: I0126 16:48:40.760423 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:40Z","lastTransitionTime":"2026-01-26T16:48:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:40 crc kubenswrapper[4754]: I0126 16:48:40.863229 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:40 crc kubenswrapper[4754]: I0126 16:48:40.863280 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:40 crc kubenswrapper[4754]: I0126 16:48:40.863291 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:40 crc kubenswrapper[4754]: I0126 16:48:40.863309 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:40 crc kubenswrapper[4754]: I0126 16:48:40.863320 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:40Z","lastTransitionTime":"2026-01-26T16:48:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:40 crc kubenswrapper[4754]: I0126 16:48:40.965820 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:40 crc kubenswrapper[4754]: I0126 16:48:40.965885 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:40 crc kubenswrapper[4754]: I0126 16:48:40.965894 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:40 crc kubenswrapper[4754]: I0126 16:48:40.965908 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:40 crc kubenswrapper[4754]: I0126 16:48:40.965918 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:40Z","lastTransitionTime":"2026-01-26T16:48:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:41 crc kubenswrapper[4754]: I0126 16:48:41.068548 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:41 crc kubenswrapper[4754]: I0126 16:48:41.068591 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:41 crc kubenswrapper[4754]: I0126 16:48:41.068600 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:41 crc kubenswrapper[4754]: I0126 16:48:41.068616 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:41 crc kubenswrapper[4754]: I0126 16:48:41.068627 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:41Z","lastTransitionTime":"2026-01-26T16:48:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:41 crc kubenswrapper[4754]: I0126 16:48:41.171097 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:41 crc kubenswrapper[4754]: I0126 16:48:41.171141 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:41 crc kubenswrapper[4754]: I0126 16:48:41.171151 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:41 crc kubenswrapper[4754]: I0126 16:48:41.171166 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:41 crc kubenswrapper[4754]: I0126 16:48:41.171176 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:41Z","lastTransitionTime":"2026-01-26T16:48:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:41 crc kubenswrapper[4754]: I0126 16:48:41.273689 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:41 crc kubenswrapper[4754]: I0126 16:48:41.273750 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:41 crc kubenswrapper[4754]: I0126 16:48:41.273761 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:41 crc kubenswrapper[4754]: I0126 16:48:41.273776 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:41 crc kubenswrapper[4754]: I0126 16:48:41.273789 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:41Z","lastTransitionTime":"2026-01-26T16:48:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:41 crc kubenswrapper[4754]: I0126 16:48:41.296872 4754 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-16 12:37:21.091536892 +0000 UTC Jan 26 16:48:41 crc kubenswrapper[4754]: I0126 16:48:41.375971 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:41 crc kubenswrapper[4754]: I0126 16:48:41.376014 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:41 crc kubenswrapper[4754]: I0126 16:48:41.376025 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:41 crc kubenswrapper[4754]: I0126 16:48:41.376041 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:41 crc kubenswrapper[4754]: I0126 16:48:41.376055 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:41Z","lastTransitionTime":"2026-01-26T16:48:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:41 crc kubenswrapper[4754]: I0126 16:48:41.478611 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:41 crc kubenswrapper[4754]: I0126 16:48:41.478654 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:41 crc kubenswrapper[4754]: I0126 16:48:41.478699 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:41 crc kubenswrapper[4754]: I0126 16:48:41.478723 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:41 crc kubenswrapper[4754]: I0126 16:48:41.478735 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:41Z","lastTransitionTime":"2026-01-26T16:48:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:41 crc kubenswrapper[4754]: I0126 16:48:41.581617 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:41 crc kubenswrapper[4754]: I0126 16:48:41.581655 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:41 crc kubenswrapper[4754]: I0126 16:48:41.581682 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:41 crc kubenswrapper[4754]: I0126 16:48:41.581698 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:41 crc kubenswrapper[4754]: I0126 16:48:41.581708 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:41Z","lastTransitionTime":"2026-01-26T16:48:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:41 crc kubenswrapper[4754]: I0126 16:48:41.684502 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:41 crc kubenswrapper[4754]: I0126 16:48:41.684548 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:41 crc kubenswrapper[4754]: I0126 16:48:41.684558 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:41 crc kubenswrapper[4754]: I0126 16:48:41.684572 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:41 crc kubenswrapper[4754]: I0126 16:48:41.684585 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:41Z","lastTransitionTime":"2026-01-26T16:48:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:41 crc kubenswrapper[4754]: I0126 16:48:41.766989 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-t4ps6" Jan 26 16:48:41 crc kubenswrapper[4754]: I0126 16:48:41.767042 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 16:48:41 crc kubenswrapper[4754]: I0126 16:48:41.767071 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 16:48:41 crc kubenswrapper[4754]: E0126 16:48:41.767131 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-t4ps6" podUID="d45179f0-9f81-4d77-8805-6b05bdb32f8f" Jan 26 16:48:41 crc kubenswrapper[4754]: E0126 16:48:41.767322 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 16:48:41 crc kubenswrapper[4754]: E0126 16:48:41.767383 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 16:48:41 crc kubenswrapper[4754]: I0126 16:48:41.767565 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 16:48:41 crc kubenswrapper[4754]: E0126 16:48:41.768754 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 16:48:41 crc kubenswrapper[4754]: I0126 16:48:41.787057 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:41 crc kubenswrapper[4754]: I0126 16:48:41.787111 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:41 crc kubenswrapper[4754]: I0126 16:48:41.787122 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:41 crc kubenswrapper[4754]: I0126 16:48:41.787139 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:41 crc kubenswrapper[4754]: I0126 16:48:41.787155 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:41Z","lastTransitionTime":"2026-01-26T16:48:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:41 crc kubenswrapper[4754]: I0126 16:48:41.889902 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:41 crc kubenswrapper[4754]: I0126 16:48:41.889940 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:41 crc kubenswrapper[4754]: I0126 16:48:41.889950 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:41 crc kubenswrapper[4754]: I0126 16:48:41.889967 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:41 crc kubenswrapper[4754]: I0126 16:48:41.889978 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:41Z","lastTransitionTime":"2026-01-26T16:48:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:41 crc kubenswrapper[4754]: I0126 16:48:41.992537 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:41 crc kubenswrapper[4754]: I0126 16:48:41.992593 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:41 crc kubenswrapper[4754]: I0126 16:48:41.992602 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:41 crc kubenswrapper[4754]: I0126 16:48:41.992620 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:41 crc kubenswrapper[4754]: I0126 16:48:41.992630 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:41Z","lastTransitionTime":"2026-01-26T16:48:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:42 crc kubenswrapper[4754]: I0126 16:48:42.094443 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:42 crc kubenswrapper[4754]: I0126 16:48:42.094477 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:42 crc kubenswrapper[4754]: I0126 16:48:42.094486 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:42 crc kubenswrapper[4754]: I0126 16:48:42.094498 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:42 crc kubenswrapper[4754]: I0126 16:48:42.094507 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:42Z","lastTransitionTime":"2026-01-26T16:48:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:42 crc kubenswrapper[4754]: I0126 16:48:42.197506 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:42 crc kubenswrapper[4754]: I0126 16:48:42.197553 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:42 crc kubenswrapper[4754]: I0126 16:48:42.197562 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:42 crc kubenswrapper[4754]: I0126 16:48:42.197574 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:42 crc kubenswrapper[4754]: I0126 16:48:42.197584 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:42Z","lastTransitionTime":"2026-01-26T16:48:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:42 crc kubenswrapper[4754]: I0126 16:48:42.297656 4754 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-27 18:37:42.310106245 +0000 UTC Jan 26 16:48:42 crc kubenswrapper[4754]: I0126 16:48:42.299278 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:42 crc kubenswrapper[4754]: I0126 16:48:42.299315 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:42 crc kubenswrapper[4754]: I0126 16:48:42.299328 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:42 crc kubenswrapper[4754]: I0126 16:48:42.299344 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:42 crc kubenswrapper[4754]: I0126 16:48:42.299355 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:42Z","lastTransitionTime":"2026-01-26T16:48:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:42 crc kubenswrapper[4754]: I0126 16:48:42.402247 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:42 crc kubenswrapper[4754]: I0126 16:48:42.402285 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:42 crc kubenswrapper[4754]: I0126 16:48:42.402293 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:42 crc kubenswrapper[4754]: I0126 16:48:42.402307 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:42 crc kubenswrapper[4754]: I0126 16:48:42.402319 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:42Z","lastTransitionTime":"2026-01-26T16:48:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:42 crc kubenswrapper[4754]: I0126 16:48:42.506821 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:42 crc kubenswrapper[4754]: I0126 16:48:42.506881 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:42 crc kubenswrapper[4754]: I0126 16:48:42.506897 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:42 crc kubenswrapper[4754]: I0126 16:48:42.506916 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:42 crc kubenswrapper[4754]: I0126 16:48:42.506932 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:42Z","lastTransitionTime":"2026-01-26T16:48:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:42 crc kubenswrapper[4754]: I0126 16:48:42.609975 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:42 crc kubenswrapper[4754]: I0126 16:48:42.610020 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:42 crc kubenswrapper[4754]: I0126 16:48:42.610030 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:42 crc kubenswrapper[4754]: I0126 16:48:42.610110 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:42 crc kubenswrapper[4754]: I0126 16:48:42.610122 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:42Z","lastTransitionTime":"2026-01-26T16:48:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:42 crc kubenswrapper[4754]: I0126 16:48:42.713737 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:42 crc kubenswrapper[4754]: I0126 16:48:42.713808 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:42 crc kubenswrapper[4754]: I0126 16:48:42.713823 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:42 crc kubenswrapper[4754]: I0126 16:48:42.713847 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:42 crc kubenswrapper[4754]: I0126 16:48:42.713858 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:42Z","lastTransitionTime":"2026-01-26T16:48:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:42 crc kubenswrapper[4754]: I0126 16:48:42.816235 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:42 crc kubenswrapper[4754]: I0126 16:48:42.816285 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:42 crc kubenswrapper[4754]: I0126 16:48:42.816297 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:42 crc kubenswrapper[4754]: I0126 16:48:42.816316 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:42 crc kubenswrapper[4754]: I0126 16:48:42.816328 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:42Z","lastTransitionTime":"2026-01-26T16:48:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:42 crc kubenswrapper[4754]: I0126 16:48:42.919755 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:42 crc kubenswrapper[4754]: I0126 16:48:42.919802 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:42 crc kubenswrapper[4754]: I0126 16:48:42.919816 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:42 crc kubenswrapper[4754]: I0126 16:48:42.919831 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:42 crc kubenswrapper[4754]: I0126 16:48:42.919841 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:42Z","lastTransitionTime":"2026-01-26T16:48:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:43 crc kubenswrapper[4754]: I0126 16:48:43.022768 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:43 crc kubenswrapper[4754]: I0126 16:48:43.022818 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:43 crc kubenswrapper[4754]: I0126 16:48:43.022834 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:43 crc kubenswrapper[4754]: I0126 16:48:43.022856 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:43 crc kubenswrapper[4754]: I0126 16:48:43.022870 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:43Z","lastTransitionTime":"2026-01-26T16:48:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:43 crc kubenswrapper[4754]: I0126 16:48:43.125625 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:43 crc kubenswrapper[4754]: I0126 16:48:43.125706 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:43 crc kubenswrapper[4754]: I0126 16:48:43.125725 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:43 crc kubenswrapper[4754]: I0126 16:48:43.125756 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:43 crc kubenswrapper[4754]: I0126 16:48:43.125774 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:43Z","lastTransitionTime":"2026-01-26T16:48:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:43 crc kubenswrapper[4754]: I0126 16:48:43.228453 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:43 crc kubenswrapper[4754]: I0126 16:48:43.228487 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:43 crc kubenswrapper[4754]: I0126 16:48:43.228495 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:43 crc kubenswrapper[4754]: I0126 16:48:43.228508 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:43 crc kubenswrapper[4754]: I0126 16:48:43.228518 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:43Z","lastTransitionTime":"2026-01-26T16:48:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:43 crc kubenswrapper[4754]: I0126 16:48:43.298739 4754 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-06 08:39:15.418120416 +0000 UTC Jan 26 16:48:43 crc kubenswrapper[4754]: I0126 16:48:43.330815 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:43 crc kubenswrapper[4754]: I0126 16:48:43.330856 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:43 crc kubenswrapper[4754]: I0126 16:48:43.330865 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:43 crc kubenswrapper[4754]: I0126 16:48:43.330880 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:43 crc kubenswrapper[4754]: I0126 16:48:43.330892 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:43Z","lastTransitionTime":"2026-01-26T16:48:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:43 crc kubenswrapper[4754]: I0126 16:48:43.433420 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:43 crc kubenswrapper[4754]: I0126 16:48:43.433466 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:43 crc kubenswrapper[4754]: I0126 16:48:43.433479 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:43 crc kubenswrapper[4754]: I0126 16:48:43.433495 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:43 crc kubenswrapper[4754]: I0126 16:48:43.433507 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:43Z","lastTransitionTime":"2026-01-26T16:48:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:43 crc kubenswrapper[4754]: I0126 16:48:43.536063 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:43 crc kubenswrapper[4754]: I0126 16:48:43.536102 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:43 crc kubenswrapper[4754]: I0126 16:48:43.536110 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:43 crc kubenswrapper[4754]: I0126 16:48:43.536149 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:43 crc kubenswrapper[4754]: I0126 16:48:43.536160 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:43Z","lastTransitionTime":"2026-01-26T16:48:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:43 crc kubenswrapper[4754]: I0126 16:48:43.639396 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:43 crc kubenswrapper[4754]: I0126 16:48:43.639480 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:43 crc kubenswrapper[4754]: I0126 16:48:43.639503 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:43 crc kubenswrapper[4754]: I0126 16:48:43.639531 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:43 crc kubenswrapper[4754]: I0126 16:48:43.639553 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:43Z","lastTransitionTime":"2026-01-26T16:48:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:43 crc kubenswrapper[4754]: I0126 16:48:43.742319 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:43 crc kubenswrapper[4754]: I0126 16:48:43.742379 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:43 crc kubenswrapper[4754]: I0126 16:48:43.742398 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:43 crc kubenswrapper[4754]: I0126 16:48:43.742421 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:43 crc kubenswrapper[4754]: I0126 16:48:43.742437 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:43Z","lastTransitionTime":"2026-01-26T16:48:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:43 crc kubenswrapper[4754]: I0126 16:48:43.766657 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 16:48:43 crc kubenswrapper[4754]: I0126 16:48:43.766747 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 16:48:43 crc kubenswrapper[4754]: E0126 16:48:43.766804 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 16:48:43 crc kubenswrapper[4754]: I0126 16:48:43.766826 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-t4ps6" Jan 26 16:48:43 crc kubenswrapper[4754]: I0126 16:48:43.766829 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 16:48:43 crc kubenswrapper[4754]: E0126 16:48:43.767006 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 16:48:43 crc kubenswrapper[4754]: E0126 16:48:43.767058 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 16:48:43 crc kubenswrapper[4754]: E0126 16:48:43.767130 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-t4ps6" podUID="d45179f0-9f81-4d77-8805-6b05bdb32f8f" Jan 26 16:48:43 crc kubenswrapper[4754]: I0126 16:48:43.789952 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=15.789920755 podStartE2EDuration="15.789920755s" podCreationTimestamp="2026-01-26 16:48:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 16:48:43.789634937 +0000 UTC m=+90.313815381" watchObservedRunningTime="2026-01-26 16:48:43.789920755 +0000 UTC m=+90.314101189" Jan 26 16:48:43 crc kubenswrapper[4754]: I0126 16:48:43.800618 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-2blzw" podStartSLOduration=70.800586434 podStartE2EDuration="1m10.800586434s" podCreationTimestamp="2026-01-26 16:47:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 16:48:43.800461521 +0000 UTC m=+90.324641955" watchObservedRunningTime="2026-01-26 16:48:43.800586434 +0000 UTC m=+90.324766868" Jan 26 16:48:43 crc kubenswrapper[4754]: I0126 16:48:43.843887 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:43 crc kubenswrapper[4754]: I0126 16:48:43.843932 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:43 crc kubenswrapper[4754]: I0126 16:48:43.843944 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:43 crc kubenswrapper[4754]: I0126 16:48:43.843959 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:43 crc kubenswrapper[4754]: I0126 16:48:43.844008 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:43Z","lastTransitionTime":"2026-01-26T16:48:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:43 crc kubenswrapper[4754]: I0126 16:48:43.852465 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-bwpd6" podStartSLOduration=70.852446362 podStartE2EDuration="1m10.852446362s" podCreationTimestamp="2026-01-26 16:47:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 16:48:43.821642235 +0000 UTC m=+90.345822689" watchObservedRunningTime="2026-01-26 16:48:43.852446362 +0000 UTC m=+90.376626796" Jan 26 16:48:43 crc kubenswrapper[4754]: I0126 16:48:43.923968 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-84lgq" podStartSLOduration=70.923919706 podStartE2EDuration="1m10.923919706s" podCreationTimestamp="2026-01-26 16:47:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 16:48:43.922256672 +0000 UTC m=+90.446437116" watchObservedRunningTime="2026-01-26 16:48:43.923919706 +0000 UTC m=+90.448100140" Jan 26 16:48:43 crc kubenswrapper[4754]: I0126 16:48:43.946884 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:43 crc kubenswrapper[4754]: I0126 16:48:43.946927 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:43 crc kubenswrapper[4754]: I0126 16:48:43.946936 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:43 crc kubenswrapper[4754]: I0126 16:48:43.946953 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:43 crc kubenswrapper[4754]: I0126 16:48:43.946962 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:43Z","lastTransitionTime":"2026-01-26T16:48:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:43 crc kubenswrapper[4754]: I0126 16:48:43.956981 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jspvh" podStartSLOduration=70.956957501 podStartE2EDuration="1m10.956957501s" podCreationTimestamp="2026-01-26 16:47:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 16:48:43.956874099 +0000 UTC m=+90.481054533" watchObservedRunningTime="2026-01-26 16:48:43.956957501 +0000 UTC m=+90.481137935" Jan 26 16:48:43 crc kubenswrapper[4754]: I0126 16:48:43.957420 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-skh5n" podStartSLOduration=70.957411593 podStartE2EDuration="1m10.957411593s" podCreationTimestamp="2026-01-26 16:47:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 16:48:43.938727503 +0000 UTC m=+90.462907937" watchObservedRunningTime="2026-01-26 16:48:43.957411593 +0000 UTC m=+90.481592027" Jan 26 16:48:43 crc kubenswrapper[4754]: I0126 16:48:43.992968 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=69.992947013 podStartE2EDuration="1m9.992947013s" podCreationTimestamp="2026-01-26 16:47:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 16:48:43.992840861 +0000 UTC m=+90.517021325" watchObservedRunningTime="2026-01-26 16:48:43.992947013 +0000 UTC m=+90.517127447" Jan 26 16:48:44 crc kubenswrapper[4754]: I0126 16:48:44.023459 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=36.023428352 podStartE2EDuration="36.023428352s" podCreationTimestamp="2026-01-26 16:48:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 16:48:44.023081273 +0000 UTC m=+90.547261707" watchObservedRunningTime="2026-01-26 16:48:44.023428352 +0000 UTC m=+90.547608786" Jan 26 16:48:44 crc kubenswrapper[4754]: I0126 16:48:44.049234 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:44 crc kubenswrapper[4754]: I0126 16:48:44.049509 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:44 crc kubenswrapper[4754]: I0126 16:48:44.049602 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:44 crc kubenswrapper[4754]: I0126 16:48:44.049717 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:44 crc kubenswrapper[4754]: I0126 16:48:44.049805 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:44Z","lastTransitionTime":"2026-01-26T16:48:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:44 crc kubenswrapper[4754]: I0126 16:48:44.069746 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podStartSLOduration=71.069718376 podStartE2EDuration="1m11.069718376s" podCreationTimestamp="2026-01-26 16:47:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 16:48:44.057543507 +0000 UTC m=+90.581723971" watchObservedRunningTime="2026-01-26 16:48:44.069718376 +0000 UTC m=+90.593898810" Jan 26 16:48:44 crc kubenswrapper[4754]: I0126 16:48:44.087005 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=73.086972807 podStartE2EDuration="1m13.086972807s" podCreationTimestamp="2026-01-26 16:47:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 16:48:44.086556787 +0000 UTC m=+90.610737241" watchObservedRunningTime="2026-01-26 16:48:44.086972807 +0000 UTC m=+90.611153241" Jan 26 16:48:44 crc kubenswrapper[4754]: I0126 16:48:44.153198 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:44 crc kubenswrapper[4754]: I0126 16:48:44.153256 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:44 crc kubenswrapper[4754]: I0126 16:48:44.153269 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:44 crc kubenswrapper[4754]: I0126 16:48:44.153289 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:44 crc kubenswrapper[4754]: I0126 16:48:44.153303 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:44Z","lastTransitionTime":"2026-01-26T16:48:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:44 crc kubenswrapper[4754]: I0126 16:48:44.255836 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:44 crc kubenswrapper[4754]: I0126 16:48:44.255877 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:44 crc kubenswrapper[4754]: I0126 16:48:44.255888 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:44 crc kubenswrapper[4754]: I0126 16:48:44.255902 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:44 crc kubenswrapper[4754]: I0126 16:48:44.255912 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:44Z","lastTransitionTime":"2026-01-26T16:48:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:44 crc kubenswrapper[4754]: I0126 16:48:44.299722 4754 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-20 06:30:57.42819854 +0000 UTC Jan 26 16:48:44 crc kubenswrapper[4754]: I0126 16:48:44.359234 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:44 crc kubenswrapper[4754]: I0126 16:48:44.359277 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:44 crc kubenswrapper[4754]: I0126 16:48:44.359289 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:44 crc kubenswrapper[4754]: I0126 16:48:44.359305 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:44 crc kubenswrapper[4754]: I0126 16:48:44.359316 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:44Z","lastTransitionTime":"2026-01-26T16:48:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:44 crc kubenswrapper[4754]: I0126 16:48:44.461569 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:44 crc kubenswrapper[4754]: I0126 16:48:44.461848 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:44 crc kubenswrapper[4754]: I0126 16:48:44.461857 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:44 crc kubenswrapper[4754]: I0126 16:48:44.461870 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:44 crc kubenswrapper[4754]: I0126 16:48:44.461881 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:44Z","lastTransitionTime":"2026-01-26T16:48:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:44 crc kubenswrapper[4754]: I0126 16:48:44.564745 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:44 crc kubenswrapper[4754]: I0126 16:48:44.564784 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:44 crc kubenswrapper[4754]: I0126 16:48:44.564795 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:44 crc kubenswrapper[4754]: I0126 16:48:44.564813 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:44 crc kubenswrapper[4754]: I0126 16:48:44.564825 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:44Z","lastTransitionTime":"2026-01-26T16:48:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:44 crc kubenswrapper[4754]: I0126 16:48:44.668540 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:44 crc kubenswrapper[4754]: I0126 16:48:44.668596 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:44 crc kubenswrapper[4754]: I0126 16:48:44.668609 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:44 crc kubenswrapper[4754]: I0126 16:48:44.668630 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:44 crc kubenswrapper[4754]: I0126 16:48:44.668642 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:44Z","lastTransitionTime":"2026-01-26T16:48:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:44 crc kubenswrapper[4754]: I0126 16:48:44.773246 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:44 crc kubenswrapper[4754]: I0126 16:48:44.773312 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:44 crc kubenswrapper[4754]: I0126 16:48:44.773327 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:44 crc kubenswrapper[4754]: I0126 16:48:44.773347 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:44 crc kubenswrapper[4754]: I0126 16:48:44.773368 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:44Z","lastTransitionTime":"2026-01-26T16:48:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:44 crc kubenswrapper[4754]: I0126 16:48:44.875544 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:44 crc kubenswrapper[4754]: I0126 16:48:44.875583 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:44 crc kubenswrapper[4754]: I0126 16:48:44.875594 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:44 crc kubenswrapper[4754]: I0126 16:48:44.875608 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:44 crc kubenswrapper[4754]: I0126 16:48:44.875619 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:44Z","lastTransitionTime":"2026-01-26T16:48:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:44 crc kubenswrapper[4754]: I0126 16:48:44.978651 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:44 crc kubenswrapper[4754]: I0126 16:48:44.978741 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:44 crc kubenswrapper[4754]: I0126 16:48:44.978755 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:44 crc kubenswrapper[4754]: I0126 16:48:44.978771 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:44 crc kubenswrapper[4754]: I0126 16:48:44.978783 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:44Z","lastTransitionTime":"2026-01-26T16:48:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:45 crc kubenswrapper[4754]: I0126 16:48:45.082113 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:45 crc kubenswrapper[4754]: I0126 16:48:45.082179 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:45 crc kubenswrapper[4754]: I0126 16:48:45.082193 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:45 crc kubenswrapper[4754]: I0126 16:48:45.082218 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:45 crc kubenswrapper[4754]: I0126 16:48:45.082238 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:45Z","lastTransitionTime":"2026-01-26T16:48:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:45 crc kubenswrapper[4754]: I0126 16:48:45.186007 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:45 crc kubenswrapper[4754]: I0126 16:48:45.186060 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:45 crc kubenswrapper[4754]: I0126 16:48:45.186075 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:45 crc kubenswrapper[4754]: I0126 16:48:45.186096 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:45 crc kubenswrapper[4754]: I0126 16:48:45.186109 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:45Z","lastTransitionTime":"2026-01-26T16:48:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:45 crc kubenswrapper[4754]: I0126 16:48:45.288377 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:45 crc kubenswrapper[4754]: I0126 16:48:45.288438 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:45 crc kubenswrapper[4754]: I0126 16:48:45.288449 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:45 crc kubenswrapper[4754]: I0126 16:48:45.288463 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:45 crc kubenswrapper[4754]: I0126 16:48:45.288475 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:45Z","lastTransitionTime":"2026-01-26T16:48:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:45 crc kubenswrapper[4754]: I0126 16:48:45.299867 4754 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-10 20:13:34.765693827 +0000 UTC Jan 26 16:48:45 crc kubenswrapper[4754]: I0126 16:48:45.390538 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:45 crc kubenswrapper[4754]: I0126 16:48:45.390595 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:45 crc kubenswrapper[4754]: I0126 16:48:45.390611 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:45 crc kubenswrapper[4754]: I0126 16:48:45.390631 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:45 crc kubenswrapper[4754]: I0126 16:48:45.390648 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:45Z","lastTransitionTime":"2026-01-26T16:48:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:45 crc kubenswrapper[4754]: I0126 16:48:45.493717 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:45 crc kubenswrapper[4754]: I0126 16:48:45.493759 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:45 crc kubenswrapper[4754]: I0126 16:48:45.493769 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:45 crc kubenswrapper[4754]: I0126 16:48:45.493796 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:45 crc kubenswrapper[4754]: I0126 16:48:45.493808 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:45Z","lastTransitionTime":"2026-01-26T16:48:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:45 crc kubenswrapper[4754]: I0126 16:48:45.596529 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:45 crc kubenswrapper[4754]: I0126 16:48:45.596571 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:45 crc kubenswrapper[4754]: I0126 16:48:45.596582 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:45 crc kubenswrapper[4754]: I0126 16:48:45.596597 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:45 crc kubenswrapper[4754]: I0126 16:48:45.596608 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:45Z","lastTransitionTime":"2026-01-26T16:48:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:45 crc kubenswrapper[4754]: I0126 16:48:45.699734 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:45 crc kubenswrapper[4754]: I0126 16:48:45.699786 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:45 crc kubenswrapper[4754]: I0126 16:48:45.699796 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:45 crc kubenswrapper[4754]: I0126 16:48:45.699814 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:45 crc kubenswrapper[4754]: I0126 16:48:45.699825 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:45Z","lastTransitionTime":"2026-01-26T16:48:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:45 crc kubenswrapper[4754]: I0126 16:48:45.767259 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 16:48:45 crc kubenswrapper[4754]: I0126 16:48:45.767418 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 16:48:45 crc kubenswrapper[4754]: I0126 16:48:45.767562 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 16:48:45 crc kubenswrapper[4754]: E0126 16:48:45.767555 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 16:48:45 crc kubenswrapper[4754]: I0126 16:48:45.767614 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-t4ps6" Jan 26 16:48:45 crc kubenswrapper[4754]: E0126 16:48:45.767764 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 16:48:45 crc kubenswrapper[4754]: E0126 16:48:45.767932 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-t4ps6" podUID="d45179f0-9f81-4d77-8805-6b05bdb32f8f" Jan 26 16:48:45 crc kubenswrapper[4754]: E0126 16:48:45.768038 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 16:48:45 crc kubenswrapper[4754]: I0126 16:48:45.802257 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:45 crc kubenswrapper[4754]: I0126 16:48:45.802316 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:45 crc kubenswrapper[4754]: I0126 16:48:45.802334 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:45 crc kubenswrapper[4754]: I0126 16:48:45.802355 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:45 crc kubenswrapper[4754]: I0126 16:48:45.802367 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:45Z","lastTransitionTime":"2026-01-26T16:48:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:45 crc kubenswrapper[4754]: I0126 16:48:45.904959 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:45 crc kubenswrapper[4754]: I0126 16:48:45.905015 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:45 crc kubenswrapper[4754]: I0126 16:48:45.905027 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:45 crc kubenswrapper[4754]: I0126 16:48:45.905042 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:45 crc kubenswrapper[4754]: I0126 16:48:45.905053 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:45Z","lastTransitionTime":"2026-01-26T16:48:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:46 crc kubenswrapper[4754]: I0126 16:48:46.007557 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:46 crc kubenswrapper[4754]: I0126 16:48:46.007600 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:46 crc kubenswrapper[4754]: I0126 16:48:46.007611 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:46 crc kubenswrapper[4754]: I0126 16:48:46.007627 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:46 crc kubenswrapper[4754]: I0126 16:48:46.007638 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:46Z","lastTransitionTime":"2026-01-26T16:48:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:46 crc kubenswrapper[4754]: I0126 16:48:46.110719 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:46 crc kubenswrapper[4754]: I0126 16:48:46.110795 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:46 crc kubenswrapper[4754]: I0126 16:48:46.110807 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:46 crc kubenswrapper[4754]: I0126 16:48:46.110826 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:46 crc kubenswrapper[4754]: I0126 16:48:46.110837 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:46Z","lastTransitionTime":"2026-01-26T16:48:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:46 crc kubenswrapper[4754]: I0126 16:48:46.214283 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:46 crc kubenswrapper[4754]: I0126 16:48:46.214348 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:46 crc kubenswrapper[4754]: I0126 16:48:46.214369 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:46 crc kubenswrapper[4754]: I0126 16:48:46.214393 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:46 crc kubenswrapper[4754]: I0126 16:48:46.214412 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:46Z","lastTransitionTime":"2026-01-26T16:48:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:46 crc kubenswrapper[4754]: I0126 16:48:46.300246 4754 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-04 22:03:29.04529161 +0000 UTC Jan 26 16:48:46 crc kubenswrapper[4754]: I0126 16:48:46.316213 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:46 crc kubenswrapper[4754]: I0126 16:48:46.316246 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:46 crc kubenswrapper[4754]: I0126 16:48:46.316255 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:46 crc kubenswrapper[4754]: I0126 16:48:46.316268 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:46 crc kubenswrapper[4754]: I0126 16:48:46.316280 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:46Z","lastTransitionTime":"2026-01-26T16:48:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:46 crc kubenswrapper[4754]: I0126 16:48:46.418634 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:46 crc kubenswrapper[4754]: I0126 16:48:46.418706 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:46 crc kubenswrapper[4754]: I0126 16:48:46.418717 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:46 crc kubenswrapper[4754]: I0126 16:48:46.418734 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:46 crc kubenswrapper[4754]: I0126 16:48:46.418797 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:46Z","lastTransitionTime":"2026-01-26T16:48:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:46 crc kubenswrapper[4754]: I0126 16:48:46.521466 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:46 crc kubenswrapper[4754]: I0126 16:48:46.521551 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:46 crc kubenswrapper[4754]: I0126 16:48:46.521568 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:46 crc kubenswrapper[4754]: I0126 16:48:46.521597 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:46 crc kubenswrapper[4754]: I0126 16:48:46.521609 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:46Z","lastTransitionTime":"2026-01-26T16:48:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:46 crc kubenswrapper[4754]: I0126 16:48:46.624599 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:46 crc kubenswrapper[4754]: I0126 16:48:46.624646 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:46 crc kubenswrapper[4754]: I0126 16:48:46.624658 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:46 crc kubenswrapper[4754]: I0126 16:48:46.624698 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:46 crc kubenswrapper[4754]: I0126 16:48:46.624710 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:46Z","lastTransitionTime":"2026-01-26T16:48:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:46 crc kubenswrapper[4754]: I0126 16:48:46.625829 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:48:46 crc kubenswrapper[4754]: I0126 16:48:46.625957 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:48:46 crc kubenswrapper[4754]: I0126 16:48:46.626032 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:48:46 crc kubenswrapper[4754]: I0126 16:48:46.626112 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:48:46 crc kubenswrapper[4754]: I0126 16:48:46.626191 4754 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:48:46Z","lastTransitionTime":"2026-01-26T16:48:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:48:46 crc kubenswrapper[4754]: I0126 16:48:46.674368 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-mzlnf"] Jan 26 16:48:46 crc kubenswrapper[4754]: I0126 16:48:46.674779 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mzlnf" Jan 26 16:48:46 crc kubenswrapper[4754]: I0126 16:48:46.677532 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Jan 26 16:48:46 crc kubenswrapper[4754]: I0126 16:48:46.677901 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Jan 26 16:48:46 crc kubenswrapper[4754]: I0126 16:48:46.677926 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Jan 26 16:48:46 crc kubenswrapper[4754]: I0126 16:48:46.679127 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Jan 26 16:48:46 crc kubenswrapper[4754]: I0126 16:48:46.747700 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7ff25627-5f82-4f4a-af84-f24fdddabc6b-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-mzlnf\" (UID: \"7ff25627-5f82-4f4a-af84-f24fdddabc6b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mzlnf" Jan 26 16:48:46 crc kubenswrapper[4754]: I0126 16:48:46.747762 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/7ff25627-5f82-4f4a-af84-f24fdddabc6b-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-mzlnf\" (UID: \"7ff25627-5f82-4f4a-af84-f24fdddabc6b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mzlnf" Jan 26 16:48:46 crc kubenswrapper[4754]: I0126 16:48:46.747792 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/7ff25627-5f82-4f4a-af84-f24fdddabc6b-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-mzlnf\" (UID: \"7ff25627-5f82-4f4a-af84-f24fdddabc6b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mzlnf" Jan 26 16:48:46 crc kubenswrapper[4754]: I0126 16:48:46.747814 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/7ff25627-5f82-4f4a-af84-f24fdddabc6b-service-ca\") pod \"cluster-version-operator-5c965bbfc6-mzlnf\" (UID: \"7ff25627-5f82-4f4a-af84-f24fdddabc6b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mzlnf" Jan 26 16:48:46 crc kubenswrapper[4754]: I0126 16:48:46.747835 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7ff25627-5f82-4f4a-af84-f24fdddabc6b-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-mzlnf\" (UID: \"7ff25627-5f82-4f4a-af84-f24fdddabc6b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mzlnf" Jan 26 16:48:46 crc kubenswrapper[4754]: I0126 16:48:46.852337 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/7ff25627-5f82-4f4a-af84-f24fdddabc6b-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-mzlnf\" (UID: \"7ff25627-5f82-4f4a-af84-f24fdddabc6b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mzlnf" Jan 26 16:48:46 crc kubenswrapper[4754]: I0126 16:48:46.852488 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/7ff25627-5f82-4f4a-af84-f24fdddabc6b-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-mzlnf\" (UID: \"7ff25627-5f82-4f4a-af84-f24fdddabc6b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mzlnf" Jan 26 16:48:46 crc kubenswrapper[4754]: I0126 16:48:46.852554 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/7ff25627-5f82-4f4a-af84-f24fdddabc6b-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-mzlnf\" (UID: \"7ff25627-5f82-4f4a-af84-f24fdddabc6b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mzlnf" Jan 26 16:48:46 crc kubenswrapper[4754]: I0126 16:48:46.852576 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/7ff25627-5f82-4f4a-af84-f24fdddabc6b-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-mzlnf\" (UID: \"7ff25627-5f82-4f4a-af84-f24fdddabc6b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mzlnf" Jan 26 16:48:46 crc kubenswrapper[4754]: I0126 16:48:46.852598 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7ff25627-5f82-4f4a-af84-f24fdddabc6b-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-mzlnf\" (UID: \"7ff25627-5f82-4f4a-af84-f24fdddabc6b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mzlnf" Jan 26 16:48:46 crc kubenswrapper[4754]: I0126 16:48:46.852771 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/7ff25627-5f82-4f4a-af84-f24fdddabc6b-service-ca\") pod \"cluster-version-operator-5c965bbfc6-mzlnf\" (UID: \"7ff25627-5f82-4f4a-af84-f24fdddabc6b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mzlnf" Jan 26 16:48:46 crc kubenswrapper[4754]: I0126 16:48:46.852834 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7ff25627-5f82-4f4a-af84-f24fdddabc6b-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-mzlnf\" (UID: \"7ff25627-5f82-4f4a-af84-f24fdddabc6b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mzlnf" Jan 26 16:48:46 crc kubenswrapper[4754]: I0126 16:48:46.854034 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/7ff25627-5f82-4f4a-af84-f24fdddabc6b-service-ca\") pod \"cluster-version-operator-5c965bbfc6-mzlnf\" (UID: \"7ff25627-5f82-4f4a-af84-f24fdddabc6b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mzlnf" Jan 26 16:48:46 crc kubenswrapper[4754]: I0126 16:48:46.861505 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7ff25627-5f82-4f4a-af84-f24fdddabc6b-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-mzlnf\" (UID: \"7ff25627-5f82-4f4a-af84-f24fdddabc6b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mzlnf" Jan 26 16:48:46 crc kubenswrapper[4754]: I0126 16:48:46.871824 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7ff25627-5f82-4f4a-af84-f24fdddabc6b-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-mzlnf\" (UID: \"7ff25627-5f82-4f4a-af84-f24fdddabc6b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mzlnf" Jan 26 16:48:46 crc kubenswrapper[4754]: I0126 16:48:46.987767 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mzlnf" Jan 26 16:48:47 crc kubenswrapper[4754]: I0126 16:48:47.301219 4754 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-17 05:58:23.897786818 +0000 UTC Jan 26 16:48:47 crc kubenswrapper[4754]: I0126 16:48:47.301298 4754 certificate_manager.go:356] kubernetes.io/kubelet-serving: Rotating certificates Jan 26 16:48:47 crc kubenswrapper[4754]: I0126 16:48:47.310821 4754 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Jan 26 16:48:47 crc kubenswrapper[4754]: I0126 16:48:47.392313 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mzlnf" event={"ID":"7ff25627-5f82-4f4a-af84-f24fdddabc6b","Type":"ContainerStarted","Data":"96fd145ae6b082d5b56735e4db0621082225c484f8822fe7ed2b6f4710182597"} Jan 26 16:48:47 crc kubenswrapper[4754]: I0126 16:48:47.392380 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mzlnf" event={"ID":"7ff25627-5f82-4f4a-af84-f24fdddabc6b","Type":"ContainerStarted","Data":"28c8032dda130eae1c6f3ba453a2c2a445fed6d8d17c85b0c547534a03a66d7c"} Jan 26 16:48:47 crc kubenswrapper[4754]: I0126 16:48:47.767049 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 16:48:47 crc kubenswrapper[4754]: I0126 16:48:47.767080 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-t4ps6" Jan 26 16:48:47 crc kubenswrapper[4754]: I0126 16:48:47.767069 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 16:48:47 crc kubenswrapper[4754]: I0126 16:48:47.767048 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 16:48:47 crc kubenswrapper[4754]: E0126 16:48:47.767174 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 16:48:47 crc kubenswrapper[4754]: E0126 16:48:47.767307 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 16:48:47 crc kubenswrapper[4754]: E0126 16:48:47.767382 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-t4ps6" podUID="d45179f0-9f81-4d77-8805-6b05bdb32f8f" Jan 26 16:48:47 crc kubenswrapper[4754]: E0126 16:48:47.767448 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 16:48:49 crc kubenswrapper[4754]: I0126 16:48:49.766801 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 16:48:49 crc kubenswrapper[4754]: I0126 16:48:49.766818 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-t4ps6" Jan 26 16:48:49 crc kubenswrapper[4754]: I0126 16:48:49.766863 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 16:48:49 crc kubenswrapper[4754]: I0126 16:48:49.766934 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 16:48:49 crc kubenswrapper[4754]: E0126 16:48:49.767123 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-t4ps6" podUID="d45179f0-9f81-4d77-8805-6b05bdb32f8f" Jan 26 16:48:49 crc kubenswrapper[4754]: E0126 16:48:49.767539 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 16:48:49 crc kubenswrapper[4754]: E0126 16:48:49.767834 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 16:48:49 crc kubenswrapper[4754]: E0126 16:48:49.767847 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 16:48:49 crc kubenswrapper[4754]: I0126 16:48:49.786551 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mzlnf" podStartSLOduration=76.786529329 podStartE2EDuration="1m16.786529329s" podCreationTimestamp="2026-01-26 16:47:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 16:48:47.406180233 +0000 UTC m=+93.930360677" watchObservedRunningTime="2026-01-26 16:48:49.786529329 +0000 UTC m=+96.310709763" Jan 26 16:48:49 crc kubenswrapper[4754]: I0126 16:48:49.787791 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Jan 26 16:48:50 crc kubenswrapper[4754]: I0126 16:48:50.767651 4754 scope.go:117] "RemoveContainer" containerID="2c059b7d618442c8e541bce1d2e576de3e9d9f9b3bd82973f0039fc6c8429d6e" Jan 26 16:48:50 crc kubenswrapper[4754]: E0126 16:48:50.768326 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-jsbxt_openshift-ovn-kubernetes(689569c4-93ee-4b82-93fc-9af0f97e6dc3)\"" pod="openshift-ovn-kubernetes/ovnkube-node-jsbxt" podUID="689569c4-93ee-4b82-93fc-9af0f97e6dc3" Jan 26 16:48:51 crc kubenswrapper[4754]: I0126 16:48:51.767186 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-t4ps6" Jan 26 16:48:51 crc kubenswrapper[4754]: E0126 16:48:51.767394 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-t4ps6" podUID="d45179f0-9f81-4d77-8805-6b05bdb32f8f" Jan 26 16:48:51 crc kubenswrapper[4754]: I0126 16:48:51.767780 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 16:48:51 crc kubenswrapper[4754]: I0126 16:48:51.767866 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 16:48:51 crc kubenswrapper[4754]: E0126 16:48:51.767915 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 16:48:51 crc kubenswrapper[4754]: I0126 16:48:51.767942 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 16:48:51 crc kubenswrapper[4754]: E0126 16:48:51.768041 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 16:48:51 crc kubenswrapper[4754]: E0126 16:48:51.768205 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 16:48:52 crc kubenswrapper[4754]: I0126 16:48:52.415008 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d45179f0-9f81-4d77-8805-6b05bdb32f8f-metrics-certs\") pod \"network-metrics-daemon-t4ps6\" (UID: \"d45179f0-9f81-4d77-8805-6b05bdb32f8f\") " pod="openshift-multus/network-metrics-daemon-t4ps6" Jan 26 16:48:52 crc kubenswrapper[4754]: E0126 16:48:52.415191 4754 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Jan 26 16:48:52 crc kubenswrapper[4754]: E0126 16:48:52.415284 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d45179f0-9f81-4d77-8805-6b05bdb32f8f-metrics-certs podName:d45179f0-9f81-4d77-8805-6b05bdb32f8f nodeName:}" failed. No retries permitted until 2026-01-26 16:49:56.415263684 +0000 UTC m=+162.939444248 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/d45179f0-9f81-4d77-8805-6b05bdb32f8f-metrics-certs") pod "network-metrics-daemon-t4ps6" (UID: "d45179f0-9f81-4d77-8805-6b05bdb32f8f") : object "openshift-multus"/"metrics-daemon-secret" not registered Jan 26 16:48:53 crc kubenswrapper[4754]: I0126 16:48:53.767168 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 16:48:53 crc kubenswrapper[4754]: I0126 16:48:53.767179 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 16:48:53 crc kubenswrapper[4754]: I0126 16:48:53.767183 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 16:48:53 crc kubenswrapper[4754]: I0126 16:48:53.767201 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-t4ps6" Jan 26 16:48:53 crc kubenswrapper[4754]: E0126 16:48:53.768380 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 16:48:53 crc kubenswrapper[4754]: E0126 16:48:53.768522 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 16:48:53 crc kubenswrapper[4754]: E0126 16:48:53.768635 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 16:48:53 crc kubenswrapper[4754]: E0126 16:48:53.768727 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-t4ps6" podUID="d45179f0-9f81-4d77-8805-6b05bdb32f8f" Jan 26 16:48:53 crc kubenswrapper[4754]: I0126 16:48:53.802511 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=4.802486259 podStartE2EDuration="4.802486259s" podCreationTimestamp="2026-01-26 16:48:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 16:48:53.80134728 +0000 UTC m=+100.325527714" watchObservedRunningTime="2026-01-26 16:48:53.802486259 +0000 UTC m=+100.326666693" Jan 26 16:48:55 crc kubenswrapper[4754]: I0126 16:48:55.767086 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-t4ps6" Jan 26 16:48:55 crc kubenswrapper[4754]: I0126 16:48:55.767172 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 16:48:55 crc kubenswrapper[4754]: I0126 16:48:55.767341 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 16:48:55 crc kubenswrapper[4754]: I0126 16:48:55.767299 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 16:48:55 crc kubenswrapper[4754]: E0126 16:48:55.767875 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-t4ps6" podUID="d45179f0-9f81-4d77-8805-6b05bdb32f8f" Jan 26 16:48:55 crc kubenswrapper[4754]: E0126 16:48:55.767939 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 16:48:55 crc kubenswrapper[4754]: E0126 16:48:55.768027 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 16:48:55 crc kubenswrapper[4754]: E0126 16:48:55.768142 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 16:48:57 crc kubenswrapper[4754]: I0126 16:48:57.766586 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 16:48:57 crc kubenswrapper[4754]: I0126 16:48:57.766847 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-t4ps6" Jan 26 16:48:57 crc kubenswrapper[4754]: I0126 16:48:57.766859 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 16:48:57 crc kubenswrapper[4754]: I0126 16:48:57.766897 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 16:48:57 crc kubenswrapper[4754]: E0126 16:48:57.766917 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 16:48:57 crc kubenswrapper[4754]: E0126 16:48:57.767049 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 16:48:57 crc kubenswrapper[4754]: E0126 16:48:57.767293 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-t4ps6" podUID="d45179f0-9f81-4d77-8805-6b05bdb32f8f" Jan 26 16:48:57 crc kubenswrapper[4754]: E0126 16:48:57.767382 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 16:48:59 crc kubenswrapper[4754]: I0126 16:48:59.767354 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 16:48:59 crc kubenswrapper[4754]: I0126 16:48:59.767485 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 16:48:59 crc kubenswrapper[4754]: E0126 16:48:59.767536 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 16:48:59 crc kubenswrapper[4754]: E0126 16:48:59.767658 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 16:48:59 crc kubenswrapper[4754]: I0126 16:48:59.767770 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-t4ps6" Jan 26 16:48:59 crc kubenswrapper[4754]: E0126 16:48:59.767933 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-t4ps6" podUID="d45179f0-9f81-4d77-8805-6b05bdb32f8f" Jan 26 16:48:59 crc kubenswrapper[4754]: I0126 16:48:59.767389 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 16:48:59 crc kubenswrapper[4754]: E0126 16:48:59.768088 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 16:49:01 crc kubenswrapper[4754]: I0126 16:49:01.767014 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-t4ps6" Jan 26 16:49:01 crc kubenswrapper[4754]: I0126 16:49:01.767064 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 16:49:01 crc kubenswrapper[4754]: I0126 16:49:01.767026 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 16:49:01 crc kubenswrapper[4754]: I0126 16:49:01.767017 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 16:49:01 crc kubenswrapper[4754]: E0126 16:49:01.767157 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-t4ps6" podUID="d45179f0-9f81-4d77-8805-6b05bdb32f8f" Jan 26 16:49:01 crc kubenswrapper[4754]: E0126 16:49:01.767223 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 16:49:01 crc kubenswrapper[4754]: E0126 16:49:01.767413 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 16:49:01 crc kubenswrapper[4754]: E0126 16:49:01.767454 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 16:49:02 crc kubenswrapper[4754]: I0126 16:49:02.767900 4754 scope.go:117] "RemoveContainer" containerID="2c059b7d618442c8e541bce1d2e576de3e9d9f9b3bd82973f0039fc6c8429d6e" Jan 26 16:49:02 crc kubenswrapper[4754]: E0126 16:49:02.768191 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-jsbxt_openshift-ovn-kubernetes(689569c4-93ee-4b82-93fc-9af0f97e6dc3)\"" pod="openshift-ovn-kubernetes/ovnkube-node-jsbxt" podUID="689569c4-93ee-4b82-93fc-9af0f97e6dc3" Jan 26 16:49:03 crc kubenswrapper[4754]: I0126 16:49:03.767017 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 16:49:03 crc kubenswrapper[4754]: I0126 16:49:03.767210 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 16:49:03 crc kubenswrapper[4754]: E0126 16:49:03.768890 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 16:49:03 crc kubenswrapper[4754]: I0126 16:49:03.768932 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-t4ps6" Jan 26 16:49:03 crc kubenswrapper[4754]: I0126 16:49:03.769004 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 16:49:03 crc kubenswrapper[4754]: E0126 16:49:03.769112 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-t4ps6" podUID="d45179f0-9f81-4d77-8805-6b05bdb32f8f" Jan 26 16:49:03 crc kubenswrapper[4754]: E0126 16:49:03.769177 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 16:49:03 crc kubenswrapper[4754]: E0126 16:49:03.769305 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 16:49:05 crc kubenswrapper[4754]: I0126 16:49:05.767455 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 16:49:05 crc kubenswrapper[4754]: I0126 16:49:05.767508 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 16:49:05 crc kubenswrapper[4754]: I0126 16:49:05.767446 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-t4ps6" Jan 26 16:49:05 crc kubenswrapper[4754]: E0126 16:49:05.767660 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 16:49:05 crc kubenswrapper[4754]: E0126 16:49:05.767778 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-t4ps6" podUID="d45179f0-9f81-4d77-8805-6b05bdb32f8f" Jan 26 16:49:05 crc kubenswrapper[4754]: E0126 16:49:05.767887 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 16:49:05 crc kubenswrapper[4754]: I0126 16:49:05.767954 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 16:49:05 crc kubenswrapper[4754]: E0126 16:49:05.768093 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 16:49:07 crc kubenswrapper[4754]: I0126 16:49:07.464013 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-skh5n_b619d233-b592-4b05-a0b7-dc094c88471e/kube-multus/1.log" Jan 26 16:49:07 crc kubenswrapper[4754]: I0126 16:49:07.464751 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-skh5n_b619d233-b592-4b05-a0b7-dc094c88471e/kube-multus/0.log" Jan 26 16:49:07 crc kubenswrapper[4754]: I0126 16:49:07.464822 4754 generic.go:334] "Generic (PLEG): container finished" podID="b619d233-b592-4b05-a0b7-dc094c88471e" containerID="c8d026fdb7ea7dc5c40add18606d8f1a7703865edf93ef19edaf811f598364c5" exitCode=1 Jan 26 16:49:07 crc kubenswrapper[4754]: I0126 16:49:07.464868 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-skh5n" event={"ID":"b619d233-b592-4b05-a0b7-dc094c88471e","Type":"ContainerDied","Data":"c8d026fdb7ea7dc5c40add18606d8f1a7703865edf93ef19edaf811f598364c5"} Jan 26 16:49:07 crc kubenswrapper[4754]: I0126 16:49:07.464916 4754 scope.go:117] "RemoveContainer" containerID="9231a4191565e2e0ef53d8a394bb34615c900b601a0dfec24c2d8e5fa7b0b659" Jan 26 16:49:07 crc kubenswrapper[4754]: I0126 16:49:07.465509 4754 scope.go:117] "RemoveContainer" containerID="c8d026fdb7ea7dc5c40add18606d8f1a7703865edf93ef19edaf811f598364c5" Jan 26 16:49:07 crc kubenswrapper[4754]: E0126 16:49:07.465701 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-skh5n_openshift-multus(b619d233-b592-4b05-a0b7-dc094c88471e)\"" pod="openshift-multus/multus-skh5n" podUID="b619d233-b592-4b05-a0b7-dc094c88471e" Jan 26 16:49:07 crc kubenswrapper[4754]: I0126 16:49:07.766844 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 16:49:07 crc kubenswrapper[4754]: E0126 16:49:07.766961 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 16:49:07 crc kubenswrapper[4754]: I0126 16:49:07.767047 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-t4ps6" Jan 26 16:49:07 crc kubenswrapper[4754]: I0126 16:49:07.767060 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 16:49:07 crc kubenswrapper[4754]: E0126 16:49:07.767267 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-t4ps6" podUID="d45179f0-9f81-4d77-8805-6b05bdb32f8f" Jan 26 16:49:07 crc kubenswrapper[4754]: E0126 16:49:07.767496 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 16:49:07 crc kubenswrapper[4754]: I0126 16:49:07.767540 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 16:49:07 crc kubenswrapper[4754]: E0126 16:49:07.767708 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 16:49:08 crc kubenswrapper[4754]: I0126 16:49:08.469572 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-skh5n_b619d233-b592-4b05-a0b7-dc094c88471e/kube-multus/1.log" Jan 26 16:49:09 crc kubenswrapper[4754]: I0126 16:49:09.766822 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 16:49:09 crc kubenswrapper[4754]: I0126 16:49:09.766861 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 16:49:09 crc kubenswrapper[4754]: I0126 16:49:09.766843 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-t4ps6" Jan 26 16:49:09 crc kubenswrapper[4754]: E0126 16:49:09.767026 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 16:49:09 crc kubenswrapper[4754]: E0126 16:49:09.767112 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 16:49:09 crc kubenswrapper[4754]: I0126 16:49:09.767155 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 16:49:09 crc kubenswrapper[4754]: E0126 16:49:09.767211 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-t4ps6" podUID="d45179f0-9f81-4d77-8805-6b05bdb32f8f" Jan 26 16:49:09 crc kubenswrapper[4754]: E0126 16:49:09.767260 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 16:49:11 crc kubenswrapper[4754]: I0126 16:49:11.767404 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-t4ps6" Jan 26 16:49:11 crc kubenswrapper[4754]: E0126 16:49:11.767610 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-t4ps6" podUID="d45179f0-9f81-4d77-8805-6b05bdb32f8f" Jan 26 16:49:11 crc kubenswrapper[4754]: I0126 16:49:11.767777 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 16:49:11 crc kubenswrapper[4754]: I0126 16:49:11.767906 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 16:49:11 crc kubenswrapper[4754]: I0126 16:49:11.767944 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 16:49:11 crc kubenswrapper[4754]: E0126 16:49:11.768088 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 16:49:11 crc kubenswrapper[4754]: E0126 16:49:11.768521 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 16:49:11 crc kubenswrapper[4754]: E0126 16:49:11.768636 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 16:49:13 crc kubenswrapper[4754]: I0126 16:49:13.767366 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-t4ps6" Jan 26 16:49:13 crc kubenswrapper[4754]: I0126 16:49:13.767434 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 16:49:13 crc kubenswrapper[4754]: I0126 16:49:13.768100 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 16:49:13 crc kubenswrapper[4754]: E0126 16:49:13.768609 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-t4ps6" podUID="d45179f0-9f81-4d77-8805-6b05bdb32f8f" Jan 26 16:49:13 crc kubenswrapper[4754]: I0126 16:49:13.768687 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 16:49:13 crc kubenswrapper[4754]: E0126 16:49:13.768843 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 16:49:13 crc kubenswrapper[4754]: E0126 16:49:13.768888 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 16:49:13 crc kubenswrapper[4754]: E0126 16:49:13.768947 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 16:49:13 crc kubenswrapper[4754]: E0126 16:49:13.794410 4754 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Jan 26 16:49:13 crc kubenswrapper[4754]: E0126 16:49:13.848510 4754 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Jan 26 16:49:15 crc kubenswrapper[4754]: I0126 16:49:15.766373 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-t4ps6" Jan 26 16:49:15 crc kubenswrapper[4754]: I0126 16:49:15.766422 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 16:49:15 crc kubenswrapper[4754]: E0126 16:49:15.766520 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-t4ps6" podUID="d45179f0-9f81-4d77-8805-6b05bdb32f8f" Jan 26 16:49:15 crc kubenswrapper[4754]: I0126 16:49:15.766382 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 16:49:15 crc kubenswrapper[4754]: I0126 16:49:15.766612 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 16:49:15 crc kubenswrapper[4754]: E0126 16:49:15.766604 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 16:49:15 crc kubenswrapper[4754]: E0126 16:49:15.767084 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 16:49:15 crc kubenswrapper[4754]: E0126 16:49:15.767174 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 16:49:17 crc kubenswrapper[4754]: I0126 16:49:17.767184 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 16:49:17 crc kubenswrapper[4754]: E0126 16:49:17.768051 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 16:49:17 crc kubenswrapper[4754]: I0126 16:49:17.767211 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-t4ps6" Jan 26 16:49:17 crc kubenswrapper[4754]: E0126 16:49:17.768126 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-t4ps6" podUID="d45179f0-9f81-4d77-8805-6b05bdb32f8f" Jan 26 16:49:17 crc kubenswrapper[4754]: I0126 16:49:17.767966 4754 scope.go:117] "RemoveContainer" containerID="2c059b7d618442c8e541bce1d2e576de3e9d9f9b3bd82973f0039fc6c8429d6e" Jan 26 16:49:17 crc kubenswrapper[4754]: I0126 16:49:17.767184 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 16:49:17 crc kubenswrapper[4754]: E0126 16:49:17.768494 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 16:49:17 crc kubenswrapper[4754]: I0126 16:49:17.767234 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 16:49:17 crc kubenswrapper[4754]: E0126 16:49:17.768827 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 16:49:18 crc kubenswrapper[4754]: I0126 16:49:18.502939 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-jsbxt_689569c4-93ee-4b82-93fc-9af0f97e6dc3/ovnkube-controller/3.log" Jan 26 16:49:18 crc kubenswrapper[4754]: I0126 16:49:18.506632 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jsbxt" event={"ID":"689569c4-93ee-4b82-93fc-9af0f97e6dc3","Type":"ContainerStarted","Data":"0d6f7aa12978cc78e269af0ca6f59f424d8898f692c6f8d07454a1868b0204cd"} Jan 26 16:49:18 crc kubenswrapper[4754]: I0126 16:49:18.507531 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-jsbxt" Jan 26 16:49:18 crc kubenswrapper[4754]: I0126 16:49:18.536833 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-jsbxt" podStartSLOduration=105.536812759 podStartE2EDuration="1m45.536812759s" podCreationTimestamp="2026-01-26 16:47:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 16:49:18.535974197 +0000 UTC m=+125.060154661" watchObservedRunningTime="2026-01-26 16:49:18.536812759 +0000 UTC m=+125.060993203" Jan 26 16:49:18 crc kubenswrapper[4754]: I0126 16:49:18.610000 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-t4ps6"] Jan 26 16:49:18 crc kubenswrapper[4754]: I0126 16:49:18.610121 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-t4ps6" Jan 26 16:49:18 crc kubenswrapper[4754]: E0126 16:49:18.610214 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-t4ps6" podUID="d45179f0-9f81-4d77-8805-6b05bdb32f8f" Jan 26 16:49:18 crc kubenswrapper[4754]: E0126 16:49:18.849419 4754 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Jan 26 16:49:19 crc kubenswrapper[4754]: I0126 16:49:19.767334 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 16:49:19 crc kubenswrapper[4754]: I0126 16:49:19.767368 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-t4ps6" Jan 26 16:49:19 crc kubenswrapper[4754]: I0126 16:49:19.767334 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 16:49:19 crc kubenswrapper[4754]: E0126 16:49:19.767504 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 16:49:19 crc kubenswrapper[4754]: E0126 16:49:19.767603 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-t4ps6" podUID="d45179f0-9f81-4d77-8805-6b05bdb32f8f" Jan 26 16:49:19 crc kubenswrapper[4754]: E0126 16:49:19.767720 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 16:49:19 crc kubenswrapper[4754]: I0126 16:49:19.767722 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 16:49:19 crc kubenswrapper[4754]: E0126 16:49:19.767857 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 16:49:20 crc kubenswrapper[4754]: I0126 16:49:20.767127 4754 scope.go:117] "RemoveContainer" containerID="c8d026fdb7ea7dc5c40add18606d8f1a7703865edf93ef19edaf811f598364c5" Jan 26 16:49:21 crc kubenswrapper[4754]: I0126 16:49:21.518399 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-skh5n_b619d233-b592-4b05-a0b7-dc094c88471e/kube-multus/1.log" Jan 26 16:49:21 crc kubenswrapper[4754]: I0126 16:49:21.518468 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-skh5n" event={"ID":"b619d233-b592-4b05-a0b7-dc094c88471e","Type":"ContainerStarted","Data":"8a1283e9d0714480b093730068bfb9fc6a36c9982a88f87e70ec9fea74bb462c"} Jan 26 16:49:21 crc kubenswrapper[4754]: I0126 16:49:21.767259 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 16:49:21 crc kubenswrapper[4754]: I0126 16:49:21.767330 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-t4ps6" Jan 26 16:49:21 crc kubenswrapper[4754]: E0126 16:49:21.768005 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 16:49:21 crc kubenswrapper[4754]: I0126 16:49:21.767400 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 16:49:21 crc kubenswrapper[4754]: E0126 16:49:21.768052 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-t4ps6" podUID="d45179f0-9f81-4d77-8805-6b05bdb32f8f" Jan 26 16:49:21 crc kubenswrapper[4754]: I0126 16:49:21.767367 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 16:49:21 crc kubenswrapper[4754]: E0126 16:49:21.768145 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 16:49:21 crc kubenswrapper[4754]: E0126 16:49:21.768319 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 16:49:23 crc kubenswrapper[4754]: I0126 16:49:23.767104 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 16:49:23 crc kubenswrapper[4754]: E0126 16:49:23.768193 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 16:49:23 crc kubenswrapper[4754]: I0126 16:49:23.768279 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 16:49:23 crc kubenswrapper[4754]: I0126 16:49:23.768316 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-t4ps6" Jan 26 16:49:23 crc kubenswrapper[4754]: I0126 16:49:23.768301 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 16:49:23 crc kubenswrapper[4754]: E0126 16:49:23.768497 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 16:49:23 crc kubenswrapper[4754]: E0126 16:49:23.768540 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-t4ps6" podUID="d45179f0-9f81-4d77-8805-6b05bdb32f8f" Jan 26 16:49:23 crc kubenswrapper[4754]: E0126 16:49:23.768585 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 16:49:25 crc kubenswrapper[4754]: I0126 16:49:25.767288 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-t4ps6" Jan 26 16:49:25 crc kubenswrapper[4754]: I0126 16:49:25.767569 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 16:49:25 crc kubenswrapper[4754]: I0126 16:49:25.767612 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 16:49:25 crc kubenswrapper[4754]: I0126 16:49:25.767848 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 16:49:25 crc kubenswrapper[4754]: I0126 16:49:25.770766 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Jan 26 16:49:25 crc kubenswrapper[4754]: I0126 16:49:25.770866 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Jan 26 16:49:25 crc kubenswrapper[4754]: I0126 16:49:25.771005 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Jan 26 16:49:25 crc kubenswrapper[4754]: I0126 16:49:25.771057 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Jan 26 16:49:25 crc kubenswrapper[4754]: I0126 16:49:25.771278 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Jan 26 16:49:25 crc kubenswrapper[4754]: I0126 16:49:25.771284 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.121989 4754 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.156648 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-kh4hh"] Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.157185 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-kh4hh" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.160813 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.161060 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.161241 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.161348 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-mj6ns"] Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.161850 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zvqdh"] Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.162147 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zvqdh" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.162485 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-mj6ns" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.162487 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-m2525"] Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.163053 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m2525" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.163930 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.164041 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.164358 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.167418 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-kjc2g"] Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.167844 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-677j9"] Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.168138 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-677j9" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.168433 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-kjc2g" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.169375 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-wjdgm"] Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.169969 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-ptrdc"] Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.170467 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-ptrdc" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.170846 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-wjdgm" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.170962 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-g64xp"] Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.171520 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-g64xp" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.172194 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.172340 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.172381 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.172473 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.172539 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.172559 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.172661 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.172695 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.172720 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.172765 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.172830 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.172854 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.172877 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.172923 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.172967 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.172990 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.173038 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.175476 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.183889 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.184324 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.184961 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.185427 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.186044 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.186250 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.186399 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.186482 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.186613 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.187530 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.187730 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-8d467"] Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.188491 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-8d467" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.189790 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.189921 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.195084 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.195517 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.200947 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.201138 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-k9sz7"] Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.202142 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-z2k6h"] Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.203259 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.203363 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-z2k6h" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.203768 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-k9sz7" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.203888 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.203274 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-6mj2d"] Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.205946 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6mj2d" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.214563 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.223189 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.223375 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.223493 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.224017 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.224209 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.224281 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-g4dbx"] Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.224711 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-5p6w2"] Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.225037 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-5p6w2" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.225339 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-g4dbx" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.225606 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-rcrn5"] Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.226918 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.230174 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-rcrn5" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.233643 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.235294 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-4hkb9"] Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.235427 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.235734 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.235773 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.235862 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.235954 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.235972 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.236057 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.236075 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.236085 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-47pnn"] Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.236164 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.236175 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-4hkb9" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.236294 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.236391 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.236166 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.236698 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.236721 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.236729 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-47pnn" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.236819 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.237166 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.237317 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.237374 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.237428 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.237511 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.237548 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.237577 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.238261 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.238349 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-t4r22"] Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.238416 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.238607 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.238659 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.238863 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-t4r22" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.245096 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.245290 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.245466 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.245578 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.245680 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.245773 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.245844 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.245892 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.245984 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.246009 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.245780 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.246069 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.246010 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.245122 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.246144 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.246154 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.245865 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.246237 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.247365 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.247493 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.247723 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-fj2tb"] Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.248679 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-sz8hj"] Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.249109 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-sz8hj" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.249341 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.251149 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-fj2tb" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.251473 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.268616 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7lgch"] Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.269723 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7lgch" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.278114 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/39a0ee2d-0f5b-4231-94a3-d11d517d3517-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-ptrdc\" (UID: \"39a0ee2d-0f5b-4231-94a3-d11d517d3517\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-ptrdc" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.278167 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/2a28f341-5f93-4d6c-bc4b-973d33dfeee1-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-k9sz7\" (UID: \"2a28f341-5f93-4d6c-bc4b-973d33dfeee1\") " pod="openshift-authentication/oauth-openshift-558db77b4-k9sz7" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.278193 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/2a28f341-5f93-4d6c-bc4b-973d33dfeee1-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-k9sz7\" (UID: \"2a28f341-5f93-4d6c-bc4b-973d33dfeee1\") " pod="openshift-authentication/oauth-openshift-558db77b4-k9sz7" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.278216 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/2a28f341-5f93-4d6c-bc4b-973d33dfeee1-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-k9sz7\" (UID: \"2a28f341-5f93-4d6c-bc4b-973d33dfeee1\") " pod="openshift-authentication/oauth-openshift-558db77b4-k9sz7" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.278242 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/10609d5f-d6dd-40a5-acc4-63e62584d18e-trusted-ca-bundle\") pod \"apiserver-76f77b778f-8d467\" (UID: \"10609d5f-d6dd-40a5-acc4-63e62584d18e\") " pod="openshift-apiserver/apiserver-76f77b778f-8d467" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.278284 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-llgrp\" (UniqueName: \"kubernetes.io/projected/587cde2e-d0a4-406b-ab72-26c7fe4b6707-kube-api-access-llgrp\") pod \"route-controller-manager-6576b87f9c-m2525\" (UID: \"587cde2e-d0a4-406b-ab72-26c7fe4b6707\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m2525" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.278312 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/10609d5f-d6dd-40a5-acc4-63e62584d18e-audit\") pod \"apiserver-76f77b778f-8d467\" (UID: \"10609d5f-d6dd-40a5-acc4-63e62584d18e\") " pod="openshift-apiserver/apiserver-76f77b778f-8d467" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.278334 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/10609d5f-d6dd-40a5-acc4-63e62584d18e-audit-dir\") pod \"apiserver-76f77b778f-8d467\" (UID: \"10609d5f-d6dd-40a5-acc4-63e62584d18e\") " pod="openshift-apiserver/apiserver-76f77b778f-8d467" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.278362 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rqfpq\" (UniqueName: \"kubernetes.io/projected/10609d5f-d6dd-40a5-acc4-63e62584d18e-kube-api-access-rqfpq\") pod \"apiserver-76f77b778f-8d467\" (UID: \"10609d5f-d6dd-40a5-acc4-63e62584d18e\") " pod="openshift-apiserver/apiserver-76f77b778f-8d467" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.278384 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/aa57d7bf-eaea-4595-9d00-5d5d4b20f059-etcd-ca\") pod \"etcd-operator-b45778765-sz8hj\" (UID: \"aa57d7bf-eaea-4595-9d00-5d5d4b20f059\") " pod="openshift-etcd-operator/etcd-operator-b45778765-sz8hj" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.278410 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/587cde2e-d0a4-406b-ab72-26c7fe4b6707-client-ca\") pod \"route-controller-manager-6576b87f9c-m2525\" (UID: \"587cde2e-d0a4-406b-ab72-26c7fe4b6707\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m2525" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.278436 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/c31a7288-dcad-4e03-beb6-22e7855a58cd-auth-proxy-config\") pod \"machine-config-operator-74547568cd-4hkb9\" (UID: \"c31a7288-dcad-4e03-beb6-22e7855a58cd\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-4hkb9" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.278463 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/10609d5f-d6dd-40a5-acc4-63e62584d18e-config\") pod \"apiserver-76f77b778f-8d467\" (UID: \"10609d5f-d6dd-40a5-acc4-63e62584d18e\") " pod="openshift-apiserver/apiserver-76f77b778f-8d467" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.278496 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/10609d5f-d6dd-40a5-acc4-63e62584d18e-etcd-client\") pod \"apiserver-76f77b778f-8d467\" (UID: \"10609d5f-d6dd-40a5-acc4-63e62584d18e\") " pod="openshift-apiserver/apiserver-76f77b778f-8d467" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.278546 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/bd64c275-cf75-4e42-8643-9178665801cb-audit-policies\") pod \"apiserver-7bbb656c7d-6mj2d\" (UID: \"bd64c275-cf75-4e42-8643-9178665801cb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6mj2d" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.278588 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/2a28f341-5f93-4d6c-bc4b-973d33dfeee1-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-k9sz7\" (UID: \"2a28f341-5f93-4d6c-bc4b-973d33dfeee1\") " pod="openshift-authentication/oauth-openshift-558db77b4-k9sz7" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.278612 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f4d41d59-ebec-455b-9a17-2ff3f9a42f04-service-ca-bundle\") pod \"authentication-operator-69f744f599-kjc2g\" (UID: \"f4d41d59-ebec-455b-9a17-2ff3f9a42f04\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-kjc2g" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.278632 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/99582129-e50d-49c0-a554-aa703fc86452-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-z2k6h\" (UID: \"99582129-e50d-49c0-a554-aa703fc86452\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-z2k6h" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.278658 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5c9zf\" (UniqueName: \"kubernetes.io/projected/aa57d7bf-eaea-4595-9d00-5d5d4b20f059-kube-api-access-5c9zf\") pod \"etcd-operator-b45778765-sz8hj\" (UID: \"aa57d7bf-eaea-4595-9d00-5d5d4b20f059\") " pod="openshift-etcd-operator/etcd-operator-b45778765-sz8hj" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.278697 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/06082490-21d0-4f81-8f02-b5ad80760b06-config\") pod \"machine-approver-56656f9798-kh4hh\" (UID: \"06082490-21d0-4f81-8f02-b5ad80760b06\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-kh4hh" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.278717 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l58br\" (UniqueName: \"kubernetes.io/projected/99582129-e50d-49c0-a554-aa703fc86452-kube-api-access-l58br\") pod \"cluster-image-registry-operator-dc59b4c8b-z2k6h\" (UID: \"99582129-e50d-49c0-a554-aa703fc86452\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-z2k6h" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.278761 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/bd64c275-cf75-4e42-8643-9178665801cb-audit-dir\") pod \"apiserver-7bbb656c7d-6mj2d\" (UID: \"bd64c275-cf75-4e42-8643-9178665801cb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6mj2d" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.278817 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/adeda50d-627e-4071-a6ab-d7712750c22d-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-zvqdh\" (UID: \"adeda50d-627e-4071-a6ab-d7712750c22d\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zvqdh" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.278936 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/528dc2c4-71cc-46cb-b847-426bb8638188-trusted-ca-bundle\") pod \"console-f9d7485db-5p6w2\" (UID: \"528dc2c4-71cc-46cb-b847-426bb8638188\") " pod="openshift-console/console-f9d7485db-5p6w2" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.286320 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.286494 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.279591 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/06082490-21d0-4f81-8f02-b5ad80760b06-auth-proxy-config\") pod \"machine-approver-56656f9798-kh4hh\" (UID: \"06082490-21d0-4f81-8f02-b5ad80760b06\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-kh4hh" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.286760 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vdvx6\" (UniqueName: \"kubernetes.io/projected/4d81d21a-83f5-402e-91fd-d8268a6b49be-kube-api-access-vdvx6\") pod \"controller-manager-879f6c89f-677j9\" (UID: \"4d81d21a-83f5-402e-91fd-d8268a6b49be\") " pod="openshift-controller-manager/controller-manager-879f6c89f-677j9" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.286789 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/06082490-21d0-4f81-8f02-b5ad80760b06-machine-approver-tls\") pod \"machine-approver-56656f9798-kh4hh\" (UID: \"06082490-21d0-4f81-8f02-b5ad80760b06\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-kh4hh" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.286972 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/2a28f341-5f93-4d6c-bc4b-973d33dfeee1-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-k9sz7\" (UID: \"2a28f341-5f93-4d6c-bc4b-973d33dfeee1\") " pod="openshift-authentication/oauth-openshift-558db77b4-k9sz7" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.287065 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-llv5g\" (UniqueName: \"kubernetes.io/projected/adeda50d-627e-4071-a6ab-d7712750c22d-kube-api-access-llv5g\") pod \"openshift-apiserver-operator-796bbdcf4f-zvqdh\" (UID: \"adeda50d-627e-4071-a6ab-d7712750c22d\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zvqdh" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.287103 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/2a28f341-5f93-4d6c-bc4b-973d33dfeee1-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-k9sz7\" (UID: \"2a28f341-5f93-4d6c-bc4b-973d33dfeee1\") " pod="openshift-authentication/oauth-openshift-558db77b4-k9sz7" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.287133 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/2a28f341-5f93-4d6c-bc4b-973d33dfeee1-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-k9sz7\" (UID: \"2a28f341-5f93-4d6c-bc4b-973d33dfeee1\") " pod="openshift-authentication/oauth-openshift-558db77b4-k9sz7" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.287168 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/6b062c17-08ee-4fda-94d2-a59a3cdf8641-metrics-tls\") pod \"dns-operator-744455d44c-fj2tb\" (UID: \"6b062c17-08ee-4fda-94d2-a59a3cdf8641\") " pod="openshift-dns-operator/dns-operator-744455d44c-fj2tb" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.287194 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/10609d5f-d6dd-40a5-acc4-63e62584d18e-etcd-serving-ca\") pod \"apiserver-76f77b778f-8d467\" (UID: \"10609d5f-d6dd-40a5-acc4-63e62584d18e\") " pod="openshift-apiserver/apiserver-76f77b778f-8d467" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.287214 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/528dc2c4-71cc-46cb-b847-426bb8638188-console-config\") pod \"console-f9d7485db-5p6w2\" (UID: \"528dc2c4-71cc-46cb-b847-426bb8638188\") " pod="openshift-console/console-f9d7485db-5p6w2" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.287241 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/bd64c275-cf75-4e42-8643-9178665801cb-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-6mj2d\" (UID: \"bd64c275-cf75-4e42-8643-9178665801cb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6mj2d" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.287360 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8jdh4\" (UniqueName: \"kubernetes.io/projected/c31a7288-dcad-4e03-beb6-22e7855a58cd-kube-api-access-8jdh4\") pod \"machine-config-operator-74547568cd-4hkb9\" (UID: \"c31a7288-dcad-4e03-beb6-22e7855a58cd\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-4hkb9" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.287912 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-8647g"] Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.288386 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/aa57d7bf-eaea-4595-9d00-5d5d4b20f059-serving-cert\") pod \"etcd-operator-b45778765-sz8hj\" (UID: \"aa57d7bf-eaea-4595-9d00-5d5d4b20f059\") " pod="openshift-etcd-operator/etcd-operator-b45778765-sz8hj" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.288428 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/aa57d7bf-eaea-4595-9d00-5d5d4b20f059-etcd-service-ca\") pod \"etcd-operator-b45778765-sz8hj\" (UID: \"aa57d7bf-eaea-4595-9d00-5d5d4b20f059\") " pod="openshift-etcd-operator/etcd-operator-b45778765-sz8hj" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.288472 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4d81d21a-83f5-402e-91fd-d8268a6b49be-serving-cert\") pod \"controller-manager-879f6c89f-677j9\" (UID: \"4d81d21a-83f5-402e-91fd-d8268a6b49be\") " pod="openshift-controller-manager/controller-manager-879f6c89f-677j9" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.288492 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/528dc2c4-71cc-46cb-b847-426bb8638188-console-oauth-config\") pod \"console-f9d7485db-5p6w2\" (UID: \"528dc2c4-71cc-46cb-b847-426bb8638188\") " pod="openshift-console/console-f9d7485db-5p6w2" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.288514 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f4d41d59-ebec-455b-9a17-2ff3f9a42f04-config\") pod \"authentication-operator-69f744f599-kjc2g\" (UID: \"f4d41d59-ebec-455b-9a17-2ff3f9a42f04\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-kjc2g" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.288539 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8zx6c\" (UniqueName: \"kubernetes.io/projected/78242150-7a64-4766-808f-fa460cd84b20-kube-api-access-8zx6c\") pod \"downloads-7954f5f757-rcrn5\" (UID: \"78242150-7a64-4766-808f-fa460cd84b20\") " pod="openshift-console/downloads-7954f5f757-rcrn5" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.288562 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/10609d5f-d6dd-40a5-acc4-63e62584d18e-image-import-ca\") pod \"apiserver-76f77b778f-8d467\" (UID: \"10609d5f-d6dd-40a5-acc4-63e62584d18e\") " pod="openshift-apiserver/apiserver-76f77b778f-8d467" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.288592 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5jq47\" (UniqueName: \"kubernetes.io/projected/528dc2c4-71cc-46cb-b847-426bb8638188-kube-api-access-5jq47\") pod \"console-f9d7485db-5p6w2\" (UID: \"528dc2c4-71cc-46cb-b847-426bb8638188\") " pod="openshift-console/console-f9d7485db-5p6w2" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.288624 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-8647g" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.288639 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4d81d21a-83f5-402e-91fd-d8268a6b49be-config\") pod \"controller-manager-879f6c89f-677j9\" (UID: \"4d81d21a-83f5-402e-91fd-d8268a6b49be\") " pod="openshift-controller-manager/controller-manager-879f6c89f-677j9" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.288724 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qb27c\" (UniqueName: \"kubernetes.io/projected/66592797-8f8e-4e3c-938b-aa6767928e3f-kube-api-access-qb27c\") pod \"openshift-config-operator-7777fb866f-mj6ns\" (UID: \"66592797-8f8e-4e3c-938b-aa6767928e3f\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-mj6ns" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.288754 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/10609d5f-d6dd-40a5-acc4-63e62584d18e-node-pullsecrets\") pod \"apiserver-76f77b778f-8d467\" (UID: \"10609d5f-d6dd-40a5-acc4-63e62584d18e\") " pod="openshift-apiserver/apiserver-76f77b778f-8d467" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.288795 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/adeda50d-627e-4071-a6ab-d7712750c22d-config\") pod \"openshift-apiserver-operator-796bbdcf4f-zvqdh\" (UID: \"adeda50d-627e-4071-a6ab-d7712750c22d\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zvqdh" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.288817 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t7bg2\" (UniqueName: \"kubernetes.io/projected/06082490-21d0-4f81-8f02-b5ad80760b06-kube-api-access-t7bg2\") pod \"machine-approver-56656f9798-kh4hh\" (UID: \"06082490-21d0-4f81-8f02-b5ad80760b06\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-kh4hh" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.288842 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/c31a7288-dcad-4e03-beb6-22e7855a58cd-images\") pod \"machine-config-operator-74547568cd-4hkb9\" (UID: \"c31a7288-dcad-4e03-beb6-22e7855a58cd\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-4hkb9" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.288883 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/66592797-8f8e-4e3c-938b-aa6767928e3f-available-featuregates\") pod \"openshift-config-operator-7777fb866f-mj6ns\" (UID: \"66592797-8f8e-4e3c-938b-aa6767928e3f\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-mj6ns" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.288904 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8tpz9\" (UniqueName: \"kubernetes.io/projected/39a0ee2d-0f5b-4231-94a3-d11d517d3517-kube-api-access-8tpz9\") pod \"cluster-samples-operator-665b6dd947-ptrdc\" (UID: \"39a0ee2d-0f5b-4231-94a3-d11d517d3517\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-ptrdc" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.288933 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/2a28f341-5f93-4d6c-bc4b-973d33dfeee1-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-k9sz7\" (UID: \"2a28f341-5f93-4d6c-bc4b-973d33dfeee1\") " pod="openshift-authentication/oauth-openshift-558db77b4-k9sz7" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.288954 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/aa57d7bf-eaea-4595-9d00-5d5d4b20f059-etcd-client\") pod \"etcd-operator-b45778765-sz8hj\" (UID: \"aa57d7bf-eaea-4595-9d00-5d5d4b20f059\") " pod="openshift-etcd-operator/etcd-operator-b45778765-sz8hj" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.288974 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/bd64c275-cf75-4e42-8643-9178665801cb-encryption-config\") pod \"apiserver-7bbb656c7d-6mj2d\" (UID: \"bd64c275-cf75-4e42-8643-9178665801cb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6mj2d" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.288984 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-7tnbd"] Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.289003 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/66592797-8f8e-4e3c-938b-aa6767928e3f-serving-cert\") pod \"openshift-config-operator-7777fb866f-mj6ns\" (UID: \"66592797-8f8e-4e3c-938b-aa6767928e3f\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-mj6ns" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.289024 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5864769c-0a13-4562-a407-f67e4348e1a6-config\") pod \"machine-api-operator-5694c8668f-g64xp\" (UID: \"5864769c-0a13-4562-a407-f67e4348e1a6\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-g64xp" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.289046 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f4d41d59-ebec-455b-9a17-2ff3f9a42f04-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-kjc2g\" (UID: \"f4d41d59-ebec-455b-9a17-2ff3f9a42f04\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-kjc2g" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.289071 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9r78f\" (UniqueName: \"kubernetes.io/projected/6b062c17-08ee-4fda-94d2-a59a3cdf8641-kube-api-access-9r78f\") pod \"dns-operator-744455d44c-fj2tb\" (UID: \"6b062c17-08ee-4fda-94d2-a59a3cdf8641\") " pod="openshift-dns-operator/dns-operator-744455d44c-fj2tb" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.289095 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/99582129-e50d-49c0-a554-aa703fc86452-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-z2k6h\" (UID: \"99582129-e50d-49c0-a554-aa703fc86452\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-z2k6h" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.289127 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/4d81d21a-83f5-402e-91fd-d8268a6b49be-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-677j9\" (UID: \"4d81d21a-83f5-402e-91fd-d8268a6b49be\") " pod="openshift-controller-manager/controller-manager-879f6c89f-677j9" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.289161 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/5864769c-0a13-4562-a407-f67e4348e1a6-images\") pod \"machine-api-operator-5694c8668f-g64xp\" (UID: \"5864769c-0a13-4562-a407-f67e4348e1a6\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-g64xp" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.289184 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/528dc2c4-71cc-46cb-b847-426bb8638188-service-ca\") pod \"console-f9d7485db-5p6w2\" (UID: \"528dc2c4-71cc-46cb-b847-426bb8638188\") " pod="openshift-console/console-f9d7485db-5p6w2" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.289201 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n2j44\" (UniqueName: \"kubernetes.io/projected/bd64c275-cf75-4e42-8643-9178665801cb-kube-api-access-n2j44\") pod \"apiserver-7bbb656c7d-6mj2d\" (UID: \"bd64c275-cf75-4e42-8643-9178665801cb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6mj2d" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.289234 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/2a28f341-5f93-4d6c-bc4b-973d33dfeee1-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-k9sz7\" (UID: \"2a28f341-5f93-4d6c-bc4b-973d33dfeee1\") " pod="openshift-authentication/oauth-openshift-558db77b4-k9sz7" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.289254 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2a28f341-5f93-4d6c-bc4b-973d33dfeee1-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-k9sz7\" (UID: \"2a28f341-5f93-4d6c-bc4b-973d33dfeee1\") " pod="openshift-authentication/oauth-openshift-558db77b4-k9sz7" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.289275 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/528dc2c4-71cc-46cb-b847-426bb8638188-console-serving-cert\") pod \"console-f9d7485db-5p6w2\" (UID: \"528dc2c4-71cc-46cb-b847-426bb8638188\") " pod="openshift-console/console-f9d7485db-5p6w2" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.289297 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bd64c275-cf75-4e42-8643-9178665801cb-serving-cert\") pod \"apiserver-7bbb656c7d-6mj2d\" (UID: \"bd64c275-cf75-4e42-8643-9178665801cb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6mj2d" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.289342 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/bd64c275-cf75-4e42-8643-9178665801cb-etcd-client\") pod \"apiserver-7bbb656c7d-6mj2d\" (UID: \"bd64c275-cf75-4e42-8643-9178665801cb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6mj2d" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.289368 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/587cde2e-d0a4-406b-ab72-26c7fe4b6707-config\") pod \"route-controller-manager-6576b87f9c-m2525\" (UID: \"587cde2e-d0a4-406b-ab72-26c7fe4b6707\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m2525" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.289391 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/99582129-e50d-49c0-a554-aa703fc86452-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-z2k6h\" (UID: \"99582129-e50d-49c0-a554-aa703fc86452\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-z2k6h" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.289478 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aa57d7bf-eaea-4595-9d00-5d5d4b20f059-config\") pod \"etcd-operator-b45778765-sz8hj\" (UID: \"aa57d7bf-eaea-4595-9d00-5d5d4b20f059\") " pod="openshift-etcd-operator/etcd-operator-b45778765-sz8hj" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.289517 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/528dc2c4-71cc-46cb-b847-426bb8638188-oauth-serving-cert\") pod \"console-f9d7485db-5p6w2\" (UID: \"528dc2c4-71cc-46cb-b847-426bb8638188\") " pod="openshift-console/console-f9d7485db-5p6w2" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.289574 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9zv2s\" (UniqueName: \"kubernetes.io/projected/5864769c-0a13-4562-a407-f67e4348e1a6-kube-api-access-9zv2s\") pod \"machine-api-operator-5694c8668f-g64xp\" (UID: \"5864769c-0a13-4562-a407-f67e4348e1a6\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-g64xp" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.289596 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/10609d5f-d6dd-40a5-acc4-63e62584d18e-encryption-config\") pod \"apiserver-76f77b778f-8d467\" (UID: \"10609d5f-d6dd-40a5-acc4-63e62584d18e\") " pod="openshift-apiserver/apiserver-76f77b778f-8d467" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.289620 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/2a28f341-5f93-4d6c-bc4b-973d33dfeee1-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-k9sz7\" (UID: \"2a28f341-5f93-4d6c-bc4b-973d33dfeee1\") " pod="openshift-authentication/oauth-openshift-558db77b4-k9sz7" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.289644 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f4d41d59-ebec-455b-9a17-2ff3f9a42f04-serving-cert\") pod \"authentication-operator-69f744f599-kjc2g\" (UID: \"f4d41d59-ebec-455b-9a17-2ff3f9a42f04\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-kjc2g" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.289809 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-7tnbd" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.291077 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.291515 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.293362 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.293826 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.294453 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-95dbg"] Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.295361 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.295535 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/bd64c275-cf75-4e42-8643-9178665801cb-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-6mj2d\" (UID: \"bd64c275-cf75-4e42-8643-9178665801cb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6mj2d" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.295625 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4d81d21a-83f5-402e-91fd-d8268a6b49be-client-ca\") pod \"controller-manager-879f6c89f-677j9\" (UID: \"4d81d21a-83f5-402e-91fd-d8268a6b49be\") " pod="openshift-controller-manager/controller-manager-879f6c89f-677j9" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.295653 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/5864769c-0a13-4562-a407-f67e4348e1a6-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-g64xp\" (UID: \"5864769c-0a13-4562-a407-f67e4348e1a6\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-g64xp" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.295688 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/2a28f341-5f93-4d6c-bc4b-973d33dfeee1-audit-dir\") pod \"oauth-openshift-558db77b4-k9sz7\" (UID: \"2a28f341-5f93-4d6c-bc4b-973d33dfeee1\") " pod="openshift-authentication/oauth-openshift-558db77b4-k9sz7" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.295719 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6mzfw\" (UniqueName: \"kubernetes.io/projected/2a28f341-5f93-4d6c-bc4b-973d33dfeee1-kube-api-access-6mzfw\") pod \"oauth-openshift-558db77b4-k9sz7\" (UID: \"2a28f341-5f93-4d6c-bc4b-973d33dfeee1\") " pod="openshift-authentication/oauth-openshift-558db77b4-k9sz7" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.295740 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9l6hw\" (UniqueName: \"kubernetes.io/projected/f4d41d59-ebec-455b-9a17-2ff3f9a42f04-kube-api-access-9l6hw\") pod \"authentication-operator-69f744f599-kjc2g\" (UID: \"f4d41d59-ebec-455b-9a17-2ff3f9a42f04\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-kjc2g" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.295773 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/c31a7288-dcad-4e03-beb6-22e7855a58cd-proxy-tls\") pod \"machine-config-operator-74547568cd-4hkb9\" (UID: \"c31a7288-dcad-4e03-beb6-22e7855a58cd\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-4hkb9" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.295793 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/10609d5f-d6dd-40a5-acc4-63e62584d18e-serving-cert\") pod \"apiserver-76f77b778f-8d467\" (UID: \"10609d5f-d6dd-40a5-acc4-63e62584d18e\") " pod="openshift-apiserver/apiserver-76f77b778f-8d467" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.295815 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/2a28f341-5f93-4d6c-bc4b-973d33dfeee1-audit-policies\") pod \"oauth-openshift-558db77b4-k9sz7\" (UID: \"2a28f341-5f93-4d6c-bc4b-973d33dfeee1\") " pod="openshift-authentication/oauth-openshift-558db77b4-k9sz7" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.295835 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/587cde2e-d0a4-406b-ab72-26c7fe4b6707-serving-cert\") pod \"route-controller-manager-6576b87f9c-m2525\" (UID: \"587cde2e-d0a4-406b-ab72-26c7fe4b6707\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m2525" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.302941 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.304457 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29490765-w8zqq"] Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.304754 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-95dbg" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.304779 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-k5sf5"] Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.304891 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29490765-w8zqq" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.305687 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-m2ktn"] Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.306087 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-2kmrb"] Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.306219 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-k5sf5" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.306305 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-m2ktn" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.307242 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8rjs9"] Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.307757 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-8rhs2"] Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.308205 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-vfldk"] Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.308659 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-472pk"] Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.308733 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-2kmrb" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.308696 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-vfldk" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.309376 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8rjs9" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.309587 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8rhs2" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.310324 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-jxzst"] Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.310891 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-jxzst" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.310930 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-472pk" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.311380 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-7xsg2"] Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.312033 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-7xsg2" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.312382 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-fbbxr"] Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.312867 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-fbbxr" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.317351 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-4px9m"] Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.318261 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-4px9m" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.321485 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.322005 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-w7zlr"] Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.322685 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-w7zlr" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.326975 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-964q6"] Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.327927 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-964q6" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.328239 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-lxx9x"] Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.328591 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-lxx9x" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.330769 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-mj6ns"] Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.332127 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-m2525"] Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.332468 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-rcrn5"] Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.333915 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-4hkb9"] Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.335024 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-t4r22"] Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.336180 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-ptrdc"] Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.338127 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-kjc2g"] Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.340067 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-8d467"] Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.343788 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8rjs9"] Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.345569 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.348819 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-g64xp"] Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.353238 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-677j9"] Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.360193 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zvqdh"] Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.362368 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.364484 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7lgch"] Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.370932 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-k9sz7"] Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.374911 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-5p6w2"] Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.376579 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-g4dbx"] Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.378855 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-k5sf5"] Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.379280 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.379749 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-7tnbd"] Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.380869 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-z2k6h"] Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.383167 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-8rhs2"] Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.384270 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-47pnn"] Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.385335 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-8647g"] Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.386373 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-2kmrb"] Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.387482 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29490765-w8zqq"] Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.389174 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-fj2tb"] Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.390406 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-wp5r6"] Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.391410 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-wp5r6" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.391441 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-9shrr"] Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.392806 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-wjdgm"] Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.392889 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-9shrr" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.394285 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-sz8hj"] Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.395632 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-6mj2d"] Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.396940 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-472pk"] Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.397396 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/aa57d7bf-eaea-4595-9d00-5d5d4b20f059-etcd-ca\") pod \"etcd-operator-b45778765-sz8hj\" (UID: \"aa57d7bf-eaea-4595-9d00-5d5d4b20f059\") " pod="openshift-etcd-operator/etcd-operator-b45778765-sz8hj" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.397574 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/587cde2e-d0a4-406b-ab72-26c7fe4b6707-client-ca\") pod \"route-controller-manager-6576b87f9c-m2525\" (UID: \"587cde2e-d0a4-406b-ab72-26c7fe4b6707\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m2525" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.397682 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/10609d5f-d6dd-40a5-acc4-63e62584d18e-audit\") pod \"apiserver-76f77b778f-8d467\" (UID: \"10609d5f-d6dd-40a5-acc4-63e62584d18e\") " pod="openshift-apiserver/apiserver-76f77b778f-8d467" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.397761 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/10609d5f-d6dd-40a5-acc4-63e62584d18e-audit-dir\") pod \"apiserver-76f77b778f-8d467\" (UID: \"10609d5f-d6dd-40a5-acc4-63e62584d18e\") " pod="openshift-apiserver/apiserver-76f77b778f-8d467" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.397844 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rqfpq\" (UniqueName: \"kubernetes.io/projected/10609d5f-d6dd-40a5-acc4-63e62584d18e-kube-api-access-rqfpq\") pod \"apiserver-76f77b778f-8d467\" (UID: \"10609d5f-d6dd-40a5-acc4-63e62584d18e\") " pod="openshift-apiserver/apiserver-76f77b778f-8d467" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.397917 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/bd64c275-cf75-4e42-8643-9178665801cb-audit-policies\") pod \"apiserver-7bbb656c7d-6mj2d\" (UID: \"bd64c275-cf75-4e42-8643-9178665801cb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6mj2d" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.397999 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/c31a7288-dcad-4e03-beb6-22e7855a58cd-auth-proxy-config\") pod \"machine-config-operator-74547568cd-4hkb9\" (UID: \"c31a7288-dcad-4e03-beb6-22e7855a58cd\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-4hkb9" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.398074 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/10609d5f-d6dd-40a5-acc4-63e62584d18e-config\") pod \"apiserver-76f77b778f-8d467\" (UID: \"10609d5f-d6dd-40a5-acc4-63e62584d18e\") " pod="openshift-apiserver/apiserver-76f77b778f-8d467" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.398151 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/10609d5f-d6dd-40a5-acc4-63e62584d18e-etcd-client\") pod \"apiserver-76f77b778f-8d467\" (UID: \"10609d5f-d6dd-40a5-acc4-63e62584d18e\") " pod="openshift-apiserver/apiserver-76f77b778f-8d467" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.398228 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/10609d5f-d6dd-40a5-acc4-63e62584d18e-audit-dir\") pod \"apiserver-76f77b778f-8d467\" (UID: \"10609d5f-d6dd-40a5-acc4-63e62584d18e\") " pod="openshift-apiserver/apiserver-76f77b778f-8d467" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.398244 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/2a28f341-5f93-4d6c-bc4b-973d33dfeee1-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-k9sz7\" (UID: \"2a28f341-5f93-4d6c-bc4b-973d33dfeee1\") " pod="openshift-authentication/oauth-openshift-558db77b4-k9sz7" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.398419 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f4d41d59-ebec-455b-9a17-2ff3f9a42f04-service-ca-bundle\") pod \"authentication-operator-69f744f599-kjc2g\" (UID: \"f4d41d59-ebec-455b-9a17-2ff3f9a42f04\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-kjc2g" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.398531 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/99582129-e50d-49c0-a554-aa703fc86452-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-z2k6h\" (UID: \"99582129-e50d-49c0-a554-aa703fc86452\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-z2k6h" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.398640 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/06082490-21d0-4f81-8f02-b5ad80760b06-config\") pod \"machine-approver-56656f9798-kh4hh\" (UID: \"06082490-21d0-4f81-8f02-b5ad80760b06\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-kh4hh" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.398787 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l58br\" (UniqueName: \"kubernetes.io/projected/99582129-e50d-49c0-a554-aa703fc86452-kube-api-access-l58br\") pod \"cluster-image-registry-operator-dc59b4c8b-z2k6h\" (UID: \"99582129-e50d-49c0-a554-aa703fc86452\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-z2k6h" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.398918 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5c9zf\" (UniqueName: \"kubernetes.io/projected/aa57d7bf-eaea-4595-9d00-5d5d4b20f059-kube-api-access-5c9zf\") pod \"etcd-operator-b45778765-sz8hj\" (UID: \"aa57d7bf-eaea-4595-9d00-5d5d4b20f059\") " pod="openshift-etcd-operator/etcd-operator-b45778765-sz8hj" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.399042 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/bd64c275-cf75-4e42-8643-9178665801cb-audit-dir\") pod \"apiserver-7bbb656c7d-6mj2d\" (UID: \"bd64c275-cf75-4e42-8643-9178665801cb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6mj2d" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.399148 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/adeda50d-627e-4071-a6ab-d7712750c22d-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-zvqdh\" (UID: \"adeda50d-627e-4071-a6ab-d7712750c22d\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zvqdh" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.399268 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/528dc2c4-71cc-46cb-b847-426bb8638188-trusted-ca-bundle\") pod \"console-f9d7485db-5p6w2\" (UID: \"528dc2c4-71cc-46cb-b847-426bb8638188\") " pod="openshift-console/console-f9d7485db-5p6w2" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.399357 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/06082490-21d0-4f81-8f02-b5ad80760b06-auth-proxy-config\") pod \"machine-approver-56656f9798-kh4hh\" (UID: \"06082490-21d0-4f81-8f02-b5ad80760b06\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-kh4hh" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.399432 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/06082490-21d0-4f81-8f02-b5ad80760b06-machine-approver-tls\") pod \"machine-approver-56656f9798-kh4hh\" (UID: \"06082490-21d0-4f81-8f02-b5ad80760b06\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-kh4hh" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.399520 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vdvx6\" (UniqueName: \"kubernetes.io/projected/4d81d21a-83f5-402e-91fd-d8268a6b49be-kube-api-access-vdvx6\") pod \"controller-manager-879f6c89f-677j9\" (UID: \"4d81d21a-83f5-402e-91fd-d8268a6b49be\") " pod="openshift-controller-manager/controller-manager-879f6c89f-677j9" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.399645 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/2a28f341-5f93-4d6c-bc4b-973d33dfeee1-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-k9sz7\" (UID: \"2a28f341-5f93-4d6c-bc4b-973d33dfeee1\") " pod="openshift-authentication/oauth-openshift-558db77b4-k9sz7" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.399797 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-llv5g\" (UniqueName: \"kubernetes.io/projected/adeda50d-627e-4071-a6ab-d7712750c22d-kube-api-access-llv5g\") pod \"openshift-apiserver-operator-796bbdcf4f-zvqdh\" (UID: \"adeda50d-627e-4071-a6ab-d7712750c22d\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zvqdh" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.399882 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/bd64c275-cf75-4e42-8643-9178665801cb-audit-dir\") pod \"apiserver-7bbb656c7d-6mj2d\" (UID: \"bd64c275-cf75-4e42-8643-9178665801cb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6mj2d" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.399893 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/6b062c17-08ee-4fda-94d2-a59a3cdf8641-metrics-tls\") pod \"dns-operator-744455d44c-fj2tb\" (UID: \"6b062c17-08ee-4fda-94d2-a59a3cdf8641\") " pod="openshift-dns-operator/dns-operator-744455d44c-fj2tb" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.400541 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/2a28f341-5f93-4d6c-bc4b-973d33dfeee1-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-k9sz7\" (UID: \"2a28f341-5f93-4d6c-bc4b-973d33dfeee1\") " pod="openshift-authentication/oauth-openshift-558db77b4-k9sz7" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.400570 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/2a28f341-5f93-4d6c-bc4b-973d33dfeee1-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-k9sz7\" (UID: \"2a28f341-5f93-4d6c-bc4b-973d33dfeee1\") " pod="openshift-authentication/oauth-openshift-558db77b4-k9sz7" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.400592 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/10609d5f-d6dd-40a5-acc4-63e62584d18e-etcd-serving-ca\") pod \"apiserver-76f77b778f-8d467\" (UID: \"10609d5f-d6dd-40a5-acc4-63e62584d18e\") " pod="openshift-apiserver/apiserver-76f77b778f-8d467" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.400636 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/528dc2c4-71cc-46cb-b847-426bb8638188-console-config\") pod \"console-f9d7485db-5p6w2\" (UID: \"528dc2c4-71cc-46cb-b847-426bb8638188\") " pod="openshift-console/console-f9d7485db-5p6w2" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.400642 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/06082490-21d0-4f81-8f02-b5ad80760b06-config\") pod \"machine-approver-56656f9798-kh4hh\" (UID: \"06082490-21d0-4f81-8f02-b5ad80760b06\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-kh4hh" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.400657 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8jdh4\" (UniqueName: \"kubernetes.io/projected/c31a7288-dcad-4e03-beb6-22e7855a58cd-kube-api-access-8jdh4\") pod \"machine-config-operator-74547568cd-4hkb9\" (UID: \"c31a7288-dcad-4e03-beb6-22e7855a58cd\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-4hkb9" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.400703 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/aa57d7bf-eaea-4595-9d00-5d5d4b20f059-serving-cert\") pod \"etcd-operator-b45778765-sz8hj\" (UID: \"aa57d7bf-eaea-4595-9d00-5d5d4b20f059\") " pod="openshift-etcd-operator/etcd-operator-b45778765-sz8hj" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.400724 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/aa57d7bf-eaea-4595-9d00-5d5d4b20f059-etcd-service-ca\") pod \"etcd-operator-b45778765-sz8hj\" (UID: \"aa57d7bf-eaea-4595-9d00-5d5d4b20f059\") " pod="openshift-etcd-operator/etcd-operator-b45778765-sz8hj" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.400745 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/bd64c275-cf75-4e42-8643-9178665801cb-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-6mj2d\" (UID: \"bd64c275-cf75-4e42-8643-9178665801cb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6mj2d" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.400788 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/528dc2c4-71cc-46cb-b847-426bb8638188-console-oauth-config\") pod \"console-f9d7485db-5p6w2\" (UID: \"528dc2c4-71cc-46cb-b847-426bb8638188\") " pod="openshift-console/console-f9d7485db-5p6w2" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.400820 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4d81d21a-83f5-402e-91fd-d8268a6b49be-serving-cert\") pod \"controller-manager-879f6c89f-677j9\" (UID: \"4d81d21a-83f5-402e-91fd-d8268a6b49be\") " pod="openshift-controller-manager/controller-manager-879f6c89f-677j9" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.400868 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f4d41d59-ebec-455b-9a17-2ff3f9a42f04-config\") pod \"authentication-operator-69f744f599-kjc2g\" (UID: \"f4d41d59-ebec-455b-9a17-2ff3f9a42f04\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-kjc2g" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.400888 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5jq47\" (UniqueName: \"kubernetes.io/projected/528dc2c4-71cc-46cb-b847-426bb8638188-kube-api-access-5jq47\") pod \"console-f9d7485db-5p6w2\" (UID: \"528dc2c4-71cc-46cb-b847-426bb8638188\") " pod="openshift-console/console-f9d7485db-5p6w2" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.400936 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8zx6c\" (UniqueName: \"kubernetes.io/projected/78242150-7a64-4766-808f-fa460cd84b20-kube-api-access-8zx6c\") pod \"downloads-7954f5f757-rcrn5\" (UID: \"78242150-7a64-4766-808f-fa460cd84b20\") " pod="openshift-console/downloads-7954f5f757-rcrn5" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.400961 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/10609d5f-d6dd-40a5-acc4-63e62584d18e-image-import-ca\") pod \"apiserver-76f77b778f-8d467\" (UID: \"10609d5f-d6dd-40a5-acc4-63e62584d18e\") " pod="openshift-apiserver/apiserver-76f77b778f-8d467" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.400992 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4d81d21a-83f5-402e-91fd-d8268a6b49be-config\") pod \"controller-manager-879f6c89f-677j9\" (UID: \"4d81d21a-83f5-402e-91fd-d8268a6b49be\") " pod="openshift-controller-manager/controller-manager-879f6c89f-677j9" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.401019 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qb27c\" (UniqueName: \"kubernetes.io/projected/66592797-8f8e-4e3c-938b-aa6767928e3f-kube-api-access-qb27c\") pod \"openshift-config-operator-7777fb866f-mj6ns\" (UID: \"66592797-8f8e-4e3c-938b-aa6767928e3f\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-mj6ns" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.401046 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/10609d5f-d6dd-40a5-acc4-63e62584d18e-node-pullsecrets\") pod \"apiserver-76f77b778f-8d467\" (UID: \"10609d5f-d6dd-40a5-acc4-63e62584d18e\") " pod="openshift-apiserver/apiserver-76f77b778f-8d467" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.398426 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/587cde2e-d0a4-406b-ab72-26c7fe4b6707-client-ca\") pod \"route-controller-manager-6576b87f9c-m2525\" (UID: \"587cde2e-d0a4-406b-ab72-26c7fe4b6707\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m2525" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.401071 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/adeda50d-627e-4071-a6ab-d7712750c22d-config\") pod \"openshift-apiserver-operator-796bbdcf4f-zvqdh\" (UID: \"adeda50d-627e-4071-a6ab-d7712750c22d\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zvqdh" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.401107 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t7bg2\" (UniqueName: \"kubernetes.io/projected/06082490-21d0-4f81-8f02-b5ad80760b06-kube-api-access-t7bg2\") pod \"machine-approver-56656f9798-kh4hh\" (UID: \"06082490-21d0-4f81-8f02-b5ad80760b06\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-kh4hh" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.401143 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/66592797-8f8e-4e3c-938b-aa6767928e3f-available-featuregates\") pod \"openshift-config-operator-7777fb866f-mj6ns\" (UID: \"66592797-8f8e-4e3c-938b-aa6767928e3f\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-mj6ns" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.401173 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8tpz9\" (UniqueName: \"kubernetes.io/projected/39a0ee2d-0f5b-4231-94a3-d11d517d3517-kube-api-access-8tpz9\") pod \"cluster-samples-operator-665b6dd947-ptrdc\" (UID: \"39a0ee2d-0f5b-4231-94a3-d11d517d3517\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-ptrdc" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.401192 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/c31a7288-dcad-4e03-beb6-22e7855a58cd-images\") pod \"machine-config-operator-74547568cd-4hkb9\" (UID: \"c31a7288-dcad-4e03-beb6-22e7855a58cd\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-4hkb9" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.401232 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/2a28f341-5f93-4d6c-bc4b-973d33dfeee1-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-k9sz7\" (UID: \"2a28f341-5f93-4d6c-bc4b-973d33dfeee1\") " pod="openshift-authentication/oauth-openshift-558db77b4-k9sz7" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.401251 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/aa57d7bf-eaea-4595-9d00-5d5d4b20f059-etcd-client\") pod \"etcd-operator-b45778765-sz8hj\" (UID: \"aa57d7bf-eaea-4595-9d00-5d5d4b20f059\") " pod="openshift-etcd-operator/etcd-operator-b45778765-sz8hj" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.401268 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/06082490-21d0-4f81-8f02-b5ad80760b06-auth-proxy-config\") pod \"machine-approver-56656f9798-kh4hh\" (UID: \"06082490-21d0-4f81-8f02-b5ad80760b06\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-kh4hh" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.401320 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5864769c-0a13-4562-a407-f67e4348e1a6-config\") pod \"machine-api-operator-5694c8668f-g64xp\" (UID: \"5864769c-0a13-4562-a407-f67e4348e1a6\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-g64xp" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.401347 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-w7zlr"] Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.399158 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/c31a7288-dcad-4e03-beb6-22e7855a58cd-auth-proxy-config\") pod \"machine-config-operator-74547568cd-4hkb9\" (UID: \"c31a7288-dcad-4e03-beb6-22e7855a58cd\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-4hkb9" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.401355 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f4d41d59-ebec-455b-9a17-2ff3f9a42f04-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-kjc2g\" (UID: \"f4d41d59-ebec-455b-9a17-2ff3f9a42f04\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-kjc2g" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.401525 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9r78f\" (UniqueName: \"kubernetes.io/projected/6b062c17-08ee-4fda-94d2-a59a3cdf8641-kube-api-access-9r78f\") pod \"dns-operator-744455d44c-fj2tb\" (UID: \"6b062c17-08ee-4fda-94d2-a59a3cdf8641\") " pod="openshift-dns-operator/dns-operator-744455d44c-fj2tb" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.401542 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-4px9m"] Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.401581 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/bd64c275-cf75-4e42-8643-9178665801cb-encryption-config\") pod \"apiserver-7bbb656c7d-6mj2d\" (UID: \"bd64c275-cf75-4e42-8643-9178665801cb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6mj2d" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.401639 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/66592797-8f8e-4e3c-938b-aa6767928e3f-serving-cert\") pod \"openshift-config-operator-7777fb866f-mj6ns\" (UID: \"66592797-8f8e-4e3c-938b-aa6767928e3f\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-mj6ns" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.401690 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/99582129-e50d-49c0-a554-aa703fc86452-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-z2k6h\" (UID: \"99582129-e50d-49c0-a554-aa703fc86452\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-z2k6h" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.401726 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/4d81d21a-83f5-402e-91fd-d8268a6b49be-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-677j9\" (UID: \"4d81d21a-83f5-402e-91fd-d8268a6b49be\") " pod="openshift-controller-manager/controller-manager-879f6c89f-677j9" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.401750 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/5864769c-0a13-4562-a407-f67e4348e1a6-images\") pod \"machine-api-operator-5694c8668f-g64xp\" (UID: \"5864769c-0a13-4562-a407-f67e4348e1a6\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-g64xp" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.401772 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/528dc2c4-71cc-46cb-b847-426bb8638188-service-ca\") pod \"console-f9d7485db-5p6w2\" (UID: \"528dc2c4-71cc-46cb-b847-426bb8638188\") " pod="openshift-console/console-f9d7485db-5p6w2" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.401795 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n2j44\" (UniqueName: \"kubernetes.io/projected/bd64c275-cf75-4e42-8643-9178665801cb-kube-api-access-n2j44\") pod \"apiserver-7bbb656c7d-6mj2d\" (UID: \"bd64c275-cf75-4e42-8643-9178665801cb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6mj2d" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.401821 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2a28f341-5f93-4d6c-bc4b-973d33dfeee1-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-k9sz7\" (UID: \"2a28f341-5f93-4d6c-bc4b-973d33dfeee1\") " pod="openshift-authentication/oauth-openshift-558db77b4-k9sz7" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.401850 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/528dc2c4-71cc-46cb-b847-426bb8638188-console-serving-cert\") pod \"console-f9d7485db-5p6w2\" (UID: \"528dc2c4-71cc-46cb-b847-426bb8638188\") " pod="openshift-console/console-f9d7485db-5p6w2" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.401873 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bd64c275-cf75-4e42-8643-9178665801cb-serving-cert\") pod \"apiserver-7bbb656c7d-6mj2d\" (UID: \"bd64c275-cf75-4e42-8643-9178665801cb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6mj2d" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.399379 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f4d41d59-ebec-455b-9a17-2ff3f9a42f04-service-ca-bundle\") pod \"authentication-operator-69f744f599-kjc2g\" (UID: \"f4d41d59-ebec-455b-9a17-2ff3f9a42f04\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-kjc2g" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.400963 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/2a28f341-5f93-4d6c-bc4b-973d33dfeee1-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-k9sz7\" (UID: \"2a28f341-5f93-4d6c-bc4b-973d33dfeee1\") " pod="openshift-authentication/oauth-openshift-558db77b4-k9sz7" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.402129 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/2a28f341-5f93-4d6c-bc4b-973d33dfeee1-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-k9sz7\" (UID: \"2a28f341-5f93-4d6c-bc4b-973d33dfeee1\") " pod="openshift-authentication/oauth-openshift-558db77b4-k9sz7" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.399392 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/10609d5f-d6dd-40a5-acc4-63e62584d18e-config\") pod \"apiserver-76f77b778f-8d467\" (UID: \"10609d5f-d6dd-40a5-acc4-63e62584d18e\") " pod="openshift-apiserver/apiserver-76f77b778f-8d467" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.398810 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/bd64c275-cf75-4e42-8643-9178665801cb-audit-policies\") pod \"apiserver-7bbb656c7d-6mj2d\" (UID: \"bd64c275-cf75-4e42-8643-9178665801cb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6mj2d" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.400115 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.399579 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/10609d5f-d6dd-40a5-acc4-63e62584d18e-audit\") pod \"apiserver-76f77b778f-8d467\" (UID: \"10609d5f-d6dd-40a5-acc4-63e62584d18e\") " pod="openshift-apiserver/apiserver-76f77b778f-8d467" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.402517 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/10609d5f-d6dd-40a5-acc4-63e62584d18e-node-pullsecrets\") pod \"apiserver-76f77b778f-8d467\" (UID: \"10609d5f-d6dd-40a5-acc4-63e62584d18e\") " pod="openshift-apiserver/apiserver-76f77b778f-8d467" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.402545 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/528dc2c4-71cc-46cb-b847-426bb8638188-trusted-ca-bundle\") pod \"console-f9d7485db-5p6w2\" (UID: \"528dc2c4-71cc-46cb-b847-426bb8638188\") " pod="openshift-console/console-f9d7485db-5p6w2" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.403056 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/66592797-8f8e-4e3c-938b-aa6767928e3f-available-featuregates\") pod \"openshift-config-operator-7777fb866f-mj6ns\" (UID: \"66592797-8f8e-4e3c-938b-aa6767928e3f\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-mj6ns" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.403551 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/528dc2c4-71cc-46cb-b847-426bb8638188-console-config\") pod \"console-f9d7485db-5p6w2\" (UID: \"528dc2c4-71cc-46cb-b847-426bb8638188\") " pod="openshift-console/console-f9d7485db-5p6w2" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.403700 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/5864769c-0a13-4562-a407-f67e4348e1a6-images\") pod \"machine-api-operator-5694c8668f-g64xp\" (UID: \"5864769c-0a13-4562-a407-f67e4348e1a6\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-g64xp" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.403890 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/10609d5f-d6dd-40a5-acc4-63e62584d18e-etcd-serving-ca\") pod \"apiserver-76f77b778f-8d467\" (UID: \"10609d5f-d6dd-40a5-acc4-63e62584d18e\") " pod="openshift-apiserver/apiserver-76f77b778f-8d467" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.404802 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5864769c-0a13-4562-a407-f67e4348e1a6-config\") pod \"machine-api-operator-5694c8668f-g64xp\" (UID: \"5864769c-0a13-4562-a407-f67e4348e1a6\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-g64xp" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.404909 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/c31a7288-dcad-4e03-beb6-22e7855a58cd-images\") pod \"machine-config-operator-74547568cd-4hkb9\" (UID: \"c31a7288-dcad-4e03-beb6-22e7855a58cd\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-4hkb9" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.405266 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f4d41d59-ebec-455b-9a17-2ff3f9a42f04-config\") pod \"authentication-operator-69f744f599-kjc2g\" (UID: \"f4d41d59-ebec-455b-9a17-2ff3f9a42f04\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-kjc2g" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.405541 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/10609d5f-d6dd-40a5-acc4-63e62584d18e-image-import-ca\") pod \"apiserver-76f77b778f-8d467\" (UID: \"10609d5f-d6dd-40a5-acc4-63e62584d18e\") " pod="openshift-apiserver/apiserver-76f77b778f-8d467" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.405589 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/99582129-e50d-49c0-a554-aa703fc86452-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-z2k6h\" (UID: \"99582129-e50d-49c0-a554-aa703fc86452\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-z2k6h" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.405885 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/528dc2c4-71cc-46cb-b847-426bb8638188-service-ca\") pod \"console-f9d7485db-5p6w2\" (UID: \"528dc2c4-71cc-46cb-b847-426bb8638188\") " pod="openshift-console/console-f9d7485db-5p6w2" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.406037 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/587cde2e-d0a4-406b-ab72-26c7fe4b6707-config\") pod \"route-controller-manager-6576b87f9c-m2525\" (UID: \"587cde2e-d0a4-406b-ab72-26c7fe4b6707\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m2525" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.406035 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2a28f341-5f93-4d6c-bc4b-973d33dfeee1-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-k9sz7\" (UID: \"2a28f341-5f93-4d6c-bc4b-973d33dfeee1\") " pod="openshift-authentication/oauth-openshift-558db77b4-k9sz7" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.406562 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/bd64c275-cf75-4e42-8643-9178665801cb-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-6mj2d\" (UID: \"bd64c275-cf75-4e42-8643-9178665801cb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6mj2d" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.406579 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f4d41d59-ebec-455b-9a17-2ff3f9a42f04-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-kjc2g\" (UID: \"f4d41d59-ebec-455b-9a17-2ff3f9a42f04\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-kjc2g" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.406605 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/06082490-21d0-4f81-8f02-b5ad80760b06-machine-approver-tls\") pod \"machine-approver-56656f9798-kh4hh\" (UID: \"06082490-21d0-4f81-8f02-b5ad80760b06\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-kh4hh" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.407060 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/10609d5f-d6dd-40a5-acc4-63e62584d18e-etcd-client\") pod \"apiserver-76f77b778f-8d467\" (UID: \"10609d5f-d6dd-40a5-acc4-63e62584d18e\") " pod="openshift-apiserver/apiserver-76f77b778f-8d467" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.407113 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/2a28f341-5f93-4d6c-bc4b-973d33dfeee1-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-k9sz7\" (UID: \"2a28f341-5f93-4d6c-bc4b-973d33dfeee1\") " pod="openshift-authentication/oauth-openshift-558db77b4-k9sz7" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.407160 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-7xsg2"] Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.407200 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-m2ktn"] Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.407329 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/99582129-e50d-49c0-a554-aa703fc86452-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-z2k6h\" (UID: \"99582129-e50d-49c0-a554-aa703fc86452\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-z2k6h" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.407393 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/99582129-e50d-49c0-a554-aa703fc86452-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-z2k6h\" (UID: \"99582129-e50d-49c0-a554-aa703fc86452\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-z2k6h" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.407555 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/bd64c275-cf75-4e42-8643-9178665801cb-etcd-client\") pod \"apiserver-7bbb656c7d-6mj2d\" (UID: \"bd64c275-cf75-4e42-8643-9178665801cb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6mj2d" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.407635 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/528dc2c4-71cc-46cb-b847-426bb8638188-oauth-serving-cert\") pod \"console-f9d7485db-5p6w2\" (UID: \"528dc2c4-71cc-46cb-b847-426bb8638188\") " pod="openshift-console/console-f9d7485db-5p6w2" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.407797 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aa57d7bf-eaea-4595-9d00-5d5d4b20f059-config\") pod \"etcd-operator-b45778765-sz8hj\" (UID: \"aa57d7bf-eaea-4595-9d00-5d5d4b20f059\") " pod="openshift-etcd-operator/etcd-operator-b45778765-sz8hj" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.407849 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/2a28f341-5f93-4d6c-bc4b-973d33dfeee1-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-k9sz7\" (UID: \"2a28f341-5f93-4d6c-bc4b-973d33dfeee1\") " pod="openshift-authentication/oauth-openshift-558db77b4-k9sz7" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.407880 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f4d41d59-ebec-455b-9a17-2ff3f9a42f04-serving-cert\") pod \"authentication-operator-69f744f599-kjc2g\" (UID: \"f4d41d59-ebec-455b-9a17-2ff3f9a42f04\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-kjc2g" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.408020 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/bd64c275-cf75-4e42-8643-9178665801cb-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-6mj2d\" (UID: \"bd64c275-cf75-4e42-8643-9178665801cb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6mj2d" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.408401 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/528dc2c4-71cc-46cb-b847-426bb8638188-oauth-serving-cert\") pod \"console-f9d7485db-5p6w2\" (UID: \"528dc2c4-71cc-46cb-b847-426bb8638188\") " pod="openshift-console/console-f9d7485db-5p6w2" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.408741 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/587cde2e-d0a4-406b-ab72-26c7fe4b6707-config\") pod \"route-controller-manager-6576b87f9c-m2525\" (UID: \"587cde2e-d0a4-406b-ab72-26c7fe4b6707\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m2525" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.408869 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bd64c275-cf75-4e42-8643-9178665801cb-serving-cert\") pod \"apiserver-7bbb656c7d-6mj2d\" (UID: \"bd64c275-cf75-4e42-8643-9178665801cb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6mj2d" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.409390 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/2a28f341-5f93-4d6c-bc4b-973d33dfeee1-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-k9sz7\" (UID: \"2a28f341-5f93-4d6c-bc4b-973d33dfeee1\") " pod="openshift-authentication/oauth-openshift-558db77b4-k9sz7" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.409479 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/bd64c275-cf75-4e42-8643-9178665801cb-encryption-config\") pod \"apiserver-7bbb656c7d-6mj2d\" (UID: \"bd64c275-cf75-4e42-8643-9178665801cb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6mj2d" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.409480 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/528dc2c4-71cc-46cb-b847-426bb8638188-console-serving-cert\") pod \"console-f9d7485db-5p6w2\" (UID: \"528dc2c4-71cc-46cb-b847-426bb8638188\") " pod="openshift-console/console-f9d7485db-5p6w2" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.410149 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/4d81d21a-83f5-402e-91fd-d8268a6b49be-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-677j9\" (UID: \"4d81d21a-83f5-402e-91fd-d8268a6b49be\") " pod="openshift-controller-manager/controller-manager-879f6c89f-677j9" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.410227 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-95dbg"] Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.410295 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9zv2s\" (UniqueName: \"kubernetes.io/projected/5864769c-0a13-4562-a407-f67e4348e1a6-kube-api-access-9zv2s\") pod \"machine-api-operator-5694c8668f-g64xp\" (UID: \"5864769c-0a13-4562-a407-f67e4348e1a6\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-g64xp" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.410335 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/10609d5f-d6dd-40a5-acc4-63e62584d18e-encryption-config\") pod \"apiserver-76f77b778f-8d467\" (UID: \"10609d5f-d6dd-40a5-acc4-63e62584d18e\") " pod="openshift-apiserver/apiserver-76f77b778f-8d467" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.410395 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/2a28f341-5f93-4d6c-bc4b-973d33dfeee1-audit-dir\") pod \"oauth-openshift-558db77b4-k9sz7\" (UID: \"2a28f341-5f93-4d6c-bc4b-973d33dfeee1\") " pod="openshift-authentication/oauth-openshift-558db77b4-k9sz7" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.410836 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/2a28f341-5f93-4d6c-bc4b-973d33dfeee1-audit-dir\") pod \"oauth-openshift-558db77b4-k9sz7\" (UID: \"2a28f341-5f93-4d6c-bc4b-973d33dfeee1\") " pod="openshift-authentication/oauth-openshift-558db77b4-k9sz7" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.410904 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4d81d21a-83f5-402e-91fd-d8268a6b49be-client-ca\") pod \"controller-manager-879f6c89f-677j9\" (UID: \"4d81d21a-83f5-402e-91fd-d8268a6b49be\") " pod="openshift-controller-manager/controller-manager-879f6c89f-677j9" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.410940 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/5864769c-0a13-4562-a407-f67e4348e1a6-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-g64xp\" (UID: \"5864769c-0a13-4562-a407-f67e4348e1a6\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-g64xp" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.410955 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/bd64c275-cf75-4e42-8643-9178665801cb-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-6mj2d\" (UID: \"bd64c275-cf75-4e42-8643-9178665801cb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6mj2d" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.410965 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/adeda50d-627e-4071-a6ab-d7712750c22d-config\") pod \"openshift-apiserver-operator-796bbdcf4f-zvqdh\" (UID: \"adeda50d-627e-4071-a6ab-d7712750c22d\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zvqdh" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.410976 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6mzfw\" (UniqueName: \"kubernetes.io/projected/2a28f341-5f93-4d6c-bc4b-973d33dfeee1-kube-api-access-6mzfw\") pod \"oauth-openshift-558db77b4-k9sz7\" (UID: \"2a28f341-5f93-4d6c-bc4b-973d33dfeee1\") " pod="openshift-authentication/oauth-openshift-558db77b4-k9sz7" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.411006 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/c31a7288-dcad-4e03-beb6-22e7855a58cd-proxy-tls\") pod \"machine-config-operator-74547568cd-4hkb9\" (UID: \"c31a7288-dcad-4e03-beb6-22e7855a58cd\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-4hkb9" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.411035 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/10609d5f-d6dd-40a5-acc4-63e62584d18e-serving-cert\") pod \"apiserver-76f77b778f-8d467\" (UID: \"10609d5f-d6dd-40a5-acc4-63e62584d18e\") " pod="openshift-apiserver/apiserver-76f77b778f-8d467" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.411060 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/2a28f341-5f93-4d6c-bc4b-973d33dfeee1-audit-policies\") pod \"oauth-openshift-558db77b4-k9sz7\" (UID: \"2a28f341-5f93-4d6c-bc4b-973d33dfeee1\") " pod="openshift-authentication/oauth-openshift-558db77b4-k9sz7" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.411088 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9l6hw\" (UniqueName: \"kubernetes.io/projected/f4d41d59-ebec-455b-9a17-2ff3f9a42f04-kube-api-access-9l6hw\") pod \"authentication-operator-69f744f599-kjc2g\" (UID: \"f4d41d59-ebec-455b-9a17-2ff3f9a42f04\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-kjc2g" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.411125 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/587cde2e-d0a4-406b-ab72-26c7fe4b6707-serving-cert\") pod \"route-controller-manager-6576b87f9c-m2525\" (UID: \"587cde2e-d0a4-406b-ab72-26c7fe4b6707\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m2525" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.411154 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/2a28f341-5f93-4d6c-bc4b-973d33dfeee1-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-k9sz7\" (UID: \"2a28f341-5f93-4d6c-bc4b-973d33dfeee1\") " pod="openshift-authentication/oauth-openshift-558db77b4-k9sz7" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.411195 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/2a28f341-5f93-4d6c-bc4b-973d33dfeee1-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-k9sz7\" (UID: \"2a28f341-5f93-4d6c-bc4b-973d33dfeee1\") " pod="openshift-authentication/oauth-openshift-558db77b4-k9sz7" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.411227 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/39a0ee2d-0f5b-4231-94a3-d11d517d3517-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-ptrdc\" (UID: \"39a0ee2d-0f5b-4231-94a3-d11d517d3517\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-ptrdc" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.411260 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/2a28f341-5f93-4d6c-bc4b-973d33dfeee1-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-k9sz7\" (UID: \"2a28f341-5f93-4d6c-bc4b-973d33dfeee1\") " pod="openshift-authentication/oauth-openshift-558db77b4-k9sz7" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.411295 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/10609d5f-d6dd-40a5-acc4-63e62584d18e-trusted-ca-bundle\") pod \"apiserver-76f77b778f-8d467\" (UID: \"10609d5f-d6dd-40a5-acc4-63e62584d18e\") " pod="openshift-apiserver/apiserver-76f77b778f-8d467" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.411328 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-llgrp\" (UniqueName: \"kubernetes.io/projected/587cde2e-d0a4-406b-ab72-26c7fe4b6707-kube-api-access-llgrp\") pod \"route-controller-manager-6576b87f9c-m2525\" (UID: \"587cde2e-d0a4-406b-ab72-26c7fe4b6707\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m2525" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.411336 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4d81d21a-83f5-402e-91fd-d8268a6b49be-serving-cert\") pod \"controller-manager-879f6c89f-677j9\" (UID: \"4d81d21a-83f5-402e-91fd-d8268a6b49be\") " pod="openshift-controller-manager/controller-manager-879f6c89f-677j9" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.411380 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/adeda50d-627e-4071-a6ab-d7712750c22d-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-zvqdh\" (UID: \"adeda50d-627e-4071-a6ab-d7712750c22d\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zvqdh" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.411377 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/2a28f341-5f93-4d6c-bc4b-973d33dfeee1-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-k9sz7\" (UID: \"2a28f341-5f93-4d6c-bc4b-973d33dfeee1\") " pod="openshift-authentication/oauth-openshift-558db77b4-k9sz7" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.411776 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/528dc2c4-71cc-46cb-b847-426bb8638188-console-oauth-config\") pod \"console-f9d7485db-5p6w2\" (UID: \"528dc2c4-71cc-46cb-b847-426bb8638188\") " pod="openshift-console/console-f9d7485db-5p6w2" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.411779 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4d81d21a-83f5-402e-91fd-d8268a6b49be-client-ca\") pod \"controller-manager-879f6c89f-677j9\" (UID: \"4d81d21a-83f5-402e-91fd-d8268a6b49be\") " pod="openshift-controller-manager/controller-manager-879f6c89f-677j9" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.412833 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/2a28f341-5f93-4d6c-bc4b-973d33dfeee1-audit-policies\") pod \"oauth-openshift-558db77b4-k9sz7\" (UID: \"2a28f341-5f93-4d6c-bc4b-973d33dfeee1\") " pod="openshift-authentication/oauth-openshift-558db77b4-k9sz7" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.413300 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/10609d5f-d6dd-40a5-acc4-63e62584d18e-trusted-ca-bundle\") pod \"apiserver-76f77b778f-8d467\" (UID: \"10609d5f-d6dd-40a5-acc4-63e62584d18e\") " pod="openshift-apiserver/apiserver-76f77b778f-8d467" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.413449 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/10609d5f-d6dd-40a5-acc4-63e62584d18e-encryption-config\") pod \"apiserver-76f77b778f-8d467\" (UID: \"10609d5f-d6dd-40a5-acc4-63e62584d18e\") " pod="openshift-apiserver/apiserver-76f77b778f-8d467" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.413497 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-vfldk"] Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.413657 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4d81d21a-83f5-402e-91fd-d8268a6b49be-config\") pod \"controller-manager-879f6c89f-677j9\" (UID: \"4d81d21a-83f5-402e-91fd-d8268a6b49be\") " pod="openshift-controller-manager/controller-manager-879f6c89f-677j9" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.414797 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-9shrr"] Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.415129 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/2a28f341-5f93-4d6c-bc4b-973d33dfeee1-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-k9sz7\" (UID: \"2a28f341-5f93-4d6c-bc4b-973d33dfeee1\") " pod="openshift-authentication/oauth-openshift-558db77b4-k9sz7" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.415549 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/10609d5f-d6dd-40a5-acc4-63e62584d18e-serving-cert\") pod \"apiserver-76f77b778f-8d467\" (UID: \"10609d5f-d6dd-40a5-acc4-63e62584d18e\") " pod="openshift-apiserver/apiserver-76f77b778f-8d467" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.416384 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/c31a7288-dcad-4e03-beb6-22e7855a58cd-proxy-tls\") pod \"machine-config-operator-74547568cd-4hkb9\" (UID: \"c31a7288-dcad-4e03-beb6-22e7855a58cd\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-4hkb9" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.416490 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/39a0ee2d-0f5b-4231-94a3-d11d517d3517-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-ptrdc\" (UID: \"39a0ee2d-0f5b-4231-94a3-d11d517d3517\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-ptrdc" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.416566 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-964q6"] Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.416818 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/66592797-8f8e-4e3c-938b-aa6767928e3f-serving-cert\") pod \"openshift-config-operator-7777fb866f-mj6ns\" (UID: \"66592797-8f8e-4e3c-938b-aa6767928e3f\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-mj6ns" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.416813 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/2a28f341-5f93-4d6c-bc4b-973d33dfeee1-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-k9sz7\" (UID: \"2a28f341-5f93-4d6c-bc4b-973d33dfeee1\") " pod="openshift-authentication/oauth-openshift-558db77b4-k9sz7" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.416954 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/2a28f341-5f93-4d6c-bc4b-973d33dfeee1-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-k9sz7\" (UID: \"2a28f341-5f93-4d6c-bc4b-973d33dfeee1\") " pod="openshift-authentication/oauth-openshift-558db77b4-k9sz7" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.416997 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f4d41d59-ebec-455b-9a17-2ff3f9a42f04-serving-cert\") pod \"authentication-operator-69f744f599-kjc2g\" (UID: \"f4d41d59-ebec-455b-9a17-2ff3f9a42f04\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-kjc2g" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.417503 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/587cde2e-d0a4-406b-ab72-26c7fe4b6707-serving-cert\") pod \"route-controller-manager-6576b87f9c-m2525\" (UID: \"587cde2e-d0a4-406b-ab72-26c7fe4b6707\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m2525" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.417525 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/bd64c275-cf75-4e42-8643-9178665801cb-etcd-client\") pod \"apiserver-7bbb656c7d-6mj2d\" (UID: \"bd64c275-cf75-4e42-8643-9178665801cb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6mj2d" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.417599 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/2a28f341-5f93-4d6c-bc4b-973d33dfeee1-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-k9sz7\" (UID: \"2a28f341-5f93-4d6c-bc4b-973d33dfeee1\") " pod="openshift-authentication/oauth-openshift-558db77b4-k9sz7" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.418028 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/2a28f341-5f93-4d6c-bc4b-973d33dfeee1-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-k9sz7\" (UID: \"2a28f341-5f93-4d6c-bc4b-973d33dfeee1\") " pod="openshift-authentication/oauth-openshift-558db77b4-k9sz7" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.418559 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-lxx9x"] Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.419188 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/5864769c-0a13-4562-a407-f67e4348e1a6-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-g64xp\" (UID: \"5864769c-0a13-4562-a407-f67e4348e1a6\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-g64xp" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.419522 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/2a28f341-5f93-4d6c-bc4b-973d33dfeee1-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-k9sz7\" (UID: \"2a28f341-5f93-4d6c-bc4b-973d33dfeee1\") " pod="openshift-authentication/oauth-openshift-558db77b4-k9sz7" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.419698 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-fbbxr"] Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.420131 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.440049 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.459693 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.479322 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.509352 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.519516 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.522764 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/aa57d7bf-eaea-4595-9d00-5d5d4b20f059-etcd-service-ca\") pod \"etcd-operator-b45778765-sz8hj\" (UID: \"aa57d7bf-eaea-4595-9d00-5d5d4b20f059\") " pod="openshift-etcd-operator/etcd-operator-b45778765-sz8hj" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.539604 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.559351 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.580417 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.587111 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/aa57d7bf-eaea-4595-9d00-5d5d4b20f059-serving-cert\") pod \"etcd-operator-b45778765-sz8hj\" (UID: \"aa57d7bf-eaea-4595-9d00-5d5d4b20f059\") " pod="openshift-etcd-operator/etcd-operator-b45778765-sz8hj" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.611865 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.620011 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/aa57d7bf-eaea-4595-9d00-5d5d4b20f059-etcd-client\") pod \"etcd-operator-b45778765-sz8hj\" (UID: \"aa57d7bf-eaea-4595-9d00-5d5d4b20f059\") " pod="openshift-etcd-operator/etcd-operator-b45778765-sz8hj" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.631168 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.639556 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.659518 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.668659 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aa57d7bf-eaea-4595-9d00-5d5d4b20f059-config\") pod \"etcd-operator-b45778765-sz8hj\" (UID: \"aa57d7bf-eaea-4595-9d00-5d5d4b20f059\") " pod="openshift-etcd-operator/etcd-operator-b45778765-sz8hj" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.679652 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.688064 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/aa57d7bf-eaea-4595-9d00-5d5d4b20f059-etcd-ca\") pod \"etcd-operator-b45778765-sz8hj\" (UID: \"aa57d7bf-eaea-4595-9d00-5d5d4b20f059\") " pod="openshift-etcd-operator/etcd-operator-b45778765-sz8hj" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.699739 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.720614 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.734765 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/6b062c17-08ee-4fda-94d2-a59a3cdf8641-metrics-tls\") pod \"dns-operator-744455d44c-fj2tb\" (UID: \"6b062c17-08ee-4fda-94d2-a59a3cdf8641\") " pod="openshift-dns-operator/dns-operator-744455d44c-fj2tb" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.740542 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.779632 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.800746 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.820390 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.838930 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.860816 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.880724 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.899958 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.921501 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.941137 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Jan 26 16:49:27 crc kubenswrapper[4754]: I0126 16:49:27.960205 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Jan 26 16:49:28 crc kubenswrapper[4754]: I0126 16:49:28.000449 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Jan 26 16:49:28 crc kubenswrapper[4754]: I0126 16:49:28.019926 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Jan 26 16:49:28 crc kubenswrapper[4754]: I0126 16:49:28.039969 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Jan 26 16:49:28 crc kubenswrapper[4754]: I0126 16:49:28.060908 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Jan 26 16:49:28 crc kubenswrapper[4754]: I0126 16:49:28.079441 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Jan 26 16:49:28 crc kubenswrapper[4754]: I0126 16:49:28.100280 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Jan 26 16:49:28 crc kubenswrapper[4754]: I0126 16:49:28.119630 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Jan 26 16:49:28 crc kubenswrapper[4754]: I0126 16:49:28.140272 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Jan 26 16:49:28 crc kubenswrapper[4754]: I0126 16:49:28.160184 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Jan 26 16:49:28 crc kubenswrapper[4754]: I0126 16:49:28.180125 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Jan 26 16:49:28 crc kubenswrapper[4754]: I0126 16:49:28.200652 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Jan 26 16:49:28 crc kubenswrapper[4754]: I0126 16:49:28.221308 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Jan 26 16:49:28 crc kubenswrapper[4754]: I0126 16:49:28.239426 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Jan 26 16:49:28 crc kubenswrapper[4754]: I0126 16:49:28.260302 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Jan 26 16:49:28 crc kubenswrapper[4754]: I0126 16:49:28.279759 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Jan 26 16:49:28 crc kubenswrapper[4754]: I0126 16:49:28.300915 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Jan 26 16:49:28 crc kubenswrapper[4754]: I0126 16:49:28.318361 4754 request.go:700] Waited for 1.008551093s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-service-ca-operator/configmaps?fieldSelector=metadata.name%3Dopenshift-service-ca.crt&limit=500&resourceVersion=0 Jan 26 16:49:28 crc kubenswrapper[4754]: I0126 16:49:28.320892 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Jan 26 16:49:28 crc kubenswrapper[4754]: I0126 16:49:28.340086 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Jan 26 16:49:28 crc kubenswrapper[4754]: I0126 16:49:28.360633 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Jan 26 16:49:28 crc kubenswrapper[4754]: I0126 16:49:28.379932 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Jan 26 16:49:28 crc kubenswrapper[4754]: I0126 16:49:28.401505 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Jan 26 16:49:28 crc kubenswrapper[4754]: I0126 16:49:28.420601 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Jan 26 16:49:28 crc kubenswrapper[4754]: I0126 16:49:28.439977 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Jan 26 16:49:28 crc kubenswrapper[4754]: I0126 16:49:28.459527 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Jan 26 16:49:28 crc kubenswrapper[4754]: I0126 16:49:28.480861 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Jan 26 16:49:28 crc kubenswrapper[4754]: I0126 16:49:28.500638 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Jan 26 16:49:28 crc kubenswrapper[4754]: I0126 16:49:28.520589 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Jan 26 16:49:28 crc kubenswrapper[4754]: I0126 16:49:28.540441 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Jan 26 16:49:28 crc kubenswrapper[4754]: I0126 16:49:28.560725 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Jan 26 16:49:28 crc kubenswrapper[4754]: I0126 16:49:28.580858 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Jan 26 16:49:28 crc kubenswrapper[4754]: I0126 16:49:28.599780 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Jan 26 16:49:28 crc kubenswrapper[4754]: I0126 16:49:28.620554 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Jan 26 16:49:28 crc kubenswrapper[4754]: I0126 16:49:28.640431 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Jan 26 16:49:28 crc kubenswrapper[4754]: I0126 16:49:28.660531 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Jan 26 16:49:28 crc kubenswrapper[4754]: I0126 16:49:28.680365 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Jan 26 16:49:28 crc kubenswrapper[4754]: I0126 16:49:28.700401 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Jan 26 16:49:28 crc kubenswrapper[4754]: I0126 16:49:28.719580 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Jan 26 16:49:28 crc kubenswrapper[4754]: I0126 16:49:28.740246 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Jan 26 16:49:28 crc kubenswrapper[4754]: I0126 16:49:28.759923 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Jan 26 16:49:28 crc kubenswrapper[4754]: I0126 16:49:28.779817 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Jan 26 16:49:28 crc kubenswrapper[4754]: I0126 16:49:28.800145 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Jan 26 16:49:28 crc kubenswrapper[4754]: I0126 16:49:28.819848 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Jan 26 16:49:28 crc kubenswrapper[4754]: I0126 16:49:28.852483 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Jan 26 16:49:28 crc kubenswrapper[4754]: I0126 16:49:28.859639 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Jan 26 16:49:28 crc kubenswrapper[4754]: I0126 16:49:28.879553 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Jan 26 16:49:28 crc kubenswrapper[4754]: I0126 16:49:28.899861 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Jan 26 16:49:28 crc kubenswrapper[4754]: I0126 16:49:28.920693 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Jan 26 16:49:28 crc kubenswrapper[4754]: I0126 16:49:28.939918 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Jan 26 16:49:28 crc kubenswrapper[4754]: I0126 16:49:28.959351 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Jan 26 16:49:28 crc kubenswrapper[4754]: I0126 16:49:28.979869 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.001333 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.020893 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.041202 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.060961 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.080518 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.099166 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.119946 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.140561 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.159302 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.179925 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.199304 4754 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.236486 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rqfpq\" (UniqueName: \"kubernetes.io/projected/10609d5f-d6dd-40a5-acc4-63e62584d18e-kube-api-access-rqfpq\") pod \"apiserver-76f77b778f-8d467\" (UID: \"10609d5f-d6dd-40a5-acc4-63e62584d18e\") " pod="openshift-apiserver/apiserver-76f77b778f-8d467" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.254242 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l58br\" (UniqueName: \"kubernetes.io/projected/99582129-e50d-49c0-a554-aa703fc86452-kube-api-access-l58br\") pod \"cluster-image-registry-operator-dc59b4c8b-z2k6h\" (UID: \"99582129-e50d-49c0-a554-aa703fc86452\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-z2k6h" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.277079 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5c9zf\" (UniqueName: \"kubernetes.io/projected/aa57d7bf-eaea-4595-9d00-5d5d4b20f059-kube-api-access-5c9zf\") pod \"etcd-operator-b45778765-sz8hj\" (UID: \"aa57d7bf-eaea-4595-9d00-5d5d4b20f059\") " pod="openshift-etcd-operator/etcd-operator-b45778765-sz8hj" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.277430 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-sz8hj" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.295012 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vdvx6\" (UniqueName: \"kubernetes.io/projected/4d81d21a-83f5-402e-91fd-d8268a6b49be-kube-api-access-vdvx6\") pod \"controller-manager-879f6c89f-677j9\" (UID: \"4d81d21a-83f5-402e-91fd-d8268a6b49be\") " pod="openshift-controller-manager/controller-manager-879f6c89f-677j9" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.315502 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-llv5g\" (UniqueName: \"kubernetes.io/projected/adeda50d-627e-4071-a6ab-d7712750c22d-kube-api-access-llv5g\") pod \"openshift-apiserver-operator-796bbdcf4f-zvqdh\" (UID: \"adeda50d-627e-4071-a6ab-d7712750c22d\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zvqdh" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.316282 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zvqdh" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.336558 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8jdh4\" (UniqueName: \"kubernetes.io/projected/c31a7288-dcad-4e03-beb6-22e7855a58cd-kube-api-access-8jdh4\") pod \"machine-config-operator-74547568cd-4hkb9\" (UID: \"c31a7288-dcad-4e03-beb6-22e7855a58cd\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-4hkb9" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.338263 4754 request.go:700] Waited for 1.936763563s due to client-side throttling, not priority and fairness, request: POST:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-console/serviceaccounts/console/token Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.358016 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5jq47\" (UniqueName: \"kubernetes.io/projected/528dc2c4-71cc-46cb-b847-426bb8638188-kube-api-access-5jq47\") pod \"console-f9d7485db-5p6w2\" (UID: \"528dc2c4-71cc-46cb-b847-426bb8638188\") " pod="openshift-console/console-f9d7485db-5p6w2" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.379029 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8zx6c\" (UniqueName: \"kubernetes.io/projected/78242150-7a64-4766-808f-fa460cd84b20-kube-api-access-8zx6c\") pod \"downloads-7954f5f757-rcrn5\" (UID: \"78242150-7a64-4766-808f-fa460cd84b20\") " pod="openshift-console/downloads-7954f5f757-rcrn5" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.385534 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-677j9" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.396546 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t7bg2\" (UniqueName: \"kubernetes.io/projected/06082490-21d0-4f81-8f02-b5ad80760b06-kube-api-access-t7bg2\") pod \"machine-approver-56656f9798-kh4hh\" (UID: \"06082490-21d0-4f81-8f02-b5ad80760b06\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-kh4hh" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.415012 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qb27c\" (UniqueName: \"kubernetes.io/projected/66592797-8f8e-4e3c-938b-aa6767928e3f-kube-api-access-qb27c\") pod \"openshift-config-operator-7777fb866f-mj6ns\" (UID: \"66592797-8f8e-4e3c-938b-aa6767928e3f\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-mj6ns" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.434610 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8tpz9\" (UniqueName: \"kubernetes.io/projected/39a0ee2d-0f5b-4231-94a3-d11d517d3517-kube-api-access-8tpz9\") pod \"cluster-samples-operator-665b6dd947-ptrdc\" (UID: \"39a0ee2d-0f5b-4231-94a3-d11d517d3517\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-ptrdc" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.455285 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-8d467" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.455502 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n2j44\" (UniqueName: \"kubernetes.io/projected/bd64c275-cf75-4e42-8643-9178665801cb-kube-api-access-n2j44\") pod \"apiserver-7bbb656c7d-6mj2d\" (UID: \"bd64c275-cf75-4e42-8643-9178665801cb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6mj2d" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.472833 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-sz8hj"] Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.476311 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9r78f\" (UniqueName: \"kubernetes.io/projected/6b062c17-08ee-4fda-94d2-a59a3cdf8641-kube-api-access-9r78f\") pod \"dns-operator-744455d44c-fj2tb\" (UID: \"6b062c17-08ee-4fda-94d2-a59a3cdf8641\") " pod="openshift-dns-operator/dns-operator-744455d44c-fj2tb" Jan 26 16:49:29 crc kubenswrapper[4754]: W0126 16:49:29.484504 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaa57d7bf_eaea_4595_9d00_5d5d4b20f059.slice/crio-60ee6cc963fa0886f62da231d88cae42d78e19457e5274e8488a5717b1dbea1d WatchSource:0}: Error finding container 60ee6cc963fa0886f62da231d88cae42d78e19457e5274e8488a5717b1dbea1d: Status 404 returned error can't find the container with id 60ee6cc963fa0886f62da231d88cae42d78e19457e5274e8488a5717b1dbea1d Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.486262 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6mj2d" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.494352 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-5p6w2" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.503995 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/99582129-e50d-49c0-a554-aa703fc86452-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-z2k6h\" (UID: \"99582129-e50d-49c0-a554-aa703fc86452\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-z2k6h" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.517074 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zvqdh"] Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.517860 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9zv2s\" (UniqueName: \"kubernetes.io/projected/5864769c-0a13-4562-a407-f67e4348e1a6-kube-api-access-9zv2s\") pod \"machine-api-operator-5694c8668f-g64xp\" (UID: \"5864769c-0a13-4562-a407-f67e4348e1a6\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-g64xp" Jan 26 16:49:29 crc kubenswrapper[4754]: W0126 16:49:29.524575 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podadeda50d_627e_4071_a6ab_d7712750c22d.slice/crio-6c1e60cbdd3f5cf03dbb8db4053cb1bc9bae29c2b93b75e47e49f889fe077782 WatchSource:0}: Error finding container 6c1e60cbdd3f5cf03dbb8db4053cb1bc9bae29c2b93b75e47e49f889fe077782: Status 404 returned error can't find the container with id 6c1e60cbdd3f5cf03dbb8db4053cb1bc9bae29c2b93b75e47e49f889fe077782 Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.543334 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-llgrp\" (UniqueName: \"kubernetes.io/projected/587cde2e-d0a4-406b-ab72-26c7fe4b6707-kube-api-access-llgrp\") pod \"route-controller-manager-6576b87f9c-m2525\" (UID: \"587cde2e-d0a4-406b-ab72-26c7fe4b6707\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m2525" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.548037 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zvqdh" event={"ID":"adeda50d-627e-4071-a6ab-d7712750c22d","Type":"ContainerStarted","Data":"6c1e60cbdd3f5cf03dbb8db4053cb1bc9bae29c2b93b75e47e49f889fe077782"} Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.548761 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-rcrn5" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.551893 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-sz8hj" event={"ID":"aa57d7bf-eaea-4595-9d00-5d5d4b20f059","Type":"ContainerStarted","Data":"60ee6cc963fa0886f62da231d88cae42d78e19457e5274e8488a5717b1dbea1d"} Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.557712 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-4hkb9" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.560055 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6mzfw\" (UniqueName: \"kubernetes.io/projected/2a28f341-5f93-4d6c-bc4b-973d33dfeee1-kube-api-access-6mzfw\") pod \"oauth-openshift-558db77b4-k9sz7\" (UID: \"2a28f341-5f93-4d6c-bc4b-973d33dfeee1\") " pod="openshift-authentication/oauth-openshift-558db77b4-k9sz7" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.579921 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9l6hw\" (UniqueName: \"kubernetes.io/projected/f4d41d59-ebec-455b-9a17-2ff3f9a42f04-kube-api-access-9l6hw\") pod \"authentication-operator-69f744f599-kjc2g\" (UID: \"f4d41d59-ebec-455b-9a17-2ff3f9a42f04\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-kjc2g" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.583248 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-kh4hh" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.587535 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-fj2tb" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.594073 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-677j9"] Jan 26 16:49:29 crc kubenswrapper[4754]: W0126 16:49:29.630416 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4d81d21a_83f5_402e_91fd_d8268a6b49be.slice/crio-df3f65a2226610c65f3906b6fb76f4060198225d530bb9b11c214dfb40d1ed9f WatchSource:0}: Error finding container df3f65a2226610c65f3906b6fb76f4060198225d530bb9b11c214dfb40d1ed9f: Status 404 returned error can't find the container with id df3f65a2226610c65f3906b6fb76f4060198225d530bb9b11c214dfb40d1ed9f Jan 26 16:49:29 crc kubenswrapper[4754]: W0126 16:49:29.633768 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod06082490_21d0_4f81_8f02_b5ad80760b06.slice/crio-7431e8c786ba61a64c687a2596e58d9ef6159f19f5babeb95b4e4bf37198cf21 WatchSource:0}: Error finding container 7431e8c786ba61a64c687a2596e58d9ef6159f19f5babeb95b4e4bf37198cf21: Status 404 returned error can't find the container with id 7431e8c786ba61a64c687a2596e58d9ef6159f19f5babeb95b4e4bf37198cf21 Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.645461 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-mj6ns" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.646158 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pmc9d\" (UniqueName: \"kubernetes.io/projected/d098371a-9920-44d2-b918-f21c8142fac2-kube-api-access-pmc9d\") pod \"image-registry-697d97f7c8-47pnn\" (UID: \"d098371a-9920-44d2-b918-f21c8142fac2\") " pod="openshift-image-registry/image-registry-697d97f7c8-47pnn" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.646186 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bhpmf\" (UniqueName: \"kubernetes.io/projected/07245d2d-bd41-4bca-9f4a-fab5f4a48eb9-kube-api-access-bhpmf\") pod \"console-operator-58897d9998-wjdgm\" (UID: \"07245d2d-bd41-4bca-9f4a-fab5f4a48eb9\") " pod="openshift-console-operator/console-operator-58897d9998-wjdgm" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.646220 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/07245d2d-bd41-4bca-9f4a-fab5f4a48eb9-config\") pod \"console-operator-58897d9998-wjdgm\" (UID: \"07245d2d-bd41-4bca-9f4a-fab5f4a48eb9\") " pod="openshift-console-operator/console-operator-58897d9998-wjdgm" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.646259 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/07245d2d-bd41-4bca-9f4a-fab5f4a48eb9-trusted-ca\") pod \"console-operator-58897d9998-wjdgm\" (UID: \"07245d2d-bd41-4bca-9f4a-fab5f4a48eb9\") " pod="openshift-console-operator/console-operator-58897d9998-wjdgm" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.646303 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/693e0080-d2bc-427b-a81b-b09d1ce0e521-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-g4dbx\" (UID: \"693e0080-d2bc-427b-a81b-b09d1ce0e521\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-g4dbx" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.646348 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/d098371a-9920-44d2-b918-f21c8142fac2-installation-pull-secrets\") pod \"image-registry-697d97f7c8-47pnn\" (UID: \"d098371a-9920-44d2-b918-f21c8142fac2\") " pod="openshift-image-registry/image-registry-697d97f7c8-47pnn" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.646396 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-47pnn\" (UID: \"d098371a-9920-44d2-b918-f21c8142fac2\") " pod="openshift-image-registry/image-registry-697d97f7c8-47pnn" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.646412 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n7xpg\" (UniqueName: \"kubernetes.io/projected/693e0080-d2bc-427b-a81b-b09d1ce0e521-kube-api-access-n7xpg\") pod \"openshift-controller-manager-operator-756b6f6bc6-g4dbx\" (UID: \"693e0080-d2bc-427b-a81b-b09d1ce0e521\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-g4dbx" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.646433 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d098371a-9920-44d2-b918-f21c8142fac2-bound-sa-token\") pod \"image-registry-697d97f7c8-47pnn\" (UID: \"d098371a-9920-44d2-b918-f21c8142fac2\") " pod="openshift-image-registry/image-registry-697d97f7c8-47pnn" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.646480 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/d098371a-9920-44d2-b918-f21c8142fac2-registry-tls\") pod \"image-registry-697d97f7c8-47pnn\" (UID: \"d098371a-9920-44d2-b918-f21c8142fac2\") " pod="openshift-image-registry/image-registry-697d97f7c8-47pnn" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.646499 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/693e0080-d2bc-427b-a81b-b09d1ce0e521-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-g4dbx\" (UID: \"693e0080-d2bc-427b-a81b-b09d1ce0e521\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-g4dbx" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.646537 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/d098371a-9920-44d2-b918-f21c8142fac2-ca-trust-extracted\") pod \"image-registry-697d97f7c8-47pnn\" (UID: \"d098371a-9920-44d2-b918-f21c8142fac2\") " pod="openshift-image-registry/image-registry-697d97f7c8-47pnn" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.646554 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/07245d2d-bd41-4bca-9f4a-fab5f4a48eb9-serving-cert\") pod \"console-operator-58897d9998-wjdgm\" (UID: \"07245d2d-bd41-4bca-9f4a-fab5f4a48eb9\") " pod="openshift-console-operator/console-operator-58897d9998-wjdgm" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.646584 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/97dfbf03-7b95-4645-afa1-f080330f4cbe-metrics-tls\") pod \"ingress-operator-5b745b69d9-t4r22\" (UID: \"97dfbf03-7b95-4645-afa1-f080330f4cbe\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-t4r22" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.646623 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/97dfbf03-7b95-4645-afa1-f080330f4cbe-trusted-ca\") pod \"ingress-operator-5b745b69d9-t4r22\" (UID: \"97dfbf03-7b95-4645-afa1-f080330f4cbe\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-t4r22" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.647561 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/97dfbf03-7b95-4645-afa1-f080330f4cbe-bound-sa-token\") pod \"ingress-operator-5b745b69d9-t4r22\" (UID: \"97dfbf03-7b95-4645-afa1-f080330f4cbe\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-t4r22" Jan 26 16:49:29 crc kubenswrapper[4754]: E0126 16:49:29.647587 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 16:49:30.147568266 +0000 UTC m=+136.671748780 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-47pnn" (UID: "d098371a-9920-44d2-b918-f21c8142fac2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.647624 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rqwvv\" (UniqueName: \"kubernetes.io/projected/97dfbf03-7b95-4645-afa1-f080330f4cbe-kube-api-access-rqwvv\") pod \"ingress-operator-5b745b69d9-t4r22\" (UID: \"97dfbf03-7b95-4645-afa1-f080330f4cbe\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-t4r22" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.647627 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-8d467"] Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.647661 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/d098371a-9920-44d2-b918-f21c8142fac2-registry-certificates\") pod \"image-registry-697d97f7c8-47pnn\" (UID: \"d098371a-9920-44d2-b918-f21c8142fac2\") " pod="openshift-image-registry/image-registry-697d97f7c8-47pnn" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.647701 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d098371a-9920-44d2-b918-f21c8142fac2-trusted-ca\") pod \"image-registry-697d97f7c8-47pnn\" (UID: \"d098371a-9920-44d2-b918-f21c8142fac2\") " pod="openshift-image-registry/image-registry-697d97f7c8-47pnn" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.654500 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m2525" Jan 26 16:49:29 crc kubenswrapper[4754]: W0126 16:49:29.673766 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod10609d5f_d6dd_40a5_acc4_63e62584d18e.slice/crio-832ebba41a7c68f635afe4ae7ff14cc3da025a4620cd4e90a572e66731b401a4 WatchSource:0}: Error finding container 832ebba41a7c68f635afe4ae7ff14cc3da025a4620cd4e90a572e66731b401a4: Status 404 returned error can't find the container with id 832ebba41a7c68f635afe4ae7ff14cc3da025a4620cd4e90a572e66731b401a4 Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.696858 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-kjc2g" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.712529 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-ptrdc" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.721424 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-6mj2d"] Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.746360 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-g64xp" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.748882 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 16:49:29 crc kubenswrapper[4754]: E0126 16:49:29.749082 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 16:49:30.249057748 +0000 UTC m=+136.773238192 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.750187 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d098371a-9920-44d2-b918-f21c8142fac2-bound-sa-token\") pod \"image-registry-697d97f7c8-47pnn\" (UID: \"d098371a-9920-44d2-b918-f21c8142fac2\") " pod="openshift-image-registry/image-registry-697d97f7c8-47pnn" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.750247 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/a411958e-77f0-45bb-a11c-abd88d756dee-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-w7zlr\" (UID: \"a411958e-77f0-45bb-a11c-abd88d756dee\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-w7zlr" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.750307 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h7gfl\" (UniqueName: \"kubernetes.io/projected/713fa157-05f9-4c30-b7e2-47a2b47cf7c4-kube-api-access-h7gfl\") pod \"migrator-59844c95c7-k5sf5\" (UID: \"713fa157-05f9-4c30-b7e2-47a2b47cf7c4\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-k5sf5" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.750325 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/12e696fe-360e-4ea7-93e6-5ca2d0fc8123-plugins-dir\") pod \"csi-hostpathplugin-9shrr\" (UID: \"12e696fe-360e-4ea7-93e6-5ca2d0fc8123\") " pod="hostpath-provisioner/csi-hostpathplugin-9shrr" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.750355 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3f0de367-82c0-4548-8f81-8c3482795003-config\") pod \"service-ca-operator-777779d784-vfldk\" (UID: \"3f0de367-82c0-4548-8f81-8c3482795003\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-vfldk" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.750811 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9caa5ea4-1245-4dc7-af31-571f68cec263-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-8rjs9\" (UID: \"9caa5ea4-1245-4dc7-af31-571f68cec263\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8rjs9" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.757507 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f5fabb97-e818-4c00-8023-db5e8477878f-config-volume\") pod \"collect-profiles-29490765-w8zqq\" (UID: \"f5fabb97-e818-4c00-8023-db5e8477878f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29490765-w8zqq" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.757822 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/d098371a-9920-44d2-b918-f21c8142fac2-registry-tls\") pod \"image-registry-697d97f7c8-47pnn\" (UID: \"d098371a-9920-44d2-b918-f21c8142fac2\") " pod="openshift-image-registry/image-registry-697d97f7c8-47pnn" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.757873 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/4d0ec6e9-f4f7-4751-8317-60940c63f320-proxy-tls\") pod \"machine-config-controller-84d6567774-8rhs2\" (UID: \"4d0ec6e9-f4f7-4751-8317-60940c63f320\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8rhs2" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.757901 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/12e696fe-360e-4ea7-93e6-5ca2d0fc8123-mountpoint-dir\") pod \"csi-hostpathplugin-9shrr\" (UID: \"12e696fe-360e-4ea7-93e6-5ca2d0fc8123\") " pod="hostpath-provisioner/csi-hostpathplugin-9shrr" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.758061 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/d098371a-9920-44d2-b918-f21c8142fac2-ca-trust-extracted\") pod \"image-registry-697d97f7c8-47pnn\" (UID: \"d098371a-9920-44d2-b918-f21c8142fac2\") " pod="openshift-image-registry/image-registry-697d97f7c8-47pnn" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.758159 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/07245d2d-bd41-4bca-9f4a-fab5f4a48eb9-serving-cert\") pod \"console-operator-58897d9998-wjdgm\" (UID: \"07245d2d-bd41-4bca-9f4a-fab5f4a48eb9\") " pod="openshift-console-operator/console-operator-58897d9998-wjdgm" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.758183 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fa9a982c-684d-497d-b8eb-f921e00fe511-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-8647g\" (UID: \"fa9a982c-684d-497d-b8eb-f921e00fe511\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-8647g" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.758221 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wc27q\" (UniqueName: \"kubernetes.io/projected/12e696fe-360e-4ea7-93e6-5ca2d0fc8123-kube-api-access-wc27q\") pod \"csi-hostpathplugin-9shrr\" (UID: \"12e696fe-360e-4ea7-93e6-5ca2d0fc8123\") " pod="hostpath-provisioner/csi-hostpathplugin-9shrr" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.758252 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/50c60179-2a08-4055-afb0-78aff31d6465-profile-collector-cert\") pod \"catalog-operator-68c6474976-4px9m\" (UID: \"50c60179-2a08-4055-afb0-78aff31d6465\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-4px9m" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.758270 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/2644a03c-ada2-45c0-ac4e-edfd969a7752-srv-cert\") pod \"olm-operator-6b444d44fb-fbbxr\" (UID: \"2644a03c-ada2-45c0-ac4e-edfd969a7752\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-fbbxr" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.758344 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/44e9e245-6889-4706-bc76-a22f2e7f6019-metrics-certs\") pod \"router-default-5444994796-jxzst\" (UID: \"44e9e245-6889-4706-bc76-a22f2e7f6019\") " pod="openshift-ingress/router-default-5444994796-jxzst" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.758394 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/593039fd-1af9-41c5-8f1a-bce2d698da75-config-volume\") pod \"dns-default-964q6\" (UID: \"593039fd-1af9-41c5-8f1a-bce2d698da75\") " pod="openshift-dns/dns-default-964q6" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.758415 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/221280be-801c-4f1c-abbc-cb8a02f47838-config\") pod \"kube-apiserver-operator-766d6c64bb-m2ktn\" (UID: \"221280be-801c-4f1c-abbc-cb8a02f47838\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-m2ktn" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.758478 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/fa9a982c-684d-497d-b8eb-f921e00fe511-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-8647g\" (UID: \"fa9a982c-684d-497d-b8eb-f921e00fe511\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-8647g" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.758498 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z8tbc\" (UniqueName: \"kubernetes.io/projected/a9e3ba79-4e68-41af-823a-be5c4494df86-kube-api-access-z8tbc\") pod \"ingress-canary-lxx9x\" (UID: \"a9e3ba79-4e68-41af-823a-be5c4494df86\") " pod="openshift-ingress-canary/ingress-canary-lxx9x" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.758638 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/97dfbf03-7b95-4645-afa1-f080330f4cbe-bound-sa-token\") pod \"ingress-operator-5b745b69d9-t4r22\" (UID: \"97dfbf03-7b95-4645-afa1-f080330f4cbe\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-t4r22" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.758686 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rqwvv\" (UniqueName: \"kubernetes.io/projected/97dfbf03-7b95-4645-afa1-f080330f4cbe-kube-api-access-rqwvv\") pod \"ingress-operator-5b745b69d9-t4r22\" (UID: \"97dfbf03-7b95-4645-afa1-f080330f4cbe\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-t4r22" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.758715 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d098371a-9920-44d2-b918-f21c8142fac2-trusted-ca\") pod \"image-registry-697d97f7c8-47pnn\" (UID: \"d098371a-9920-44d2-b918-f21c8142fac2\") " pod="openshift-image-registry/image-registry-697d97f7c8-47pnn" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.758828 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/44e9e245-6889-4706-bc76-a22f2e7f6019-service-ca-bundle\") pod \"router-default-5444994796-jxzst\" (UID: \"44e9e245-6889-4706-bc76-a22f2e7f6019\") " pod="openshift-ingress/router-default-5444994796-jxzst" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.758880 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fmqgc\" (UniqueName: \"kubernetes.io/projected/4d0ec6e9-f4f7-4751-8317-60940c63f320-kube-api-access-fmqgc\") pod \"machine-config-controller-84d6567774-8rhs2\" (UID: \"4d0ec6e9-f4f7-4751-8317-60940c63f320\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8rhs2" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.758966 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/e89d71d4-488f-4bed-9356-5dc92d5a8eae-signing-key\") pod \"service-ca-9c57cc56f-472pk\" (UID: \"e89d71d4-488f-4bed-9356-5dc92d5a8eae\") " pod="openshift-service-ca/service-ca-9c57cc56f-472pk" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.758990 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/e89d71d4-488f-4bed-9356-5dc92d5a8eae-signing-cabundle\") pod \"service-ca-9c57cc56f-472pk\" (UID: \"e89d71d4-488f-4bed-9356-5dc92d5a8eae\") " pod="openshift-service-ca/service-ca-9c57cc56f-472pk" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.759047 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vlhnl\" (UniqueName: \"kubernetes.io/projected/e89d71d4-488f-4bed-9356-5dc92d5a8eae-kube-api-access-vlhnl\") pod \"service-ca-9c57cc56f-472pk\" (UID: \"e89d71d4-488f-4bed-9356-5dc92d5a8eae\") " pod="openshift-service-ca/service-ca-9c57cc56f-472pk" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.759072 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/d0295627-eb92-49f1-9e39-82fd0d1ec864-certs\") pod \"machine-config-server-wp5r6\" (UID: \"d0295627-eb92-49f1-9e39-82fd0d1ec864\") " pod="openshift-machine-config-operator/machine-config-server-wp5r6" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.759127 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9caa5ea4-1245-4dc7-af31-571f68cec263-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-8rjs9\" (UID: \"9caa5ea4-1245-4dc7-af31-571f68cec263\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8rjs9" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.759203 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/221280be-801c-4f1c-abbc-cb8a02f47838-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-m2ktn\" (UID: \"221280be-801c-4f1c-abbc-cb8a02f47838\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-m2ktn" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.759231 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/d0295627-eb92-49f1-9e39-82fd0d1ec864-node-bootstrap-token\") pod \"machine-config-server-wp5r6\" (UID: \"d0295627-eb92-49f1-9e39-82fd0d1ec864\") " pod="openshift-machine-config-operator/machine-config-server-wp5r6" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.759280 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qsqtw\" (UniqueName: \"kubernetes.io/projected/a411958e-77f0-45bb-a11c-abd88d756dee-kube-api-access-qsqtw\") pod \"control-plane-machine-set-operator-78cbb6b69f-w7zlr\" (UID: \"a411958e-77f0-45bb-a11c-abd88d756dee\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-w7zlr" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.759315 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pmc9d\" (UniqueName: \"kubernetes.io/projected/d098371a-9920-44d2-b918-f21c8142fac2-kube-api-access-pmc9d\") pod \"image-registry-697d97f7c8-47pnn\" (UID: \"d098371a-9920-44d2-b918-f21c8142fac2\") " pod="openshift-image-registry/image-registry-697d97f7c8-47pnn" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.759354 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bhpmf\" (UniqueName: \"kubernetes.io/projected/07245d2d-bd41-4bca-9f4a-fab5f4a48eb9-kube-api-access-bhpmf\") pod \"console-operator-58897d9998-wjdgm\" (UID: \"07245d2d-bd41-4bca-9f4a-fab5f4a48eb9\") " pod="openshift-console-operator/console-operator-58897d9998-wjdgm" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.759376 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bdd2421f-ccc8-4500-8fd2-23a5f58173df-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-2kmrb\" (UID: \"bdd2421f-ccc8-4500-8fd2-23a5f58173df\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-2kmrb" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.759404 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/07245d2d-bd41-4bca-9f4a-fab5f4a48eb9-config\") pod \"console-operator-58897d9998-wjdgm\" (UID: \"07245d2d-bd41-4bca-9f4a-fab5f4a48eb9\") " pod="openshift-console-operator/console-operator-58897d9998-wjdgm" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.759453 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8nwnd\" (UniqueName: \"kubernetes.io/projected/2644a03c-ada2-45c0-ac4e-edfd969a7752-kube-api-access-8nwnd\") pod \"olm-operator-6b444d44fb-fbbxr\" (UID: \"2644a03c-ada2-45c0-ac4e-edfd969a7752\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-fbbxr" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.759480 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/12e696fe-360e-4ea7-93e6-5ca2d0fc8123-registration-dir\") pod \"csi-hostpathplugin-9shrr\" (UID: \"12e696fe-360e-4ea7-93e6-5ca2d0fc8123\") " pod="hostpath-provisioner/csi-hostpathplugin-9shrr" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.759636 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/07245d2d-bd41-4bca-9f4a-fab5f4a48eb9-trusted-ca\") pod \"console-operator-58897d9998-wjdgm\" (UID: \"07245d2d-bd41-4bca-9f4a-fab5f4a48eb9\") " pod="openshift-console-operator/console-operator-58897d9998-wjdgm" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.759691 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/221280be-801c-4f1c-abbc-cb8a02f47838-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-m2ktn\" (UID: \"221280be-801c-4f1c-abbc-cb8a02f47838\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-m2ktn" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.759723 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cgtt5\" (UniqueName: \"kubernetes.io/projected/50c60179-2a08-4055-afb0-78aff31d6465-kube-api-access-cgtt5\") pod \"catalog-operator-68c6474976-4px9m\" (UID: \"50c60179-2a08-4055-afb0-78aff31d6465\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-4px9m" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.759790 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/693e0080-d2bc-427b-a81b-b09d1ce0e521-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-g4dbx\" (UID: \"693e0080-d2bc-427b-a81b-b09d1ce0e521\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-g4dbx" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.759815 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/44e9e245-6889-4706-bc76-a22f2e7f6019-default-certificate\") pod \"router-default-5444994796-jxzst\" (UID: \"44e9e245-6889-4706-bc76-a22f2e7f6019\") " pod="openshift-ingress/router-default-5444994796-jxzst" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.759901 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kmmqb\" (UniqueName: \"kubernetes.io/projected/d0295627-eb92-49f1-9e39-82fd0d1ec864-kube-api-access-kmmqb\") pod \"machine-config-server-wp5r6\" (UID: \"d0295627-eb92-49f1-9e39-82fd0d1ec864\") " pod="openshift-machine-config-operator/machine-config-server-wp5r6" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.759994 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/d098371a-9920-44d2-b918-f21c8142fac2-installation-pull-secrets\") pod \"image-registry-697d97f7c8-47pnn\" (UID: \"d098371a-9920-44d2-b918-f21c8142fac2\") " pod="openshift-image-registry/image-registry-697d97f7c8-47pnn" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.760057 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/50c60179-2a08-4055-afb0-78aff31d6465-srv-cert\") pod \"catalog-operator-68c6474976-4px9m\" (UID: \"50c60179-2a08-4055-afb0-78aff31d6465\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-4px9m" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.760062 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/d098371a-9920-44d2-b918-f21c8142fac2-ca-trust-extracted\") pod \"image-registry-697d97f7c8-47pnn\" (UID: \"d098371a-9920-44d2-b918-f21c8142fac2\") " pod="openshift-image-registry/image-registry-697d97f7c8-47pnn" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.760107 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n7xpg\" (UniqueName: \"kubernetes.io/projected/693e0080-d2bc-427b-a81b-b09d1ce0e521-kube-api-access-n7xpg\") pod \"openshift-controller-manager-operator-756b6f6bc6-g4dbx\" (UID: \"693e0080-d2bc-427b-a81b-b09d1ce0e521\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-g4dbx" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.760136 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3f0de367-82c0-4548-8f81-8c3482795003-serving-cert\") pod \"service-ca-operator-777779d784-vfldk\" (UID: \"3f0de367-82c0-4548-8f81-8c3482795003\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-vfldk" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.760233 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/4d0ec6e9-f4f7-4751-8317-60940c63f320-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-8rhs2\" (UID: \"4d0ec6e9-f4f7-4751-8317-60940c63f320\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8rhs2" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.760288 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h4gqm\" (UniqueName: \"kubernetes.io/projected/44e9e245-6889-4706-bc76-a22f2e7f6019-kube-api-access-h4gqm\") pod \"router-default-5444994796-jxzst\" (UID: \"44e9e245-6889-4706-bc76-a22f2e7f6019\") " pod="openshift-ingress/router-default-5444994796-jxzst" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.760350 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gl9hl\" (UniqueName: \"kubernetes.io/projected/be3e5760-2f70-423b-9dd8-b5eeb0482b80-kube-api-access-gl9hl\") pod \"packageserver-d55dfcdfc-7lgch\" (UID: \"be3e5760-2f70-423b-9dd8-b5eeb0482b80\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7lgch" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.760458 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/432c543c-59d1-49c9-ba43-91be5e113975-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-7tnbd\" (UID: \"432c543c-59d1-49c9-ba43-91be5e113975\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-7tnbd" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.760482 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/be3e5760-2f70-423b-9dd8-b5eeb0482b80-webhook-cert\") pod \"packageserver-d55dfcdfc-7lgch\" (UID: \"be3e5760-2f70-423b-9dd8-b5eeb0482b80\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7lgch" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.760532 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gtt7p\" (UniqueName: \"kubernetes.io/projected/5de8942d-833c-499b-842a-ddb2eeeab601-kube-api-access-gtt7p\") pod \"package-server-manager-789f6589d5-95dbg\" (UID: \"5de8942d-833c-499b-842a-ddb2eeeab601\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-95dbg" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.760612 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/693e0080-d2bc-427b-a81b-b09d1ce0e521-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-g4dbx\" (UID: \"693e0080-d2bc-427b-a81b-b09d1ce0e521\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-g4dbx" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.760783 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/97dfbf03-7b95-4645-afa1-f080330f4cbe-metrics-tls\") pod \"ingress-operator-5b745b69d9-t4r22\" (UID: \"97dfbf03-7b95-4645-afa1-f080330f4cbe\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-t4r22" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.760835 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bcsps\" (UniqueName: \"kubernetes.io/projected/3f0de367-82c0-4548-8f81-8c3482795003-kube-api-access-bcsps\") pod \"service-ca-operator-777779d784-vfldk\" (UID: \"3f0de367-82c0-4548-8f81-8c3482795003\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-vfldk" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.761721 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d098371a-9920-44d2-b918-f21c8142fac2-trusted-ca\") pod \"image-registry-697d97f7c8-47pnn\" (UID: \"d098371a-9920-44d2-b918-f21c8142fac2\") " pod="openshift-image-registry/image-registry-697d97f7c8-47pnn" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.761872 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/97dfbf03-7b95-4645-afa1-f080330f4cbe-trusted-ca\") pod \"ingress-operator-5b745b69d9-t4r22\" (UID: \"97dfbf03-7b95-4645-afa1-f080330f4cbe\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-t4r22" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.761963 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mt7kw\" (UniqueName: \"kubernetes.io/projected/f5fabb97-e818-4c00-8023-db5e8477878f-kube-api-access-mt7kw\") pod \"collect-profiles-29490765-w8zqq\" (UID: \"f5fabb97-e818-4c00-8023-db5e8477878f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29490765-w8zqq" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.761998 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f06be1d1-4134-4ad7-9233-b4099687bfe2-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-7xsg2\" (UID: \"f06be1d1-4134-4ad7-9233-b4099687bfe2\") " pod="openshift-marketplace/marketplace-operator-79b997595-7xsg2" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.762575 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/07245d2d-bd41-4bca-9f4a-fab5f4a48eb9-config\") pod \"console-operator-58897d9998-wjdgm\" (UID: \"07245d2d-bd41-4bca-9f4a-fab5f4a48eb9\") " pod="openshift-console-operator/console-operator-58897d9998-wjdgm" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.762885 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/d098371a-9920-44d2-b918-f21c8142fac2-registry-tls\") pod \"image-registry-697d97f7c8-47pnn\" (UID: \"d098371a-9920-44d2-b918-f21c8142fac2\") " pod="openshift-image-registry/image-registry-697d97f7c8-47pnn" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.765568 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/07245d2d-bd41-4bca-9f4a-fab5f4a48eb9-serving-cert\") pod \"console-operator-58897d9998-wjdgm\" (UID: \"07245d2d-bd41-4bca-9f4a-fab5f4a48eb9\") " pod="openshift-console-operator/console-operator-58897d9998-wjdgm" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.766167 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/07245d2d-bd41-4bca-9f4a-fab5f4a48eb9-trusted-ca\") pod \"console-operator-58897d9998-wjdgm\" (UID: \"07245d2d-bd41-4bca-9f4a-fab5f4a48eb9\") " pod="openshift-console-operator/console-operator-58897d9998-wjdgm" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.767522 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/97dfbf03-7b95-4645-afa1-f080330f4cbe-trusted-ca\") pod \"ingress-operator-5b745b69d9-t4r22\" (UID: \"97dfbf03-7b95-4645-afa1-f080330f4cbe\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-t4r22" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.768680 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bdd2421f-ccc8-4500-8fd2-23a5f58173df-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-2kmrb\" (UID: \"bdd2421f-ccc8-4500-8fd2-23a5f58173df\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-2kmrb" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.768728 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/2644a03c-ada2-45c0-ac4e-edfd969a7752-profile-collector-cert\") pod \"olm-operator-6b444d44fb-fbbxr\" (UID: \"2644a03c-ada2-45c0-ac4e-edfd969a7752\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-fbbxr" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.769055 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/d098371a-9920-44d2-b918-f21c8142fac2-registry-certificates\") pod \"image-registry-697d97f7c8-47pnn\" (UID: \"d098371a-9920-44d2-b918-f21c8142fac2\") " pod="openshift-image-registry/image-registry-697d97f7c8-47pnn" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.769102 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bdd2421f-ccc8-4500-8fd2-23a5f58173df-config\") pod \"kube-controller-manager-operator-78b949d7b-2kmrb\" (UID: \"bdd2421f-ccc8-4500-8fd2-23a5f58173df\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-2kmrb" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.769132 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/12e696fe-360e-4ea7-93e6-5ca2d0fc8123-csi-data-dir\") pod \"csi-hostpathplugin-9shrr\" (UID: \"12e696fe-360e-4ea7-93e6-5ca2d0fc8123\") " pod="hostpath-provisioner/csi-hostpathplugin-9shrr" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.769164 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/f06be1d1-4134-4ad7-9233-b4099687bfe2-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-7xsg2\" (UID: \"f06be1d1-4134-4ad7-9233-b4099687bfe2\") " pod="openshift-marketplace/marketplace-operator-79b997595-7xsg2" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.769199 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p8pxk\" (UniqueName: \"kubernetes.io/projected/9caa5ea4-1245-4dc7-af31-571f68cec263-kube-api-access-p8pxk\") pod \"kube-storage-version-migrator-operator-b67b599dd-8rjs9\" (UID: \"9caa5ea4-1245-4dc7-af31-571f68cec263\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8rjs9" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.769283 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wgkrw\" (UniqueName: \"kubernetes.io/projected/593039fd-1af9-41c5-8f1a-bce2d698da75-kube-api-access-wgkrw\") pod \"dns-default-964q6\" (UID: \"593039fd-1af9-41c5-8f1a-bce2d698da75\") " pod="openshift-dns/dns-default-964q6" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.769470 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a9e3ba79-4e68-41af-823a-be5c4494df86-cert\") pod \"ingress-canary-lxx9x\" (UID: \"a9e3ba79-4e68-41af-823a-be5c4494df86\") " pod="openshift-ingress-canary/ingress-canary-lxx9x" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.769522 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/44e9e245-6889-4706-bc76-a22f2e7f6019-stats-auth\") pod \"router-default-5444994796-jxzst\" (UID: \"44e9e245-6889-4706-bc76-a22f2e7f6019\") " pod="openshift-ingress/router-default-5444994796-jxzst" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.769550 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zszpm\" (UniqueName: \"kubernetes.io/projected/f06be1d1-4134-4ad7-9233-b4099687bfe2-kube-api-access-zszpm\") pod \"marketplace-operator-79b997595-7xsg2\" (UID: \"f06be1d1-4134-4ad7-9233-b4099687bfe2\") " pod="openshift-marketplace/marketplace-operator-79b997595-7xsg2" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.769597 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fa9a982c-684d-497d-b8eb-f921e00fe511-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-8647g\" (UID: \"fa9a982c-684d-497d-b8eb-f921e00fe511\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-8647g" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.771177 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/be3e5760-2f70-423b-9dd8-b5eeb0482b80-tmpfs\") pod \"packageserver-d55dfcdfc-7lgch\" (UID: \"be3e5760-2f70-423b-9dd8-b5eeb0482b80\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7lgch" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.771210 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/593039fd-1af9-41c5-8f1a-bce2d698da75-metrics-tls\") pod \"dns-default-964q6\" (UID: \"593039fd-1af9-41c5-8f1a-bce2d698da75\") " pod="openshift-dns/dns-default-964q6" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.771370 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4chnr\" (UniqueName: \"kubernetes.io/projected/432c543c-59d1-49c9-ba43-91be5e113975-kube-api-access-4chnr\") pod \"multus-admission-controller-857f4d67dd-7tnbd\" (UID: \"432c543c-59d1-49c9-ba43-91be5e113975\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-7tnbd" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.771404 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/12e696fe-360e-4ea7-93e6-5ca2d0fc8123-socket-dir\") pod \"csi-hostpathplugin-9shrr\" (UID: \"12e696fe-360e-4ea7-93e6-5ca2d0fc8123\") " pod="hostpath-provisioner/csi-hostpathplugin-9shrr" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.771549 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/5de8942d-833c-499b-842a-ddb2eeeab601-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-95dbg\" (UID: \"5de8942d-833c-499b-842a-ddb2eeeab601\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-95dbg" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.771605 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f5fabb97-e818-4c00-8023-db5e8477878f-secret-volume\") pod \"collect-profiles-29490765-w8zqq\" (UID: \"f5fabb97-e818-4c00-8023-db5e8477878f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29490765-w8zqq" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.771953 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-47pnn\" (UID: \"d098371a-9920-44d2-b918-f21c8142fac2\") " pod="openshift-image-registry/image-registry-697d97f7c8-47pnn" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.771992 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/be3e5760-2f70-423b-9dd8-b5eeb0482b80-apiservice-cert\") pod \"packageserver-d55dfcdfc-7lgch\" (UID: \"be3e5760-2f70-423b-9dd8-b5eeb0482b80\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7lgch" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.776404 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/d098371a-9920-44d2-b918-f21c8142fac2-installation-pull-secrets\") pod \"image-registry-697d97f7c8-47pnn\" (UID: \"d098371a-9920-44d2-b918-f21c8142fac2\") " pod="openshift-image-registry/image-registry-697d97f7c8-47pnn" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.778870 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/d098371a-9920-44d2-b918-f21c8142fac2-registry-certificates\") pod \"image-registry-697d97f7c8-47pnn\" (UID: \"d098371a-9920-44d2-b918-f21c8142fac2\") " pod="openshift-image-registry/image-registry-697d97f7c8-47pnn" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.782645 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-z2k6h" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.784295 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/693e0080-d2bc-427b-a81b-b09d1ce0e521-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-g4dbx\" (UID: \"693e0080-d2bc-427b-a81b-b09d1ce0e521\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-g4dbx" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.786762 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-k9sz7" Jan 26 16:49:29 crc kubenswrapper[4754]: E0126 16:49:29.787407 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 16:49:30.287171105 +0000 UTC m=+136.811351539 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-47pnn" (UID: "d098371a-9920-44d2-b918-f21c8142fac2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.801260 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/97dfbf03-7b95-4645-afa1-f080330f4cbe-metrics-tls\") pod \"ingress-operator-5b745b69d9-t4r22\" (UID: \"97dfbf03-7b95-4645-afa1-f080330f4cbe\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-t4r22" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.802401 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/693e0080-d2bc-427b-a81b-b09d1ce0e521-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-g4dbx\" (UID: \"693e0080-d2bc-427b-a81b-b09d1ce0e521\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-g4dbx" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.807134 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d098371a-9920-44d2-b918-f21c8142fac2-bound-sa-token\") pod \"image-registry-697d97f7c8-47pnn\" (UID: \"d098371a-9920-44d2-b918-f21c8142fac2\") " pod="openshift-image-registry/image-registry-697d97f7c8-47pnn" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.812883 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-5p6w2"] Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.812917 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-rcrn5"] Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.816620 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/97dfbf03-7b95-4645-afa1-f080330f4cbe-bound-sa-token\") pod \"ingress-operator-5b745b69d9-t4r22\" (UID: \"97dfbf03-7b95-4645-afa1-f080330f4cbe\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-t4r22" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.837111 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rqwvv\" (UniqueName: \"kubernetes.io/projected/97dfbf03-7b95-4645-afa1-f080330f4cbe-kube-api-access-rqwvv\") pod \"ingress-operator-5b745b69d9-t4r22\" (UID: \"97dfbf03-7b95-4645-afa1-f080330f4cbe\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-t4r22" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.861718 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pmc9d\" (UniqueName: \"kubernetes.io/projected/d098371a-9920-44d2-b918-f21c8142fac2-kube-api-access-pmc9d\") pod \"image-registry-697d97f7c8-47pnn\" (UID: \"d098371a-9920-44d2-b918-f21c8142fac2\") " pod="openshift-image-registry/image-registry-697d97f7c8-47pnn" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.867560 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-fj2tb"] Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.871996 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-t4r22" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.872836 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.873152 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/12e696fe-360e-4ea7-93e6-5ca2d0fc8123-registration-dir\") pod \"csi-hostpathplugin-9shrr\" (UID: \"12e696fe-360e-4ea7-93e6-5ca2d0fc8123\") " pod="hostpath-provisioner/csi-hostpathplugin-9shrr" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.873193 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cgtt5\" (UniqueName: \"kubernetes.io/projected/50c60179-2a08-4055-afb0-78aff31d6465-kube-api-access-cgtt5\") pod \"catalog-operator-68c6474976-4px9m\" (UID: \"50c60179-2a08-4055-afb0-78aff31d6465\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-4px9m" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.873224 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/221280be-801c-4f1c-abbc-cb8a02f47838-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-m2ktn\" (UID: \"221280be-801c-4f1c-abbc-cb8a02f47838\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-m2ktn" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.873251 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/44e9e245-6889-4706-bc76-a22f2e7f6019-default-certificate\") pod \"router-default-5444994796-jxzst\" (UID: \"44e9e245-6889-4706-bc76-a22f2e7f6019\") " pod="openshift-ingress/router-default-5444994796-jxzst" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.873275 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kmmqb\" (UniqueName: \"kubernetes.io/projected/d0295627-eb92-49f1-9e39-82fd0d1ec864-kube-api-access-kmmqb\") pod \"machine-config-server-wp5r6\" (UID: \"d0295627-eb92-49f1-9e39-82fd0d1ec864\") " pod="openshift-machine-config-operator/machine-config-server-wp5r6" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.873300 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/50c60179-2a08-4055-afb0-78aff31d6465-srv-cert\") pod \"catalog-operator-68c6474976-4px9m\" (UID: \"50c60179-2a08-4055-afb0-78aff31d6465\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-4px9m" Jan 26 16:49:29 crc kubenswrapper[4754]: E0126 16:49:29.873404 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 16:49:30.373357468 +0000 UTC m=+136.897537902 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.873518 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3f0de367-82c0-4548-8f81-8c3482795003-serving-cert\") pod \"service-ca-operator-777779d784-vfldk\" (UID: \"3f0de367-82c0-4548-8f81-8c3482795003\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-vfldk" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.873561 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/4d0ec6e9-f4f7-4751-8317-60940c63f320-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-8rhs2\" (UID: \"4d0ec6e9-f4f7-4751-8317-60940c63f320\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8rhs2" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.873590 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h4gqm\" (UniqueName: \"kubernetes.io/projected/44e9e245-6889-4706-bc76-a22f2e7f6019-kube-api-access-h4gqm\") pod \"router-default-5444994796-jxzst\" (UID: \"44e9e245-6889-4706-bc76-a22f2e7f6019\") " pod="openshift-ingress/router-default-5444994796-jxzst" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.873620 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gl9hl\" (UniqueName: \"kubernetes.io/projected/be3e5760-2f70-423b-9dd8-b5eeb0482b80-kube-api-access-gl9hl\") pod \"packageserver-d55dfcdfc-7lgch\" (UID: \"be3e5760-2f70-423b-9dd8-b5eeb0482b80\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7lgch" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.873661 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/be3e5760-2f70-423b-9dd8-b5eeb0482b80-webhook-cert\") pod \"packageserver-d55dfcdfc-7lgch\" (UID: \"be3e5760-2f70-423b-9dd8-b5eeb0482b80\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7lgch" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.873696 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gtt7p\" (UniqueName: \"kubernetes.io/projected/5de8942d-833c-499b-842a-ddb2eeeab601-kube-api-access-gtt7p\") pod \"package-server-manager-789f6589d5-95dbg\" (UID: \"5de8942d-833c-499b-842a-ddb2eeeab601\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-95dbg" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.873717 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/432c543c-59d1-49c9-ba43-91be5e113975-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-7tnbd\" (UID: \"432c543c-59d1-49c9-ba43-91be5e113975\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-7tnbd" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.873751 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bcsps\" (UniqueName: \"kubernetes.io/projected/3f0de367-82c0-4548-8f81-8c3482795003-kube-api-access-bcsps\") pod \"service-ca-operator-777779d784-vfldk\" (UID: \"3f0de367-82c0-4548-8f81-8c3482795003\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-vfldk" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.873799 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mt7kw\" (UniqueName: \"kubernetes.io/projected/f5fabb97-e818-4c00-8023-db5e8477878f-kube-api-access-mt7kw\") pod \"collect-profiles-29490765-w8zqq\" (UID: \"f5fabb97-e818-4c00-8023-db5e8477878f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29490765-w8zqq" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.873820 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f06be1d1-4134-4ad7-9233-b4099687bfe2-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-7xsg2\" (UID: \"f06be1d1-4134-4ad7-9233-b4099687bfe2\") " pod="openshift-marketplace/marketplace-operator-79b997595-7xsg2" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.873850 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bdd2421f-ccc8-4500-8fd2-23a5f58173df-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-2kmrb\" (UID: \"bdd2421f-ccc8-4500-8fd2-23a5f58173df\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-2kmrb" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.873859 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/12e696fe-360e-4ea7-93e6-5ca2d0fc8123-registration-dir\") pod \"csi-hostpathplugin-9shrr\" (UID: \"12e696fe-360e-4ea7-93e6-5ca2d0fc8123\") " pod="hostpath-provisioner/csi-hostpathplugin-9shrr" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.873886 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bdd2421f-ccc8-4500-8fd2-23a5f58173df-config\") pod \"kube-controller-manager-operator-78b949d7b-2kmrb\" (UID: \"bdd2421f-ccc8-4500-8fd2-23a5f58173df\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-2kmrb" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.873911 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/2644a03c-ada2-45c0-ac4e-edfd969a7752-profile-collector-cert\") pod \"olm-operator-6b444d44fb-fbbxr\" (UID: \"2644a03c-ada2-45c0-ac4e-edfd969a7752\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-fbbxr" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.873939 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/12e696fe-360e-4ea7-93e6-5ca2d0fc8123-csi-data-dir\") pod \"csi-hostpathplugin-9shrr\" (UID: \"12e696fe-360e-4ea7-93e6-5ca2d0fc8123\") " pod="hostpath-provisioner/csi-hostpathplugin-9shrr" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.873974 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/f06be1d1-4134-4ad7-9233-b4099687bfe2-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-7xsg2\" (UID: \"f06be1d1-4134-4ad7-9233-b4099687bfe2\") " pod="openshift-marketplace/marketplace-operator-79b997595-7xsg2" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.873998 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p8pxk\" (UniqueName: \"kubernetes.io/projected/9caa5ea4-1245-4dc7-af31-571f68cec263-kube-api-access-p8pxk\") pod \"kube-storage-version-migrator-operator-b67b599dd-8rjs9\" (UID: \"9caa5ea4-1245-4dc7-af31-571f68cec263\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8rjs9" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.874025 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wgkrw\" (UniqueName: \"kubernetes.io/projected/593039fd-1af9-41c5-8f1a-bce2d698da75-kube-api-access-wgkrw\") pod \"dns-default-964q6\" (UID: \"593039fd-1af9-41c5-8f1a-bce2d698da75\") " pod="openshift-dns/dns-default-964q6" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.874048 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a9e3ba79-4e68-41af-823a-be5c4494df86-cert\") pod \"ingress-canary-lxx9x\" (UID: \"a9e3ba79-4e68-41af-823a-be5c4494df86\") " pod="openshift-ingress-canary/ingress-canary-lxx9x" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.874068 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/44e9e245-6889-4706-bc76-a22f2e7f6019-stats-auth\") pod \"router-default-5444994796-jxzst\" (UID: \"44e9e245-6889-4706-bc76-a22f2e7f6019\") " pod="openshift-ingress/router-default-5444994796-jxzst" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.874098 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zszpm\" (UniqueName: \"kubernetes.io/projected/f06be1d1-4134-4ad7-9233-b4099687bfe2-kube-api-access-zszpm\") pod \"marketplace-operator-79b997595-7xsg2\" (UID: \"f06be1d1-4134-4ad7-9233-b4099687bfe2\") " pod="openshift-marketplace/marketplace-operator-79b997595-7xsg2" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.874131 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/be3e5760-2f70-423b-9dd8-b5eeb0482b80-tmpfs\") pod \"packageserver-d55dfcdfc-7lgch\" (UID: \"be3e5760-2f70-423b-9dd8-b5eeb0482b80\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7lgch" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.874150 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/593039fd-1af9-41c5-8f1a-bce2d698da75-metrics-tls\") pod \"dns-default-964q6\" (UID: \"593039fd-1af9-41c5-8f1a-bce2d698da75\") " pod="openshift-dns/dns-default-964q6" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.874172 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fa9a982c-684d-497d-b8eb-f921e00fe511-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-8647g\" (UID: \"fa9a982c-684d-497d-b8eb-f921e00fe511\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-8647g" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.874197 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4chnr\" (UniqueName: \"kubernetes.io/projected/432c543c-59d1-49c9-ba43-91be5e113975-kube-api-access-4chnr\") pod \"multus-admission-controller-857f4d67dd-7tnbd\" (UID: \"432c543c-59d1-49c9-ba43-91be5e113975\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-7tnbd" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.874220 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/12e696fe-360e-4ea7-93e6-5ca2d0fc8123-socket-dir\") pod \"csi-hostpathplugin-9shrr\" (UID: \"12e696fe-360e-4ea7-93e6-5ca2d0fc8123\") " pod="hostpath-provisioner/csi-hostpathplugin-9shrr" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.874242 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/5de8942d-833c-499b-842a-ddb2eeeab601-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-95dbg\" (UID: \"5de8942d-833c-499b-842a-ddb2eeeab601\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-95dbg" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.874275 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f5fabb97-e818-4c00-8023-db5e8477878f-secret-volume\") pod \"collect-profiles-29490765-w8zqq\" (UID: \"f5fabb97-e818-4c00-8023-db5e8477878f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29490765-w8zqq" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.874306 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-47pnn\" (UID: \"d098371a-9920-44d2-b918-f21c8142fac2\") " pod="openshift-image-registry/image-registry-697d97f7c8-47pnn" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.874324 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/be3e5760-2f70-423b-9dd8-b5eeb0482b80-apiservice-cert\") pod \"packageserver-d55dfcdfc-7lgch\" (UID: \"be3e5760-2f70-423b-9dd8-b5eeb0482b80\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7lgch" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.874353 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/a411958e-77f0-45bb-a11c-abd88d756dee-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-w7zlr\" (UID: \"a411958e-77f0-45bb-a11c-abd88d756dee\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-w7zlr" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.874379 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h7gfl\" (UniqueName: \"kubernetes.io/projected/713fa157-05f9-4c30-b7e2-47a2b47cf7c4-kube-api-access-h7gfl\") pod \"migrator-59844c95c7-k5sf5\" (UID: \"713fa157-05f9-4c30-b7e2-47a2b47cf7c4\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-k5sf5" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.874397 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/12e696fe-360e-4ea7-93e6-5ca2d0fc8123-plugins-dir\") pod \"csi-hostpathplugin-9shrr\" (UID: \"12e696fe-360e-4ea7-93e6-5ca2d0fc8123\") " pod="hostpath-provisioner/csi-hostpathplugin-9shrr" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.874423 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3f0de367-82c0-4548-8f81-8c3482795003-config\") pod \"service-ca-operator-777779d784-vfldk\" (UID: \"3f0de367-82c0-4548-8f81-8c3482795003\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-vfldk" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.874459 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9caa5ea4-1245-4dc7-af31-571f68cec263-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-8rjs9\" (UID: \"9caa5ea4-1245-4dc7-af31-571f68cec263\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8rjs9" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.874494 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/4d0ec6e9-f4f7-4751-8317-60940c63f320-proxy-tls\") pod \"machine-config-controller-84d6567774-8rhs2\" (UID: \"4d0ec6e9-f4f7-4751-8317-60940c63f320\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8rhs2" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.874516 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/12e696fe-360e-4ea7-93e6-5ca2d0fc8123-mountpoint-dir\") pod \"csi-hostpathplugin-9shrr\" (UID: \"12e696fe-360e-4ea7-93e6-5ca2d0fc8123\") " pod="hostpath-provisioner/csi-hostpathplugin-9shrr" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.874537 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f5fabb97-e818-4c00-8023-db5e8477878f-config-volume\") pod \"collect-profiles-29490765-w8zqq\" (UID: \"f5fabb97-e818-4c00-8023-db5e8477878f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29490765-w8zqq" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.874565 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fa9a982c-684d-497d-b8eb-f921e00fe511-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-8647g\" (UID: \"fa9a982c-684d-497d-b8eb-f921e00fe511\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-8647g" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.874588 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wc27q\" (UniqueName: \"kubernetes.io/projected/12e696fe-360e-4ea7-93e6-5ca2d0fc8123-kube-api-access-wc27q\") pod \"csi-hostpathplugin-9shrr\" (UID: \"12e696fe-360e-4ea7-93e6-5ca2d0fc8123\") " pod="hostpath-provisioner/csi-hostpathplugin-9shrr" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.874626 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/50c60179-2a08-4055-afb0-78aff31d6465-profile-collector-cert\") pod \"catalog-operator-68c6474976-4px9m\" (UID: \"50c60179-2a08-4055-afb0-78aff31d6465\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-4px9m" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.874648 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/2644a03c-ada2-45c0-ac4e-edfd969a7752-srv-cert\") pod \"olm-operator-6b444d44fb-fbbxr\" (UID: \"2644a03c-ada2-45c0-ac4e-edfd969a7752\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-fbbxr" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.874691 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/44e9e245-6889-4706-bc76-a22f2e7f6019-metrics-certs\") pod \"router-default-5444994796-jxzst\" (UID: \"44e9e245-6889-4706-bc76-a22f2e7f6019\") " pod="openshift-ingress/router-default-5444994796-jxzst" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.874712 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/593039fd-1af9-41c5-8f1a-bce2d698da75-config-volume\") pod \"dns-default-964q6\" (UID: \"593039fd-1af9-41c5-8f1a-bce2d698da75\") " pod="openshift-dns/dns-default-964q6" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.874737 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/221280be-801c-4f1c-abbc-cb8a02f47838-config\") pod \"kube-apiserver-operator-766d6c64bb-m2ktn\" (UID: \"221280be-801c-4f1c-abbc-cb8a02f47838\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-m2ktn" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.874755 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/fa9a982c-684d-497d-b8eb-f921e00fe511-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-8647g\" (UID: \"fa9a982c-684d-497d-b8eb-f921e00fe511\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-8647g" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.874780 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z8tbc\" (UniqueName: \"kubernetes.io/projected/a9e3ba79-4e68-41af-823a-be5c4494df86-kube-api-access-z8tbc\") pod \"ingress-canary-lxx9x\" (UID: \"a9e3ba79-4e68-41af-823a-be5c4494df86\") " pod="openshift-ingress-canary/ingress-canary-lxx9x" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.874820 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/44e9e245-6889-4706-bc76-a22f2e7f6019-service-ca-bundle\") pod \"router-default-5444994796-jxzst\" (UID: \"44e9e245-6889-4706-bc76-a22f2e7f6019\") " pod="openshift-ingress/router-default-5444994796-jxzst" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.874841 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fmqgc\" (UniqueName: \"kubernetes.io/projected/4d0ec6e9-f4f7-4751-8317-60940c63f320-kube-api-access-fmqgc\") pod \"machine-config-controller-84d6567774-8rhs2\" (UID: \"4d0ec6e9-f4f7-4751-8317-60940c63f320\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8rhs2" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.874863 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/d0295627-eb92-49f1-9e39-82fd0d1ec864-certs\") pod \"machine-config-server-wp5r6\" (UID: \"d0295627-eb92-49f1-9e39-82fd0d1ec864\") " pod="openshift-machine-config-operator/machine-config-server-wp5r6" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.874880 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9caa5ea4-1245-4dc7-af31-571f68cec263-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-8rjs9\" (UID: \"9caa5ea4-1245-4dc7-af31-571f68cec263\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8rjs9" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.874897 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/e89d71d4-488f-4bed-9356-5dc92d5a8eae-signing-key\") pod \"service-ca-9c57cc56f-472pk\" (UID: \"e89d71d4-488f-4bed-9356-5dc92d5a8eae\") " pod="openshift-service-ca/service-ca-9c57cc56f-472pk" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.874919 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/e89d71d4-488f-4bed-9356-5dc92d5a8eae-signing-cabundle\") pod \"service-ca-9c57cc56f-472pk\" (UID: \"e89d71d4-488f-4bed-9356-5dc92d5a8eae\") " pod="openshift-service-ca/service-ca-9c57cc56f-472pk" Jan 26 16:49:29 crc kubenswrapper[4754]: E0126 16:49:29.874949 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 16:49:30.374932285 +0000 UTC m=+136.899112709 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-47pnn" (UID: "d098371a-9920-44d2-b918-f21c8142fac2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.874980 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vlhnl\" (UniqueName: \"kubernetes.io/projected/e89d71d4-488f-4bed-9356-5dc92d5a8eae-kube-api-access-vlhnl\") pod \"service-ca-9c57cc56f-472pk\" (UID: \"e89d71d4-488f-4bed-9356-5dc92d5a8eae\") " pod="openshift-service-ca/service-ca-9c57cc56f-472pk" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.875008 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/221280be-801c-4f1c-abbc-cb8a02f47838-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-m2ktn\" (UID: \"221280be-801c-4f1c-abbc-cb8a02f47838\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-m2ktn" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.875037 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qsqtw\" (UniqueName: \"kubernetes.io/projected/a411958e-77f0-45bb-a11c-abd88d756dee-kube-api-access-qsqtw\") pod \"control-plane-machine-set-operator-78cbb6b69f-w7zlr\" (UID: \"a411958e-77f0-45bb-a11c-abd88d756dee\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-w7zlr" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.875072 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/d0295627-eb92-49f1-9e39-82fd0d1ec864-node-bootstrap-token\") pod \"machine-config-server-wp5r6\" (UID: \"d0295627-eb92-49f1-9e39-82fd0d1ec864\") " pod="openshift-machine-config-operator/machine-config-server-wp5r6" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.875112 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bdd2421f-ccc8-4500-8fd2-23a5f58173df-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-2kmrb\" (UID: \"bdd2421f-ccc8-4500-8fd2-23a5f58173df\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-2kmrb" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.875139 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8nwnd\" (UniqueName: \"kubernetes.io/projected/2644a03c-ada2-45c0-ac4e-edfd969a7752-kube-api-access-8nwnd\") pod \"olm-operator-6b444d44fb-fbbxr\" (UID: \"2644a03c-ada2-45c0-ac4e-edfd969a7752\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-fbbxr" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.875473 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/4d0ec6e9-f4f7-4751-8317-60940c63f320-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-8rhs2\" (UID: \"4d0ec6e9-f4f7-4751-8317-60940c63f320\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8rhs2" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.875990 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/e89d71d4-488f-4bed-9356-5dc92d5a8eae-signing-cabundle\") pod \"service-ca-9c57cc56f-472pk\" (UID: \"e89d71d4-488f-4bed-9356-5dc92d5a8eae\") " pod="openshift-service-ca/service-ca-9c57cc56f-472pk" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.875997 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/12e696fe-360e-4ea7-93e6-5ca2d0fc8123-plugins-dir\") pod \"csi-hostpathplugin-9shrr\" (UID: \"12e696fe-360e-4ea7-93e6-5ca2d0fc8123\") " pod="hostpath-provisioner/csi-hostpathplugin-9shrr" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.876066 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/12e696fe-360e-4ea7-93e6-5ca2d0fc8123-mountpoint-dir\") pod \"csi-hostpathplugin-9shrr\" (UID: \"12e696fe-360e-4ea7-93e6-5ca2d0fc8123\") " pod="hostpath-provisioner/csi-hostpathplugin-9shrr" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.876110 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bhpmf\" (UniqueName: \"kubernetes.io/projected/07245d2d-bd41-4bca-9f4a-fab5f4a48eb9-kube-api-access-bhpmf\") pod \"console-operator-58897d9998-wjdgm\" (UID: \"07245d2d-bd41-4bca-9f4a-fab5f4a48eb9\") " pod="openshift-console-operator/console-operator-58897d9998-wjdgm" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.876337 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f06be1d1-4134-4ad7-9233-b4099687bfe2-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-7xsg2\" (UID: \"f06be1d1-4134-4ad7-9233-b4099687bfe2\") " pod="openshift-marketplace/marketplace-operator-79b997595-7xsg2" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.876732 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bdd2421f-ccc8-4500-8fd2-23a5f58173df-config\") pod \"kube-controller-manager-operator-78b949d7b-2kmrb\" (UID: \"bdd2421f-ccc8-4500-8fd2-23a5f58173df\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-2kmrb" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.877037 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3f0de367-82c0-4548-8f81-8c3482795003-serving-cert\") pod \"service-ca-operator-777779d784-vfldk\" (UID: \"3f0de367-82c0-4548-8f81-8c3482795003\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-vfldk" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.877373 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3f0de367-82c0-4548-8f81-8c3482795003-config\") pod \"service-ca-operator-777779d784-vfldk\" (UID: \"3f0de367-82c0-4548-8f81-8c3482795003\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-vfldk" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.877748 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/12e696fe-360e-4ea7-93e6-5ca2d0fc8123-csi-data-dir\") pod \"csi-hostpathplugin-9shrr\" (UID: \"12e696fe-360e-4ea7-93e6-5ca2d0fc8123\") " pod="hostpath-provisioner/csi-hostpathplugin-9shrr" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.878407 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/be3e5760-2f70-423b-9dd8-b5eeb0482b80-webhook-cert\") pod \"packageserver-d55dfcdfc-7lgch\" (UID: \"be3e5760-2f70-423b-9dd8-b5eeb0482b80\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7lgch" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.878455 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9caa5ea4-1245-4dc7-af31-571f68cec263-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-8rjs9\" (UID: \"9caa5ea4-1245-4dc7-af31-571f68cec263\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8rjs9" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.878776 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/432c543c-59d1-49c9-ba43-91be5e113975-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-7tnbd\" (UID: \"432c543c-59d1-49c9-ba43-91be5e113975\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-7tnbd" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.879080 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/be3e5760-2f70-423b-9dd8-b5eeb0482b80-apiservice-cert\") pod \"packageserver-d55dfcdfc-7lgch\" (UID: \"be3e5760-2f70-423b-9dd8-b5eeb0482b80\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7lgch" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.879342 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/44e9e245-6889-4706-bc76-a22f2e7f6019-service-ca-bundle\") pod \"router-default-5444994796-jxzst\" (UID: \"44e9e245-6889-4706-bc76-a22f2e7f6019\") " pod="openshift-ingress/router-default-5444994796-jxzst" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.879349 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/12e696fe-360e-4ea7-93e6-5ca2d0fc8123-socket-dir\") pod \"csi-hostpathplugin-9shrr\" (UID: \"12e696fe-360e-4ea7-93e6-5ca2d0fc8123\") " pod="hostpath-provisioner/csi-hostpathplugin-9shrr" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.879952 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/4d0ec6e9-f4f7-4751-8317-60940c63f320-proxy-tls\") pod \"machine-config-controller-84d6567774-8rhs2\" (UID: \"4d0ec6e9-f4f7-4751-8317-60940c63f320\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8rhs2" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.880075 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fa9a982c-684d-497d-b8eb-f921e00fe511-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-8647g\" (UID: \"fa9a982c-684d-497d-b8eb-f921e00fe511\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-8647g" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.880863 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/593039fd-1af9-41c5-8f1a-bce2d698da75-config-volume\") pod \"dns-default-964q6\" (UID: \"593039fd-1af9-41c5-8f1a-bce2d698da75\") " pod="openshift-dns/dns-default-964q6" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.880903 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a9e3ba79-4e68-41af-823a-be5c4494df86-cert\") pod \"ingress-canary-lxx9x\" (UID: \"a9e3ba79-4e68-41af-823a-be5c4494df86\") " pod="openshift-ingress-canary/ingress-canary-lxx9x" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.882279 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/f06be1d1-4134-4ad7-9233-b4099687bfe2-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-7xsg2\" (UID: \"f06be1d1-4134-4ad7-9233-b4099687bfe2\") " pod="openshift-marketplace/marketplace-operator-79b997595-7xsg2" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.882375 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/50c60179-2a08-4055-afb0-78aff31d6465-srv-cert\") pod \"catalog-operator-68c6474976-4px9m\" (UID: \"50c60179-2a08-4055-afb0-78aff31d6465\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-4px9m" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.882524 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/2644a03c-ada2-45c0-ac4e-edfd969a7752-profile-collector-cert\") pod \"olm-operator-6b444d44fb-fbbxr\" (UID: \"2644a03c-ada2-45c0-ac4e-edfd969a7752\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-fbbxr" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.882974 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/44e9e245-6889-4706-bc76-a22f2e7f6019-default-certificate\") pod \"router-default-5444994796-jxzst\" (UID: \"44e9e245-6889-4706-bc76-a22f2e7f6019\") " pod="openshift-ingress/router-default-5444994796-jxzst" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.883423 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/e89d71d4-488f-4bed-9356-5dc92d5a8eae-signing-key\") pod \"service-ca-9c57cc56f-472pk\" (UID: \"e89d71d4-488f-4bed-9356-5dc92d5a8eae\") " pod="openshift-service-ca/service-ca-9c57cc56f-472pk" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.883474 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/2644a03c-ada2-45c0-ac4e-edfd969a7752-srv-cert\") pod \"olm-operator-6b444d44fb-fbbxr\" (UID: \"2644a03c-ada2-45c0-ac4e-edfd969a7752\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-fbbxr" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.884780 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/593039fd-1af9-41c5-8f1a-bce2d698da75-metrics-tls\") pod \"dns-default-964q6\" (UID: \"593039fd-1af9-41c5-8f1a-bce2d698da75\") " pod="openshift-dns/dns-default-964q6" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.885101 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9caa5ea4-1245-4dc7-af31-571f68cec263-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-8rjs9\" (UID: \"9caa5ea4-1245-4dc7-af31-571f68cec263\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8rjs9" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.886267 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/44e9e245-6889-4706-bc76-a22f2e7f6019-stats-auth\") pod \"router-default-5444994796-jxzst\" (UID: \"44e9e245-6889-4706-bc76-a22f2e7f6019\") " pod="openshift-ingress/router-default-5444994796-jxzst" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.886712 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/d0295627-eb92-49f1-9e39-82fd0d1ec864-certs\") pod \"machine-config-server-wp5r6\" (UID: \"d0295627-eb92-49f1-9e39-82fd0d1ec864\") " pod="openshift-machine-config-operator/machine-config-server-wp5r6" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.887189 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fa9a982c-684d-497d-b8eb-f921e00fe511-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-8647g\" (UID: \"fa9a982c-684d-497d-b8eb-f921e00fe511\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-8647g" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.887329 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/50c60179-2a08-4055-afb0-78aff31d6465-profile-collector-cert\") pod \"catalog-operator-68c6474976-4px9m\" (UID: \"50c60179-2a08-4055-afb0-78aff31d6465\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-4px9m" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.887485 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/d0295627-eb92-49f1-9e39-82fd0d1ec864-node-bootstrap-token\") pod \"machine-config-server-wp5r6\" (UID: \"d0295627-eb92-49f1-9e39-82fd0d1ec864\") " pod="openshift-machine-config-operator/machine-config-server-wp5r6" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.887994 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/a411958e-77f0-45bb-a11c-abd88d756dee-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-w7zlr\" (UID: \"a411958e-77f0-45bb-a11c-abd88d756dee\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-w7zlr" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.888942 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bdd2421f-ccc8-4500-8fd2-23a5f58173df-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-2kmrb\" (UID: \"bdd2421f-ccc8-4500-8fd2-23a5f58173df\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-2kmrb" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.897357 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/44e9e245-6889-4706-bc76-a22f2e7f6019-metrics-certs\") pod \"router-default-5444994796-jxzst\" (UID: \"44e9e245-6889-4706-bc76-a22f2e7f6019\") " pod="openshift-ingress/router-default-5444994796-jxzst" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.898839 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-4hkb9"] Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.913634 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n7xpg\" (UniqueName: \"kubernetes.io/projected/693e0080-d2bc-427b-a81b-b09d1ce0e521-kube-api-access-n7xpg\") pod \"openshift-controller-manager-operator-756b6f6bc6-g4dbx\" (UID: \"693e0080-d2bc-427b-a81b-b09d1ce0e521\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-g4dbx" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.932313 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cgtt5\" (UniqueName: \"kubernetes.io/projected/50c60179-2a08-4055-afb0-78aff31d6465-kube-api-access-cgtt5\") pod \"catalog-operator-68c6474976-4px9m\" (UID: \"50c60179-2a08-4055-afb0-78aff31d6465\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-4px9m" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.951784 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bdd2421f-ccc8-4500-8fd2-23a5f58173df-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-2kmrb\" (UID: \"bdd2421f-ccc8-4500-8fd2-23a5f58173df\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-2kmrb" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.957649 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-2kmrb" Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.976641 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 16:49:29 crc kubenswrapper[4754]: E0126 16:49:29.976827 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 16:49:30.47679509 +0000 UTC m=+137.000975524 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.976953 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-47pnn\" (UID: \"d098371a-9920-44d2-b918-f21c8142fac2\") " pod="openshift-image-registry/image-registry-697d97f7c8-47pnn" Jan 26 16:49:29 crc kubenswrapper[4754]: E0126 16:49:29.977280 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 16:49:30.477267497 +0000 UTC m=+137.001447931 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-47pnn" (UID: "d098371a-9920-44d2-b918-f21c8142fac2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 16:49:29 crc kubenswrapper[4754]: I0126 16:49:29.992113 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bcsps\" (UniqueName: \"kubernetes.io/projected/3f0de367-82c0-4548-8f81-8c3482795003-kube-api-access-bcsps\") pod \"service-ca-operator-777779d784-vfldk\" (UID: \"3f0de367-82c0-4548-8f81-8c3482795003\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-vfldk" Jan 26 16:49:30 crc kubenswrapper[4754]: I0126 16:49:30.020980 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/221280be-801c-4f1c-abbc-cb8a02f47838-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-m2ktn\" (UID: \"221280be-801c-4f1c-abbc-cb8a02f47838\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-m2ktn" Jan 26 16:49:30 crc kubenswrapper[4754]: I0126 16:49:30.021439 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f5fabb97-e818-4c00-8023-db5e8477878f-config-volume\") pod \"collect-profiles-29490765-w8zqq\" (UID: \"f5fabb97-e818-4c00-8023-db5e8477878f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29490765-w8zqq" Jan 26 16:49:30 crc kubenswrapper[4754]: I0126 16:49:30.021519 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/221280be-801c-4f1c-abbc-cb8a02f47838-config\") pod \"kube-apiserver-operator-766d6c64bb-m2ktn\" (UID: \"221280be-801c-4f1c-abbc-cb8a02f47838\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-m2ktn" Jan 26 16:49:30 crc kubenswrapper[4754]: I0126 16:49:30.021591 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/be3e5760-2f70-423b-9dd8-b5eeb0482b80-tmpfs\") pod \"packageserver-d55dfcdfc-7lgch\" (UID: \"be3e5760-2f70-423b-9dd8-b5eeb0482b80\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7lgch" Jan 26 16:49:30 crc kubenswrapper[4754]: I0126 16:49:30.024485 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/5de8942d-833c-499b-842a-ddb2eeeab601-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-95dbg\" (UID: \"5de8942d-833c-499b-842a-ddb2eeeab601\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-95dbg" Jan 26 16:49:30 crc kubenswrapper[4754]: I0126 16:49:30.025374 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f5fabb97-e818-4c00-8023-db5e8477878f-secret-volume\") pod \"collect-profiles-29490765-w8zqq\" (UID: \"f5fabb97-e818-4c00-8023-db5e8477878f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29490765-w8zqq" Jan 26 16:49:30 crc kubenswrapper[4754]: I0126 16:49:30.026120 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gl9hl\" (UniqueName: \"kubernetes.io/projected/be3e5760-2f70-423b-9dd8-b5eeb0482b80-kube-api-access-gl9hl\") pod \"packageserver-d55dfcdfc-7lgch\" (UID: \"be3e5760-2f70-423b-9dd8-b5eeb0482b80\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7lgch" Jan 26 16:49:30 crc kubenswrapper[4754]: I0126 16:49:30.026789 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mt7kw\" (UniqueName: \"kubernetes.io/projected/f5fabb97-e818-4c00-8023-db5e8477878f-kube-api-access-mt7kw\") pod \"collect-profiles-29490765-w8zqq\" (UID: \"f5fabb97-e818-4c00-8023-db5e8477878f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29490765-w8zqq" Jan 26 16:49:30 crc kubenswrapper[4754]: I0126 16:49:30.028910 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-wjdgm" Jan 26 16:49:30 crc kubenswrapper[4754]: W0126 16:49:30.029366 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6b062c17_08ee_4fda_94d2_a59a3cdf8641.slice/crio-376f50ff150ec5b207ebb8d7914151c0b82bba9e689612140e97739a1c86000c WatchSource:0}: Error finding container 376f50ff150ec5b207ebb8d7914151c0b82bba9e689612140e97739a1c86000c: Status 404 returned error can't find the container with id 376f50ff150ec5b207ebb8d7914151c0b82bba9e689612140e97739a1c86000c Jan 26 16:49:30 crc kubenswrapper[4754]: I0126 16:49:30.034687 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h4gqm\" (UniqueName: \"kubernetes.io/projected/44e9e245-6889-4706-bc76-a22f2e7f6019-kube-api-access-h4gqm\") pod \"router-default-5444994796-jxzst\" (UID: \"44e9e245-6889-4706-bc76-a22f2e7f6019\") " pod="openshift-ingress/router-default-5444994796-jxzst" Jan 26 16:49:30 crc kubenswrapper[4754]: W0126 16:49:30.035815 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod528dc2c4_71cc_46cb_b847_426bb8638188.slice/crio-3f08ff855a9c086991567039090cb37f1bc644f627cc0ba3983e9ab1d9ca0f38 WatchSource:0}: Error finding container 3f08ff855a9c086991567039090cb37f1bc644f627cc0ba3983e9ab1d9ca0f38: Status 404 returned error can't find the container with id 3f08ff855a9c086991567039090cb37f1bc644f627cc0ba3983e9ab1d9ca0f38 Jan 26 16:49:30 crc kubenswrapper[4754]: W0126 16:49:30.039045 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc31a7288_dcad_4e03_beb6_22e7855a58cd.slice/crio-d4c5475f656776cd5c55f1b5eb90f94244397fd0a661659dcf4d91c33684110b WatchSource:0}: Error finding container d4c5475f656776cd5c55f1b5eb90f94244397fd0a661659dcf4d91c33684110b: Status 404 returned error can't find the container with id d4c5475f656776cd5c55f1b5eb90f94244397fd0a661659dcf4d91c33684110b Jan 26 16:49:30 crc kubenswrapper[4754]: W0126 16:49:30.040980 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod78242150_7a64_4766_808f_fa460cd84b20.slice/crio-194300c1e4947d37e4c01ae5a470449f42d1eb1dc19dd79d9ab2ce26a7c87ebc WatchSource:0}: Error finding container 194300c1e4947d37e4c01ae5a470449f42d1eb1dc19dd79d9ab2ce26a7c87ebc: Status 404 returned error can't find the container with id 194300c1e4947d37e4c01ae5a470449f42d1eb1dc19dd79d9ab2ce26a7c87ebc Jan 26 16:49:30 crc kubenswrapper[4754]: I0126 16:49:30.055475 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-4px9m" Jan 26 16:49:30 crc kubenswrapper[4754]: I0126 16:49:30.059212 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/221280be-801c-4f1c-abbc-cb8a02f47838-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-m2ktn\" (UID: \"221280be-801c-4f1c-abbc-cb8a02f47838\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-m2ktn" Jan 26 16:49:30 crc kubenswrapper[4754]: I0126 16:49:30.076709 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kmmqb\" (UniqueName: \"kubernetes.io/projected/d0295627-eb92-49f1-9e39-82fd0d1ec864-kube-api-access-kmmqb\") pod \"machine-config-server-wp5r6\" (UID: \"d0295627-eb92-49f1-9e39-82fd0d1ec864\") " pod="openshift-machine-config-operator/machine-config-server-wp5r6" Jan 26 16:49:30 crc kubenswrapper[4754]: I0126 16:49:30.079925 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 16:49:30 crc kubenswrapper[4754]: E0126 16:49:30.080121 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 16:49:30.580085747 +0000 UTC m=+137.104266181 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 16:49:30 crc kubenswrapper[4754]: I0126 16:49:30.080253 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-47pnn\" (UID: \"d098371a-9920-44d2-b918-f21c8142fac2\") " pod="openshift-image-registry/image-registry-697d97f7c8-47pnn" Jan 26 16:49:30 crc kubenswrapper[4754]: E0126 16:49:30.081057 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 16:49:30.581027311 +0000 UTC m=+137.105207745 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-47pnn" (UID: "d098371a-9920-44d2-b918-f21c8142fac2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 16:49:30 crc kubenswrapper[4754]: I0126 16:49:30.093572 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-wp5r6" Jan 26 16:49:30 crc kubenswrapper[4754]: I0126 16:49:30.095941 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8nwnd\" (UniqueName: \"kubernetes.io/projected/2644a03c-ada2-45c0-ac4e-edfd969a7752-kube-api-access-8nwnd\") pod \"olm-operator-6b444d44fb-fbbxr\" (UID: \"2644a03c-ada2-45c0-ac4e-edfd969a7752\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-fbbxr" Jan 26 16:49:30 crc kubenswrapper[4754]: I0126 16:49:30.101016 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-g4dbx" Jan 26 16:49:30 crc kubenswrapper[4754]: I0126 16:49:30.117211 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p8pxk\" (UniqueName: \"kubernetes.io/projected/9caa5ea4-1245-4dc7-af31-571f68cec263-kube-api-access-p8pxk\") pod \"kube-storage-version-migrator-operator-b67b599dd-8rjs9\" (UID: \"9caa5ea4-1245-4dc7-af31-571f68cec263\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8rjs9" Jan 26 16:49:30 crc kubenswrapper[4754]: I0126 16:49:30.137882 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wgkrw\" (UniqueName: \"kubernetes.io/projected/593039fd-1af9-41c5-8f1a-bce2d698da75-kube-api-access-wgkrw\") pod \"dns-default-964q6\" (UID: \"593039fd-1af9-41c5-8f1a-bce2d698da75\") " pod="openshift-dns/dns-default-964q6" Jan 26 16:49:30 crc kubenswrapper[4754]: I0126 16:49:30.159118 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h7gfl\" (UniqueName: \"kubernetes.io/projected/713fa157-05f9-4c30-b7e2-47a2b47cf7c4-kube-api-access-h7gfl\") pod \"migrator-59844c95c7-k5sf5\" (UID: \"713fa157-05f9-4c30-b7e2-47a2b47cf7c4\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-k5sf5" Jan 26 16:49:30 crc kubenswrapper[4754]: I0126 16:49:30.182339 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 16:49:30 crc kubenswrapper[4754]: E0126 16:49:30.182574 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 16:49:30.682553315 +0000 UTC m=+137.206733749 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 16:49:30 crc kubenswrapper[4754]: I0126 16:49:30.182896 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-47pnn\" (UID: \"d098371a-9920-44d2-b918-f21c8142fac2\") " pod="openshift-image-registry/image-registry-697d97f7c8-47pnn" Jan 26 16:49:30 crc kubenswrapper[4754]: E0126 16:49:30.183544 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 16:49:30.683490868 +0000 UTC m=+137.207671312 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-47pnn" (UID: "d098371a-9920-44d2-b918-f21c8142fac2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 16:49:30 crc kubenswrapper[4754]: I0126 16:49:30.196114 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7lgch" Jan 26 16:49:30 crc kubenswrapper[4754]: I0126 16:49:30.209569 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fmqgc\" (UniqueName: \"kubernetes.io/projected/4d0ec6e9-f4f7-4751-8317-60940c63f320-kube-api-access-fmqgc\") pod \"machine-config-controller-84d6567774-8rhs2\" (UID: \"4d0ec6e9-f4f7-4751-8317-60940c63f320\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8rhs2" Jan 26 16:49:30 crc kubenswrapper[4754]: I0126 16:49:30.211062 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gtt7p\" (UniqueName: \"kubernetes.io/projected/5de8942d-833c-499b-842a-ddb2eeeab601-kube-api-access-gtt7p\") pod \"package-server-manager-789f6589d5-95dbg\" (UID: \"5de8942d-833c-499b-842a-ddb2eeeab601\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-95dbg" Jan 26 16:49:30 crc kubenswrapper[4754]: I0126 16:49:30.222302 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-95dbg" Jan 26 16:49:30 crc kubenswrapper[4754]: I0126 16:49:30.223368 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/fa9a982c-684d-497d-b8eb-f921e00fe511-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-8647g\" (UID: \"fa9a982c-684d-497d-b8eb-f921e00fe511\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-8647g" Jan 26 16:49:30 crc kubenswrapper[4754]: I0126 16:49:30.232767 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29490765-w8zqq" Jan 26 16:49:30 crc kubenswrapper[4754]: I0126 16:49:30.241506 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-k5sf5" Jan 26 16:49:30 crc kubenswrapper[4754]: I0126 16:49:30.243031 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z8tbc\" (UniqueName: \"kubernetes.io/projected/a9e3ba79-4e68-41af-823a-be5c4494df86-kube-api-access-z8tbc\") pod \"ingress-canary-lxx9x\" (UID: \"a9e3ba79-4e68-41af-823a-be5c4494df86\") " pod="openshift-ingress-canary/ingress-canary-lxx9x" Jan 26 16:49:30 crc kubenswrapper[4754]: I0126 16:49:30.249368 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-m2ktn" Jan 26 16:49:30 crc kubenswrapper[4754]: I0126 16:49:30.262539 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vlhnl\" (UniqueName: \"kubernetes.io/projected/e89d71d4-488f-4bed-9356-5dc92d5a8eae-kube-api-access-vlhnl\") pod \"service-ca-9c57cc56f-472pk\" (UID: \"e89d71d4-488f-4bed-9356-5dc92d5a8eae\") " pod="openshift-service-ca/service-ca-9c57cc56f-472pk" Jan 26 16:49:30 crc kubenswrapper[4754]: I0126 16:49:30.265147 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-vfldk" Jan 26 16:49:30 crc kubenswrapper[4754]: I0126 16:49:30.279196 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8rjs9" Jan 26 16:49:30 crc kubenswrapper[4754]: I0126 16:49:30.280600 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8rhs2" Jan 26 16:49:30 crc kubenswrapper[4754]: I0126 16:49:30.290620 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zszpm\" (UniqueName: \"kubernetes.io/projected/f06be1d1-4134-4ad7-9233-b4099687bfe2-kube-api-access-zszpm\") pod \"marketplace-operator-79b997595-7xsg2\" (UID: \"f06be1d1-4134-4ad7-9233-b4099687bfe2\") " pod="openshift-marketplace/marketplace-operator-79b997595-7xsg2" Jan 26 16:49:30 crc kubenswrapper[4754]: I0126 16:49:30.291111 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 16:49:30 crc kubenswrapper[4754]: E0126 16:49:30.291520 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 16:49:30.791493413 +0000 UTC m=+137.315673847 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 16:49:30 crc kubenswrapper[4754]: I0126 16:49:30.291711 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-47pnn\" (UID: \"d098371a-9920-44d2-b918-f21c8142fac2\") " pod="openshift-image-registry/image-registry-697d97f7c8-47pnn" Jan 26 16:49:30 crc kubenswrapper[4754]: E0126 16:49:30.292142 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 16:49:30.792134177 +0000 UTC m=+137.316314611 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-47pnn" (UID: "d098371a-9920-44d2-b918-f21c8142fac2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 16:49:30 crc kubenswrapper[4754]: I0126 16:49:30.293083 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-jxzst" Jan 26 16:49:30 crc kubenswrapper[4754]: I0126 16:49:30.313335 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-mj6ns"] Jan 26 16:49:30 crc kubenswrapper[4754]: I0126 16:49:30.314793 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qsqtw\" (UniqueName: \"kubernetes.io/projected/a411958e-77f0-45bb-a11c-abd88d756dee-kube-api-access-qsqtw\") pod \"control-plane-machine-set-operator-78cbb6b69f-w7zlr\" (UID: \"a411958e-77f0-45bb-a11c-abd88d756dee\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-w7zlr" Jan 26 16:49:30 crc kubenswrapper[4754]: I0126 16:49:30.332342 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-472pk" Jan 26 16:49:30 crc kubenswrapper[4754]: I0126 16:49:30.338146 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-7xsg2" Jan 26 16:49:30 crc kubenswrapper[4754]: I0126 16:49:30.345731 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wc27q\" (UniqueName: \"kubernetes.io/projected/12e696fe-360e-4ea7-93e6-5ca2d0fc8123-kube-api-access-wc27q\") pod \"csi-hostpathplugin-9shrr\" (UID: \"12e696fe-360e-4ea7-93e6-5ca2d0fc8123\") " pod="hostpath-provisioner/csi-hostpathplugin-9shrr" Jan 26 16:49:30 crc kubenswrapper[4754]: I0126 16:49:30.349301 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-fbbxr" Jan 26 16:49:30 crc kubenswrapper[4754]: I0126 16:49:30.364412 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-w7zlr" Jan 26 16:49:30 crc kubenswrapper[4754]: I0126 16:49:30.366127 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4chnr\" (UniqueName: \"kubernetes.io/projected/432c543c-59d1-49c9-ba43-91be5e113975-kube-api-access-4chnr\") pod \"multus-admission-controller-857f4d67dd-7tnbd\" (UID: \"432c543c-59d1-49c9-ba43-91be5e113975\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-7tnbd" Jan 26 16:49:30 crc kubenswrapper[4754]: I0126 16:49:30.374450 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-964q6" Jan 26 16:49:30 crc kubenswrapper[4754]: I0126 16:49:30.384153 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-lxx9x" Jan 26 16:49:30 crc kubenswrapper[4754]: I0126 16:49:30.396994 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 16:49:30 crc kubenswrapper[4754]: E0126 16:49:30.397425 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 16:49:30.897406254 +0000 UTC m=+137.421586688 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 16:49:30 crc kubenswrapper[4754]: I0126 16:49:30.422524 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-9shrr" Jan 26 16:49:30 crc kubenswrapper[4754]: I0126 16:49:30.483442 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-m2525"] Jan 26 16:49:30 crc kubenswrapper[4754]: I0126 16:49:30.498468 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-47pnn\" (UID: \"d098371a-9920-44d2-b918-f21c8142fac2\") " pod="openshift-image-registry/image-registry-697d97f7c8-47pnn" Jan 26 16:49:30 crc kubenswrapper[4754]: E0126 16:49:30.498890 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 16:49:30.998876976 +0000 UTC m=+137.523057410 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-47pnn" (UID: "d098371a-9920-44d2-b918-f21c8142fac2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 16:49:30 crc kubenswrapper[4754]: I0126 16:49:30.503591 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-8647g" Jan 26 16:49:30 crc kubenswrapper[4754]: I0126 16:49:30.516017 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-7tnbd" Jan 26 16:49:30 crc kubenswrapper[4754]: I0126 16:49:30.581254 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-fj2tb" event={"ID":"6b062c17-08ee-4fda-94d2-a59a3cdf8641","Type":"ContainerStarted","Data":"376f50ff150ec5b207ebb8d7914151c0b82bba9e689612140e97739a1c86000c"} Jan 26 16:49:30 crc kubenswrapper[4754]: I0126 16:49:30.593479 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-677j9" event={"ID":"4d81d21a-83f5-402e-91fd-d8268a6b49be","Type":"ContainerStarted","Data":"9ba363f3729d6d97c45b7a749c56183deb835a07527b3f7da097b0eabf031c27"} Jan 26 16:49:30 crc kubenswrapper[4754]: I0126 16:49:30.593527 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-677j9" event={"ID":"4d81d21a-83f5-402e-91fd-d8268a6b49be","Type":"ContainerStarted","Data":"df3f65a2226610c65f3906b6fb76f4060198225d530bb9b11c214dfb40d1ed9f"} Jan 26 16:49:30 crc kubenswrapper[4754]: I0126 16:49:30.594410 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-677j9" Jan 26 16:49:30 crc kubenswrapper[4754]: I0126 16:49:30.605309 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 16:49:30 crc kubenswrapper[4754]: I0126 16:49:30.605939 4754 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-677j9 container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.11:8443/healthz\": dial tcp 10.217.0.11:8443: connect: connection refused" start-of-body= Jan 26 16:49:30 crc kubenswrapper[4754]: I0126 16:49:30.606007 4754 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-677j9" podUID="4d81d21a-83f5-402e-91fd-d8268a6b49be" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.11:8443/healthz\": dial tcp 10.217.0.11:8443: connect: connection refused" Jan 26 16:49:30 crc kubenswrapper[4754]: E0126 16:49:30.606396 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 16:49:31.106355793 +0000 UTC m=+137.630536237 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 16:49:30 crc kubenswrapper[4754]: I0126 16:49:30.609632 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-47pnn\" (UID: \"d098371a-9920-44d2-b918-f21c8142fac2\") " pod="openshift-image-registry/image-registry-697d97f7c8-47pnn" Jan 26 16:49:30 crc kubenswrapper[4754]: E0126 16:49:30.610325 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 16:49:31.110302774 +0000 UTC m=+137.634483208 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-47pnn" (UID: "d098371a-9920-44d2-b918-f21c8142fac2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 16:49:30 crc kubenswrapper[4754]: I0126 16:49:30.624898 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zvqdh" event={"ID":"adeda50d-627e-4071-a6ab-d7712750c22d","Type":"ContainerStarted","Data":"566d20448dc4862d4a5a5c5040c2f8aeeb5e2531565f0e9eec5fddd3022ab82c"} Jan 26 16:49:30 crc kubenswrapper[4754]: I0126 16:49:30.638160 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-2kmrb"] Jan 26 16:49:30 crc kubenswrapper[4754]: I0126 16:49:30.650084 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-mj6ns" event={"ID":"66592797-8f8e-4e3c-938b-aa6767928e3f","Type":"ContainerStarted","Data":"bb1e33b012a74ebbb597b88bc92bd4566daf64cccf44b97b889de8f3097f43de"} Jan 26 16:49:30 crc kubenswrapper[4754]: I0126 16:49:30.656588 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-5p6w2" event={"ID":"528dc2c4-71cc-46cb-b847-426bb8638188","Type":"ContainerStarted","Data":"3f08ff855a9c086991567039090cb37f1bc644f627cc0ba3983e9ab1d9ca0f38"} Jan 26 16:49:30 crc kubenswrapper[4754]: I0126 16:49:30.658702 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-4hkb9" event={"ID":"c31a7288-dcad-4e03-beb6-22e7855a58cd","Type":"ContainerStarted","Data":"d4c5475f656776cd5c55f1b5eb90f94244397fd0a661659dcf4d91c33684110b"} Jan 26 16:49:30 crc kubenswrapper[4754]: I0126 16:49:30.660361 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6mj2d" event={"ID":"bd64c275-cf75-4e42-8643-9178665801cb","Type":"ContainerStarted","Data":"2d80ac9c767728b639b6d4cacae13aae1ff235d122ecf741b359d67cdc107d94"} Jan 26 16:49:30 crc kubenswrapper[4754]: I0126 16:49:30.661937 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-8d467" event={"ID":"10609d5f-d6dd-40a5-acc4-63e62584d18e","Type":"ContainerStarted","Data":"832ebba41a7c68f635afe4ae7ff14cc3da025a4620cd4e90a572e66731b401a4"} Jan 26 16:49:30 crc kubenswrapper[4754]: I0126 16:49:30.664096 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-sz8hj" event={"ID":"aa57d7bf-eaea-4595-9d00-5d5d4b20f059","Type":"ContainerStarted","Data":"033b589914de1dc668275707d71ae2c8b28c3dc266a04bbec730ea0c8d1cee11"} Jan 26 16:49:30 crc kubenswrapper[4754]: I0126 16:49:30.681307 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-ptrdc"] Jan 26 16:49:30 crc kubenswrapper[4754]: I0126 16:49:30.694959 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-wp5r6" event={"ID":"d0295627-eb92-49f1-9e39-82fd0d1ec864","Type":"ContainerStarted","Data":"19af7af73f21e7a777d70436daf577293e082a4cfdc20dd559bd37d8cde3f86d"} Jan 26 16:49:30 crc kubenswrapper[4754]: I0126 16:49:30.700741 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-rcrn5" event={"ID":"78242150-7a64-4766-808f-fa460cd84b20","Type":"ContainerStarted","Data":"194300c1e4947d37e4c01ae5a470449f42d1eb1dc19dd79d9ab2ce26a7c87ebc"} Jan 26 16:49:30 crc kubenswrapper[4754]: I0126 16:49:30.711838 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-kh4hh" event={"ID":"06082490-21d0-4f81-8f02-b5ad80760b06","Type":"ContainerStarted","Data":"22e1cc2a46182da7fdcdf9a2113953ce59d54bccc0b1752d74d3e2dc1a62a166"} Jan 26 16:49:30 crc kubenswrapper[4754]: I0126 16:49:30.711913 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-kh4hh" event={"ID":"06082490-21d0-4f81-8f02-b5ad80760b06","Type":"ContainerStarted","Data":"7431e8c786ba61a64c687a2596e58d9ef6159f19f5babeb95b4e4bf37198cf21"} Jan 26 16:49:30 crc kubenswrapper[4754]: I0126 16:49:30.713236 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 16:49:30 crc kubenswrapper[4754]: E0126 16:49:30.714489 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 16:49:31.214451212 +0000 UTC m=+137.738631646 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 16:49:30 crc kubenswrapper[4754]: I0126 16:49:30.716086 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-47pnn\" (UID: \"d098371a-9920-44d2-b918-f21c8142fac2\") " pod="openshift-image-registry/image-registry-697d97f7c8-47pnn" Jan 26 16:49:30 crc kubenswrapper[4754]: I0126 16:49:30.716873 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-t4r22"] Jan 26 16:49:30 crc kubenswrapper[4754]: E0126 16:49:30.717098 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 16:49:31.217078466 +0000 UTC m=+137.741258890 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-47pnn" (UID: "d098371a-9920-44d2-b918-f21c8142fac2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 16:49:30 crc kubenswrapper[4754]: I0126 16:49:30.780508 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-g64xp"] Jan 26 16:49:30 crc kubenswrapper[4754]: I0126 16:49:30.818142 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-z2k6h"] Jan 26 16:49:30 crc kubenswrapper[4754]: I0126 16:49:30.818610 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 16:49:30 crc kubenswrapper[4754]: E0126 16:49:30.818868 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 16:49:31.318838308 +0000 UTC m=+137.843018742 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 16:49:30 crc kubenswrapper[4754]: I0126 16:49:30.819413 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-47pnn\" (UID: \"d098371a-9920-44d2-b918-f21c8142fac2\") " pod="openshift-image-registry/image-registry-697d97f7c8-47pnn" Jan 26 16:49:30 crc kubenswrapper[4754]: E0126 16:49:30.820978 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 16:49:31.320955584 +0000 UTC m=+137.845136018 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-47pnn" (UID: "d098371a-9920-44d2-b918-f21c8142fac2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 16:49:30 crc kubenswrapper[4754]: I0126 16:49:30.920092 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 16:49:30 crc kubenswrapper[4754]: E0126 16:49:30.922018 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 16:49:31.42197934 +0000 UTC m=+137.946159774 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 16:49:30 crc kubenswrapper[4754]: I0126 16:49:30.928689 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-47pnn\" (UID: \"d098371a-9920-44d2-b918-f21c8142fac2\") " pod="openshift-image-registry/image-registry-697d97f7c8-47pnn" Jan 26 16:49:30 crc kubenswrapper[4754]: E0126 16:49:30.929087 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 16:49:31.429069934 +0000 UTC m=+137.953250368 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-47pnn" (UID: "d098371a-9920-44d2-b918-f21c8142fac2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 16:49:30 crc kubenswrapper[4754]: I0126 16:49:30.947815 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-4px9m"] Jan 26 16:49:30 crc kubenswrapper[4754]: I0126 16:49:30.987488 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-wjdgm"] Jan 26 16:49:31 crc kubenswrapper[4754]: I0126 16:49:31.026758 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-kjc2g"] Jan 26 16:49:31 crc kubenswrapper[4754]: I0126 16:49:31.029403 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 16:49:31 crc kubenswrapper[4754]: E0126 16:49:31.029833 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 16:49:31.529813729 +0000 UTC m=+138.053994163 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 16:49:31 crc kubenswrapper[4754]: I0126 16:49:31.054446 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-g4dbx"] Jan 26 16:49:31 crc kubenswrapper[4754]: I0126 16:49:31.090268 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-k9sz7"] Jan 26 16:49:31 crc kubenswrapper[4754]: I0126 16:49:31.130921 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-47pnn\" (UID: \"d098371a-9920-44d2-b918-f21c8142fac2\") " pod="openshift-image-registry/image-registry-697d97f7c8-47pnn" Jan 26 16:49:31 crc kubenswrapper[4754]: E0126 16:49:31.131334 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 16:49:31.631318672 +0000 UTC m=+138.155499106 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-47pnn" (UID: "d098371a-9920-44d2-b918-f21c8142fac2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 16:49:31 crc kubenswrapper[4754]: I0126 16:49:31.161348 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-sz8hj" podStartSLOduration=118.161327559 podStartE2EDuration="1m58.161327559s" podCreationTimestamp="2026-01-26 16:47:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 16:49:31.11483086 +0000 UTC m=+137.639011294" watchObservedRunningTime="2026-01-26 16:49:31.161327559 +0000 UTC m=+137.685507993" Jan 26 16:49:31 crc kubenswrapper[4754]: I0126 16:49:31.232184 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 16:49:31 crc kubenswrapper[4754]: E0126 16:49:31.232604 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 16:49:31.732589326 +0000 UTC m=+138.256769760 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 16:49:31 crc kubenswrapper[4754]: I0126 16:49:31.334756 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-47pnn\" (UID: \"d098371a-9920-44d2-b918-f21c8142fac2\") " pod="openshift-image-registry/image-registry-697d97f7c8-47pnn" Jan 26 16:49:31 crc kubenswrapper[4754]: E0126 16:49:31.335207 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 16:49:31.835195048 +0000 UTC m=+138.359375472 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-47pnn" (UID: "d098371a-9920-44d2-b918-f21c8142fac2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 16:49:31 crc kubenswrapper[4754]: I0126 16:49:31.436595 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 16:49:31 crc kubenswrapper[4754]: E0126 16:49:31.436835 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 16:49:31.936795924 +0000 UTC m=+138.460976358 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 16:49:31 crc kubenswrapper[4754]: I0126 16:49:31.437107 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-47pnn\" (UID: \"d098371a-9920-44d2-b918-f21c8142fac2\") " pod="openshift-image-registry/image-registry-697d97f7c8-47pnn" Jan 26 16:49:31 crc kubenswrapper[4754]: E0126 16:49:31.437465 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 16:49:31.937449808 +0000 UTC m=+138.461630242 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-47pnn" (UID: "d098371a-9920-44d2-b918-f21c8142fac2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 16:49:31 crc kubenswrapper[4754]: I0126 16:49:31.541573 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 16:49:31 crc kubenswrapper[4754]: E0126 16:49:31.542080 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 16:49:32.042059972 +0000 UTC m=+138.566240406 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 16:49:31 crc kubenswrapper[4754]: I0126 16:49:31.654382 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-47pnn\" (UID: \"d098371a-9920-44d2-b918-f21c8142fac2\") " pod="openshift-image-registry/image-registry-697d97f7c8-47pnn" Jan 26 16:49:31 crc kubenswrapper[4754]: E0126 16:49:31.660090 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 16:49:32.160073197 +0000 UTC m=+138.684253631 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-47pnn" (UID: "d098371a-9920-44d2-b918-f21c8142fac2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 16:49:31 crc kubenswrapper[4754]: I0126 16:49:31.763764 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 16:49:31 crc kubenswrapper[4754]: E0126 16:49:31.764005 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 16:49:32.263961925 +0000 UTC m=+138.788142359 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 16:49:31 crc kubenswrapper[4754]: I0126 16:49:31.764328 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-47pnn\" (UID: \"d098371a-9920-44d2-b918-f21c8142fac2\") " pod="openshift-image-registry/image-registry-697d97f7c8-47pnn" Jan 26 16:49:31 crc kubenswrapper[4754]: E0126 16:49:31.764855 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 16:49:32.264848646 +0000 UTC m=+138.789029080 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-47pnn" (UID: "d098371a-9920-44d2-b918-f21c8142fac2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 16:49:31 crc kubenswrapper[4754]: I0126 16:49:31.834696 4754 patch_prober.go:28] interesting pod/downloads-7954f5f757-rcrn5 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Jan 26 16:49:31 crc kubenswrapper[4754]: I0126 16:49:31.835290 4754 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-rcrn5" podUID="78242150-7a64-4766-808f-fa460cd84b20" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Jan 26 16:49:31 crc kubenswrapper[4754]: I0126 16:49:31.838629 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zvqdh" podStartSLOduration=118.838614854 podStartE2EDuration="1m58.838614854s" podCreationTimestamp="2026-01-26 16:47:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 16:49:31.83821851 +0000 UTC m=+138.362398944" watchObservedRunningTime="2026-01-26 16:49:31.838614854 +0000 UTC m=+138.362795288" Jan 26 16:49:31 crc kubenswrapper[4754]: I0126 16:49:31.869062 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 16:49:31 crc kubenswrapper[4754]: E0126 16:49:31.875889 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 16:49:32.37585873 +0000 UTC m=+138.900039164 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 16:49:31 crc kubenswrapper[4754]: I0126 16:49:31.896826 4754 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-m2525 container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.9:8443/healthz\": dial tcp 10.217.0.9:8443: connect: connection refused" start-of-body= Jan 26 16:49:31 crc kubenswrapper[4754]: I0126 16:49:31.896920 4754 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m2525" podUID="587cde2e-d0a4-406b-ab72-26c7fe4b6707" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.9:8443/healthz\": dial tcp 10.217.0.9:8443: connect: connection refused" Jan 26 16:49:31 crc kubenswrapper[4754]: I0126 16:49:31.978082 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-677j9" podStartSLOduration=118.978060789 podStartE2EDuration="1m58.978060789s" podCreationTimestamp="2026-01-26 16:47:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 16:49:31.928451678 +0000 UTC m=+138.452632122" watchObservedRunningTime="2026-01-26 16:49:31.978060789 +0000 UTC m=+138.502241223" Jan 26 16:49:31 crc kubenswrapper[4754]: I0126 16:49:31.983694 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-47pnn\" (UID: \"d098371a-9920-44d2-b918-f21c8142fac2\") " pod="openshift-image-registry/image-registry-697d97f7c8-47pnn" Jan 26 16:49:31 crc kubenswrapper[4754]: E0126 16:49:31.985727 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 16:49:32.485711392 +0000 UTC m=+139.009891826 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-47pnn" (UID: "d098371a-9920-44d2-b918-f21c8142fac2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 16:49:32 crc kubenswrapper[4754]: I0126 16:49:32.028110 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m2525" podStartSLOduration=119.028072633 podStartE2EDuration="1m59.028072633s" podCreationTimestamp="2026-01-26 16:47:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 16:49:32.026617381 +0000 UTC m=+138.550797835" watchObservedRunningTime="2026-01-26 16:49:32.028072633 +0000 UTC m=+138.552253067" Jan 26 16:49:32 crc kubenswrapper[4754]: I0126 16:49:32.078472 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-rcrn5" podStartSLOduration=119.07844561 podStartE2EDuration="1m59.07844561s" podCreationTimestamp="2026-01-26 16:47:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 16:49:32.056887717 +0000 UTC m=+138.581068161" watchObservedRunningTime="2026-01-26 16:49:32.07844561 +0000 UTC m=+138.602626044" Jan 26 16:49:32 crc kubenswrapper[4754]: I0126 16:49:32.090793 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 16:49:32 crc kubenswrapper[4754]: E0126 16:49:32.091757 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 16:49:32.591702956 +0000 UTC m=+139.115883390 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 16:49:32 crc kubenswrapper[4754]: I0126 16:49:32.119681 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-wp5r6" podStartSLOduration=5.119645419 podStartE2EDuration="5.119645419s" podCreationTimestamp="2026-01-26 16:49:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 16:49:32.118910393 +0000 UTC m=+138.643090827" watchObservedRunningTime="2026-01-26 16:49:32.119645419 +0000 UTC m=+138.643825853" Jan 26 16:49:32 crc kubenswrapper[4754]: I0126 16:49:32.154117 4754 generic.go:334] "Generic (PLEG): container finished" podID="bd64c275-cf75-4e42-8643-9178665801cb" containerID="596c7eb6408bf683a4b0dba994b7a889b732f10a343a9422dfc10b6f5c4f70f0" exitCode=0 Jan 26 16:49:32 crc kubenswrapper[4754]: I0126 16:49:32.167467 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-4hkb9" podStartSLOduration=119.167450905 podStartE2EDuration="1m59.167450905s" podCreationTimestamp="2026-01-26 16:47:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 16:49:32.167376112 +0000 UTC m=+138.691556546" watchObservedRunningTime="2026-01-26 16:49:32.167450905 +0000 UTC m=+138.691631339" Jan 26 16:49:32 crc kubenswrapper[4754]: I0126 16:49:32.173321 4754 generic.go:334] "Generic (PLEG): container finished" podID="10609d5f-d6dd-40a5-acc4-63e62584d18e" containerID="c5c960012e7121969152474c0caab666c09613457c09af1cd618717c7e85341a" exitCode=0 Jan 26 16:49:32 crc kubenswrapper[4754]: I0126 16:49:32.192800 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-47pnn\" (UID: \"d098371a-9920-44d2-b918-f21c8142fac2\") " pod="openshift-image-registry/image-registry-697d97f7c8-47pnn" Jan 26 16:49:32 crc kubenswrapper[4754]: E0126 16:49:32.193742 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 16:49:32.69352939 +0000 UTC m=+139.217709824 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-47pnn" (UID: "d098371a-9920-44d2-b918-f21c8142fac2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 16:49:32 crc kubenswrapper[4754]: I0126 16:49:32.250461 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-rcrn5" event={"ID":"78242150-7a64-4766-808f-fa460cd84b20","Type":"ContainerStarted","Data":"119db5bab480b9bf962a3bc2027f9b42309188203141622fbf10435b45ec1eee"} Jan 26 16:49:32 crc kubenswrapper[4754]: I0126 16:49:32.250501 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-kjc2g" event={"ID":"f4d41d59-ebec-455b-9a17-2ff3f9a42f04","Type":"ContainerStarted","Data":"38344e5350aaaee87afeb61da92c81d58356e924be5976be9a4ef510575a85e1"} Jan 26 16:49:32 crc kubenswrapper[4754]: I0126 16:49:32.250520 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-rcrn5" Jan 26 16:49:32 crc kubenswrapper[4754]: I0126 16:49:32.250539 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m2525" Jan 26 16:49:32 crc kubenswrapper[4754]: I0126 16:49:32.250553 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m2525" event={"ID":"587cde2e-d0a4-406b-ab72-26c7fe4b6707","Type":"ContainerStarted","Data":"2f4c3587fd3d9e40fdd038a747a6ef5602c25a1ee3b35c91c551084566eed866"} Jan 26 16:49:32 crc kubenswrapper[4754]: I0126 16:49:32.250590 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-677j9" Jan 26 16:49:32 crc kubenswrapper[4754]: I0126 16:49:32.250602 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-4hkb9" event={"ID":"c31a7288-dcad-4e03-beb6-22e7855a58cd","Type":"ContainerStarted","Data":"48e2f6c067d666853d32daf9549980e55280c6c6be918965b2a07cfc3206e2fb"} Jan 26 16:49:32 crc kubenswrapper[4754]: I0126 16:49:32.250614 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-wp5r6" event={"ID":"d0295627-eb92-49f1-9e39-82fd0d1ec864","Type":"ContainerStarted","Data":"e02057ea772c59e779797cb230e1c230141e5687e5f49d522dd4995ec23ccda0"} Jan 26 16:49:32 crc kubenswrapper[4754]: I0126 16:49:32.250628 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-fj2tb" event={"ID":"6b062c17-08ee-4fda-94d2-a59a3cdf8641","Type":"ContainerStarted","Data":"bf78b74dbfa7d378c27639710c2f8f81981891aa1e4e775dc9e7e228603a5c5d"} Jan 26 16:49:32 crc kubenswrapper[4754]: I0126 16:49:32.250646 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-t4r22" event={"ID":"97dfbf03-7b95-4645-afa1-f080330f4cbe","Type":"ContainerStarted","Data":"d7f43a2de2abe6d4b8cbabef9ec4565e96d80b085b0c017f9310d900d0487e4e"} Jan 26 16:49:32 crc kubenswrapper[4754]: I0126 16:49:32.250696 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-t4r22" event={"ID":"97dfbf03-7b95-4645-afa1-f080330f4cbe","Type":"ContainerStarted","Data":"797e3d0af1f0c7ddb981ed463a58836e71b5ef29091bfcafeb268163c29ed330"} Jan 26 16:49:32 crc kubenswrapper[4754]: I0126 16:49:32.250716 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-mj6ns" event={"ID":"66592797-8f8e-4e3c-938b-aa6767928e3f","Type":"ContainerStarted","Data":"0e1f89e5009c15f541211d5605c51fbf9e968514a0af885d251acc6d2b5220db"} Jan 26 16:49:32 crc kubenswrapper[4754]: I0126 16:49:32.250728 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-z2k6h" event={"ID":"99582129-e50d-49c0-a554-aa703fc86452","Type":"ContainerStarted","Data":"5280d34689c16b1baafca1d5759417b0a53a3889077352dc89d41a00bc1e99db"} Jan 26 16:49:32 crc kubenswrapper[4754]: I0126 16:49:32.250739 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-ptrdc" event={"ID":"39a0ee2d-0f5b-4231-94a3-d11d517d3517","Type":"ContainerStarted","Data":"6d5f24bafc1219e0edeb57bc4721a6f38a9ee2d7ce2c270245f23f6f56c89e3c"} Jan 26 16:49:32 crc kubenswrapper[4754]: I0126 16:49:32.250750 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-k9sz7" event={"ID":"2a28f341-5f93-4d6c-bc4b-973d33dfeee1","Type":"ContainerStarted","Data":"6547b7c8ba1bb7c76eff300312d4989db2181deca6d0c6b53251d128f9854ded"} Jan 26 16:49:32 crc kubenswrapper[4754]: I0126 16:49:32.250762 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-g64xp" event={"ID":"5864769c-0a13-4562-a407-f67e4348e1a6","Type":"ContainerStarted","Data":"adee967a9fdaa9550f96a544668ace2eef9434adf283a5ad0ce3da2fa76a9011"} Jan 26 16:49:32 crc kubenswrapper[4754]: I0126 16:49:32.250775 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-5p6w2" event={"ID":"528dc2c4-71cc-46cb-b847-426bb8638188","Type":"ContainerStarted","Data":"283c7af4a93330b2596db693edb773b68c3ccab0a3c72eaf457b334081943e8f"} Jan 26 16:49:32 crc kubenswrapper[4754]: I0126 16:49:32.250789 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-g4dbx" event={"ID":"693e0080-d2bc-427b-a81b-b09d1ce0e521","Type":"ContainerStarted","Data":"bc3fa51b9398e0fc9ca8ec2adf12e5ca14ffc9571a15d9ff52800b7d8181f297"} Jan 26 16:49:32 crc kubenswrapper[4754]: I0126 16:49:32.250802 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6mj2d" event={"ID":"bd64c275-cf75-4e42-8643-9178665801cb","Type":"ContainerDied","Data":"596c7eb6408bf683a4b0dba994b7a889b732f10a343a9422dfc10b6f5c4f70f0"} Jan 26 16:49:32 crc kubenswrapper[4754]: I0126 16:49:32.250815 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-4px9m" event={"ID":"50c60179-2a08-4055-afb0-78aff31d6465","Type":"ContainerStarted","Data":"fc74612cfb67158d9cc06ee0f514d4a380ca1c544a845e7ba243eb388f0c1e68"} Jan 26 16:49:32 crc kubenswrapper[4754]: I0126 16:49:32.250826 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-2kmrb" event={"ID":"bdd2421f-ccc8-4500-8fd2-23a5f58173df","Type":"ContainerStarted","Data":"ddc04f4bd127d269d37e17442e3b8af7d67093437f4c51ea009e7a39a45fb860"} Jan 26 16:49:32 crc kubenswrapper[4754]: I0126 16:49:32.250839 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-jxzst" event={"ID":"44e9e245-6889-4706-bc76-a22f2e7f6019","Type":"ContainerStarted","Data":"10ff60bd2d2e37763e4ea3498f07c2b0ccfbf063d1afa44031a16cd2f32c0f01"} Jan 26 16:49:32 crc kubenswrapper[4754]: I0126 16:49:32.250852 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-8d467" event={"ID":"10609d5f-d6dd-40a5-acc4-63e62584d18e","Type":"ContainerDied","Data":"c5c960012e7121969152474c0caab666c09613457c09af1cd618717c7e85341a"} Jan 26 16:49:32 crc kubenswrapper[4754]: I0126 16:49:32.250866 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-wjdgm" event={"ID":"07245d2d-bd41-4bca-9f4a-fab5f4a48eb9","Type":"ContainerStarted","Data":"cfb9ea7bf626e0ee158e4fb89f221fd26dff7d295b5470be3fff8f7b63d17cd7"} Jan 26 16:49:32 crc kubenswrapper[4754]: I0126 16:49:32.290208 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-5p6w2" podStartSLOduration=119.290183069 podStartE2EDuration="1m59.290183069s" podCreationTimestamp="2026-01-26 16:47:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 16:49:32.288489588 +0000 UTC m=+138.812670022" watchObservedRunningTime="2026-01-26 16:49:32.290183069 +0000 UTC m=+138.814363503" Jan 26 16:49:32 crc kubenswrapper[4754]: I0126 16:49:32.295143 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 16:49:32 crc kubenswrapper[4754]: I0126 16:49:32.296201 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-jxzst" Jan 26 16:49:32 crc kubenswrapper[4754]: E0126 16:49:32.299444 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 16:49:32.79942057 +0000 UTC m=+139.323601004 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 16:49:32 crc kubenswrapper[4754]: I0126 16:49:32.321060 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-jxzst" podStartSLOduration=119.321041697 podStartE2EDuration="1m59.321041697s" podCreationTimestamp="2026-01-26 16:47:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 16:49:32.319047215 +0000 UTC m=+138.843227679" watchObservedRunningTime="2026-01-26 16:49:32.321041697 +0000 UTC m=+138.845222131" Jan 26 16:49:32 crc kubenswrapper[4754]: I0126 16:49:32.403516 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-47pnn\" (UID: \"d098371a-9920-44d2-b918-f21c8142fac2\") " pod="openshift-image-registry/image-registry-697d97f7c8-47pnn" Jan 26 16:49:32 crc kubenswrapper[4754]: E0126 16:49:32.403959 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 16:49:32.903944422 +0000 UTC m=+139.428124856 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-47pnn" (UID: "d098371a-9920-44d2-b918-f21c8142fac2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 16:49:32 crc kubenswrapper[4754]: I0126 16:49:32.415249 4754 patch_prober.go:28] interesting pod/router-default-5444994796-jxzst container/router namespace/openshift-ingress: Startup probe status=failure output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" start-of-body= Jan 26 16:49:32 crc kubenswrapper[4754]: I0126 16:49:32.415311 4754 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-jxzst" podUID="44e9e245-6889-4706-bc76-a22f2e7f6019" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" Jan 26 16:49:32 crc kubenswrapper[4754]: I0126 16:49:32.509214 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 16:49:32 crc kubenswrapper[4754]: E0126 16:49:32.509436 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 16:49:33.009417986 +0000 UTC m=+139.533598420 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 16:49:32 crc kubenswrapper[4754]: I0126 16:49:32.509462 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-47pnn\" (UID: \"d098371a-9920-44d2-b918-f21c8142fac2\") " pod="openshift-image-registry/image-registry-697d97f7c8-47pnn" Jan 26 16:49:32 crc kubenswrapper[4754]: E0126 16:49:32.516043 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 16:49:33.015751223 +0000 UTC m=+139.539931677 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-47pnn" (UID: "d098371a-9920-44d2-b918-f21c8142fac2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 16:49:32 crc kubenswrapper[4754]: I0126 16:49:32.614006 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 16:49:32 crc kubenswrapper[4754]: E0126 16:49:32.614581 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 16:49:33.11455544 +0000 UTC m=+139.638735874 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 16:49:32 crc kubenswrapper[4754]: I0126 16:49:32.717364 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-47pnn\" (UID: \"d098371a-9920-44d2-b918-f21c8142fac2\") " pod="openshift-image-registry/image-registry-697d97f7c8-47pnn" Jan 26 16:49:32 crc kubenswrapper[4754]: E0126 16:49:32.728015 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 16:49:33.22798905 +0000 UTC m=+139.752169484 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-47pnn" (UID: "d098371a-9920-44d2-b918-f21c8142fac2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 16:49:32 crc kubenswrapper[4754]: I0126 16:49:32.749120 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-7xsg2"] Jan 26 16:49:32 crc kubenswrapper[4754]: I0126 16:49:32.760805 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-964q6"] Jan 26 16:49:32 crc kubenswrapper[4754]: I0126 16:49:32.819341 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 16:49:32 crc kubenswrapper[4754]: E0126 16:49:32.820004 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 16:49:33.319980181 +0000 UTC m=+139.844160615 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 16:49:32 crc kubenswrapper[4754]: I0126 16:49:32.849207 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-95dbg"] Jan 26 16:49:32 crc kubenswrapper[4754]: I0126 16:49:32.892316 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-lxx9x"] Jan 26 16:49:32 crc kubenswrapper[4754]: I0126 16:49:32.921058 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-47pnn\" (UID: \"d098371a-9920-44d2-b918-f21c8142fac2\") " pod="openshift-image-registry/image-registry-697d97f7c8-47pnn" Jan 26 16:49:32 crc kubenswrapper[4754]: E0126 16:49:32.925271 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 16:49:33.425243569 +0000 UTC m=+139.949424003 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-47pnn" (UID: "d098371a-9920-44d2-b918-f21c8142fac2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 16:49:32 crc kubenswrapper[4754]: W0126 16:49:32.942908 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda9e3ba79_4e68_41af_823a_be5c4494df86.slice/crio-3d37ad7850a485ec87b749bafe96fb4592fcfe6ae4bc32b21fbfac1ce7648cbf WatchSource:0}: Error finding container 3d37ad7850a485ec87b749bafe96fb4592fcfe6ae4bc32b21fbfac1ce7648cbf: Status 404 returned error can't find the container with id 3d37ad7850a485ec87b749bafe96fb4592fcfe6ae4bc32b21fbfac1ce7648cbf Jan 26 16:49:32 crc kubenswrapper[4754]: I0126 16:49:32.967973 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29490765-w8zqq"] Jan 26 16:49:33 crc kubenswrapper[4754]: I0126 16:49:33.022534 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 16:49:33 crc kubenswrapper[4754]: E0126 16:49:33.024091 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 16:49:33.524036514 +0000 UTC m=+140.048216948 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 16:49:33 crc kubenswrapper[4754]: I0126 16:49:33.071152 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-fbbxr"] Jan 26 16:49:33 crc kubenswrapper[4754]: I0126 16:49:33.106749 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-vfldk"] Jan 26 16:49:33 crc kubenswrapper[4754]: I0126 16:49:33.109774 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-w7zlr"] Jan 26 16:49:33 crc kubenswrapper[4754]: I0126 16:49:33.126789 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-47pnn\" (UID: \"d098371a-9920-44d2-b918-f21c8142fac2\") " pod="openshift-image-registry/image-registry-697d97f7c8-47pnn" Jan 26 16:49:33 crc kubenswrapper[4754]: E0126 16:49:33.127273 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 16:49:33.627256869 +0000 UTC m=+140.151437303 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-47pnn" (UID: "d098371a-9920-44d2-b918-f21c8142fac2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 16:49:33 crc kubenswrapper[4754]: I0126 16:49:33.182057 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-m2ktn"] Jan 26 16:49:33 crc kubenswrapper[4754]: I0126 16:49:33.187241 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-472pk"] Jan 26 16:49:33 crc kubenswrapper[4754]: I0126 16:49:33.187256 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7lgch"] Jan 26 16:49:33 crc kubenswrapper[4754]: I0126 16:49:33.191172 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-k5sf5"] Jan 26 16:49:33 crc kubenswrapper[4754]: I0126 16:49:33.191210 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-8rhs2"] Jan 26 16:49:33 crc kubenswrapper[4754]: I0126 16:49:33.226742 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8rjs9"] Jan 26 16:49:33 crc kubenswrapper[4754]: I0126 16:49:33.231213 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 16:49:33 crc kubenswrapper[4754]: E0126 16:49:33.233177 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 16:49:33.733145809 +0000 UTC m=+140.257326243 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 16:49:33 crc kubenswrapper[4754]: I0126 16:49:33.237450 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-47pnn\" (UID: \"d098371a-9920-44d2-b918-f21c8142fac2\") " pod="openshift-image-registry/image-registry-697d97f7c8-47pnn" Jan 26 16:49:33 crc kubenswrapper[4754]: E0126 16:49:33.238143 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 16:49:33.738121667 +0000 UTC m=+140.262302111 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-47pnn" (UID: "d098371a-9920-44d2-b918-f21c8142fac2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 16:49:33 crc kubenswrapper[4754]: I0126 16:49:33.254062 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-wjdgm" event={"ID":"07245d2d-bd41-4bca-9f4a-fab5f4a48eb9","Type":"ContainerStarted","Data":"eaf46cb5dbba3e353e2c50cd2e60fe5ee3a4de86dce0bdf93be9296ad76da6e3"} Jan 26 16:49:33 crc kubenswrapper[4754]: I0126 16:49:33.255455 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-wjdgm" Jan 26 16:49:33 crc kubenswrapper[4754]: I0126 16:49:33.261744 4754 patch_prober.go:28] interesting pod/console-operator-58897d9998-wjdgm container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.32:8443/readyz\": dial tcp 10.217.0.32:8443: connect: connection refused" start-of-body= Jan 26 16:49:33 crc kubenswrapper[4754]: I0126 16:49:33.261836 4754 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-wjdgm" podUID="07245d2d-bd41-4bca-9f4a-fab5f4a48eb9" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.32:8443/readyz\": dial tcp 10.217.0.32:8443: connect: connection refused" Jan 26 16:49:33 crc kubenswrapper[4754]: W0126 16:49:33.263832 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3f0de367_82c0_4548_8f81_8c3482795003.slice/crio-1941dd9bbb8f03f3e8ac4ebd6fc671dc7e779011e854bb8db5affd9b76525a39 WatchSource:0}: Error finding container 1941dd9bbb8f03f3e8ac4ebd6fc671dc7e779011e854bb8db5affd9b76525a39: Status 404 returned error can't find the container with id 1941dd9bbb8f03f3e8ac4ebd6fc671dc7e779011e854bb8db5affd9b76525a39 Jan 26 16:49:33 crc kubenswrapper[4754]: I0126 16:49:33.264142 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-8647g"] Jan 26 16:49:33 crc kubenswrapper[4754]: W0126 16:49:33.268077 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod221280be_801c_4f1c_abbc_cb8a02f47838.slice/crio-c08bff35cf42d985a590c15f582b6447b0e88bb1690b0b83d04a40136110fc02 WatchSource:0}: Error finding container c08bff35cf42d985a590c15f582b6447b0e88bb1690b0b83d04a40136110fc02: Status 404 returned error can't find the container with id c08bff35cf42d985a590c15f582b6447b0e88bb1690b0b83d04a40136110fc02 Jan 26 16:49:33 crc kubenswrapper[4754]: I0126 16:49:33.272993 4754 generic.go:334] "Generic (PLEG): container finished" podID="66592797-8f8e-4e3c-938b-aa6767928e3f" containerID="0e1f89e5009c15f541211d5605c51fbf9e968514a0af885d251acc6d2b5220db" exitCode=0 Jan 26 16:49:33 crc kubenswrapper[4754]: I0126 16:49:33.273072 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-mj6ns" event={"ID":"66592797-8f8e-4e3c-938b-aa6767928e3f","Type":"ContainerDied","Data":"0e1f89e5009c15f541211d5605c51fbf9e968514a0af885d251acc6d2b5220db"} Jan 26 16:49:33 crc kubenswrapper[4754]: I0126 16:49:33.273105 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-mj6ns" event={"ID":"66592797-8f8e-4e3c-938b-aa6767928e3f","Type":"ContainerStarted","Data":"7a37c25689aab8950c97f446b4b5c98c134ded7205a40c12e49e7bcd8a5c089c"} Jan 26 16:49:33 crc kubenswrapper[4754]: I0126 16:49:33.273735 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-mj6ns" Jan 26 16:49:33 crc kubenswrapper[4754]: W0126 16:49:33.313097 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod713fa157_05f9_4c30_b7e2_47a2b47cf7c4.slice/crio-a83904b787cef4f892853cf5bf8697e3cd8a4e6b98f1df011ba5cdf56a9a1e73 WatchSource:0}: Error finding container a83904b787cef4f892853cf5bf8697e3cd8a4e6b98f1df011ba5cdf56a9a1e73: Status 404 returned error can't find the container with id a83904b787cef4f892853cf5bf8697e3cd8a4e6b98f1df011ba5cdf56a9a1e73 Jan 26 16:49:33 crc kubenswrapper[4754]: I0126 16:49:33.314053 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-4hkb9" event={"ID":"c31a7288-dcad-4e03-beb6-22e7855a58cd","Type":"ContainerStarted","Data":"99180d8dcf8f37550e52585e07f3666ffd4304a5bc4d0c2e4ce0e7f901d78eb1"} Jan 26 16:49:33 crc kubenswrapper[4754]: I0126 16:49:33.318454 4754 patch_prober.go:28] interesting pod/router-default-5444994796-jxzst container/router namespace/openshift-ingress: Startup probe status=failure output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" start-of-body= Jan 26 16:49:33 crc kubenswrapper[4754]: I0126 16:49:33.318492 4754 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-jxzst" podUID="44e9e245-6889-4706-bc76-a22f2e7f6019" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" Jan 26 16:49:33 crc kubenswrapper[4754]: I0126 16:49:33.322382 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29490765-w8zqq" event={"ID":"f5fabb97-e818-4c00-8023-db5e8477878f","Type":"ContainerStarted","Data":"a31fd3a98cc51d16587d5922ed23a21a41a4042ccf85c441b26d01ca1ee5c458"} Jan 26 16:49:33 crc kubenswrapper[4754]: I0126 16:49:33.326232 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-9shrr"] Jan 26 16:49:33 crc kubenswrapper[4754]: I0126 16:49:33.335374 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-wjdgm" podStartSLOduration=120.335353966 podStartE2EDuration="2m0.335353966s" podCreationTimestamp="2026-01-26 16:47:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 16:49:33.313849544 +0000 UTC m=+139.838029998" watchObservedRunningTime="2026-01-26 16:49:33.335353966 +0000 UTC m=+139.859534400" Jan 26 16:49:33 crc kubenswrapper[4754]: I0126 16:49:33.338347 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 16:49:33 crc kubenswrapper[4754]: E0126 16:49:33.339753 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 16:49:33.839731754 +0000 UTC m=+140.363912188 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 16:49:33 crc kubenswrapper[4754]: I0126 16:49:33.345845 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-jxzst" event={"ID":"44e9e245-6889-4706-bc76-a22f2e7f6019","Type":"ContainerStarted","Data":"e55d82b35a061c49ee68a084b1b23851470bdf041424ad58c170bb7177f1cc49"} Jan 26 16:49:33 crc kubenswrapper[4754]: W0126 16:49:33.360473 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9caa5ea4_1245_4dc7_af31_571f68cec263.slice/crio-a1698d010bebe35d2c5d019e1c493aef359ec4335e7c3249e1869bf5dd21b0d9 WatchSource:0}: Error finding container a1698d010bebe35d2c5d019e1c493aef359ec4335e7c3249e1869bf5dd21b0d9: Status 404 returned error can't find the container with id a1698d010bebe35d2c5d019e1c493aef359ec4335e7c3249e1869bf5dd21b0d9 Jan 26 16:49:33 crc kubenswrapper[4754]: I0126 16:49:33.375370 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-fbbxr" event={"ID":"2644a03c-ada2-45c0-ac4e-edfd969a7752","Type":"ContainerStarted","Data":"2ba0b396952389a67967fe46b5190241ac67f1a57bbf962ce1d9bf57c3cc7985"} Jan 26 16:49:33 crc kubenswrapper[4754]: I0126 16:49:33.385210 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-mj6ns" podStartSLOduration=120.385186355 podStartE2EDuration="2m0.385186355s" podCreationTimestamp="2026-01-26 16:47:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 16:49:33.349158502 +0000 UTC m=+139.873338936" watchObservedRunningTime="2026-01-26 16:49:33.385186355 +0000 UTC m=+139.909366779" Jan 26 16:49:33 crc kubenswrapper[4754]: I0126 16:49:33.385554 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-7tnbd"] Jan 26 16:49:33 crc kubenswrapper[4754]: I0126 16:49:33.411198 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-fj2tb" event={"ID":"6b062c17-08ee-4fda-94d2-a59a3cdf8641","Type":"ContainerStarted","Data":"f7c21bc070f62f9a5e4ba0e1e5281410514a76613725b875e5e1240d78369840"} Jan 26 16:49:33 crc kubenswrapper[4754]: I0126 16:49:33.414613 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-lxx9x" event={"ID":"a9e3ba79-4e68-41af-823a-be5c4494df86","Type":"ContainerStarted","Data":"3d37ad7850a485ec87b749bafe96fb4592fcfe6ae4bc32b21fbfac1ce7648cbf"} Jan 26 16:49:33 crc kubenswrapper[4754]: I0126 16:49:33.438005 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-ptrdc" event={"ID":"39a0ee2d-0f5b-4231-94a3-d11d517d3517","Type":"ContainerStarted","Data":"bb16cb8697ab376a529f78ef086d1f8e516ad0711d1d28ed6209eeb0c059e35c"} Jan 26 16:49:33 crc kubenswrapper[4754]: I0126 16:49:33.439755 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-47pnn\" (UID: \"d098371a-9920-44d2-b918-f21c8142fac2\") " pod="openshift-image-registry/image-registry-697d97f7c8-47pnn" Jan 26 16:49:33 crc kubenswrapper[4754]: E0126 16:49:33.440151 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 16:49:33.940136536 +0000 UTC m=+140.464316970 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-47pnn" (UID: "d098371a-9920-44d2-b918-f21c8142fac2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 16:49:33 crc kubenswrapper[4754]: I0126 16:49:33.443620 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-fj2tb" podStartSLOduration=120.443589211 podStartE2EDuration="2m0.443589211s" podCreationTimestamp="2026-01-26 16:47:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 16:49:33.44025928 +0000 UTC m=+139.964439734" watchObservedRunningTime="2026-01-26 16:49:33.443589211 +0000 UTC m=+139.967769645" Jan 26 16:49:33 crc kubenswrapper[4754]: I0126 16:49:33.487908 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-4px9m" event={"ID":"50c60179-2a08-4055-afb0-78aff31d6465","Type":"ContainerStarted","Data":"0148bae6006fc8cf9785fd33e0f9a7d8a1a57d35f3422ebc9a81f5cb9b43d143"} Jan 26 16:49:33 crc kubenswrapper[4754]: I0126 16:49:33.489439 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-4px9m" Jan 26 16:49:33 crc kubenswrapper[4754]: I0126 16:49:33.509326 4754 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-4px9m container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.37:8443/healthz\": dial tcp 10.217.0.37:8443: connect: connection refused" start-of-body= Jan 26 16:49:33 crc kubenswrapper[4754]: I0126 16:49:33.509384 4754 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-4px9m" podUID="50c60179-2a08-4055-afb0-78aff31d6465" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.37:8443/healthz\": dial tcp 10.217.0.37:8443: connect: connection refused" Jan 26 16:49:33 crc kubenswrapper[4754]: I0126 16:49:33.526494 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-4px9m" podStartSLOduration=120.526479575 podStartE2EDuration="2m0.526479575s" podCreationTimestamp="2026-01-26 16:47:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 16:49:33.526013738 +0000 UTC m=+140.050194172" watchObservedRunningTime="2026-01-26 16:49:33.526479575 +0000 UTC m=+140.050660009" Jan 26 16:49:33 crc kubenswrapper[4754]: I0126 16:49:33.545050 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 16:49:33 crc kubenswrapper[4754]: E0126 16:49:33.547269 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 16:49:34.04723698 +0000 UTC m=+140.571417424 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 16:49:33 crc kubenswrapper[4754]: I0126 16:49:33.567707 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-2kmrb" event={"ID":"bdd2421f-ccc8-4500-8fd2-23a5f58173df","Type":"ContainerStarted","Data":"b358db905d01a3ec4595284d48a6c2a4eceeddac3ac250a3f4d51a552d051224"} Jan 26 16:49:33 crc kubenswrapper[4754]: I0126 16:49:33.603424 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-2kmrb" podStartSLOduration=120.603401106 podStartE2EDuration="2m0.603401106s" podCreationTimestamp="2026-01-26 16:47:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 16:49:33.601056031 +0000 UTC m=+140.125236465" watchObservedRunningTime="2026-01-26 16:49:33.603401106 +0000 UTC m=+140.127581540" Jan 26 16:49:33 crc kubenswrapper[4754]: I0126 16:49:33.619297 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-kh4hh" event={"ID":"06082490-21d0-4f81-8f02-b5ad80760b06","Type":"ContainerStarted","Data":"c29411a659eace6731c6c160c938726e5177cfc8519d3ae86cd37da412dd49dc"} Jan 26 16:49:33 crc kubenswrapper[4754]: I0126 16:49:33.638872 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-kjc2g" event={"ID":"f4d41d59-ebec-455b-9a17-2ff3f9a42f04","Type":"ContainerStarted","Data":"85f004ef39a1fe5485cd1f9d4a073cda2cf9313eadaa5f4ecb32a9814c0d91c3"} Jan 26 16:49:33 crc kubenswrapper[4754]: I0126 16:49:33.644072 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m2525" event={"ID":"587cde2e-d0a4-406b-ab72-26c7fe4b6707","Type":"ContainerStarted","Data":"10faa4395f7c506f3b98bb9824f69a2aedc5484d3299d85f666cb4dbbab259c8"} Jan 26 16:49:33 crc kubenswrapper[4754]: I0126 16:49:33.648514 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-47pnn\" (UID: \"d098371a-9920-44d2-b918-f21c8142fac2\") " pod="openshift-image-registry/image-registry-697d97f7c8-47pnn" Jan 26 16:49:33 crc kubenswrapper[4754]: E0126 16:49:33.649754 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 16:49:34.149737098 +0000 UTC m=+140.673917532 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-47pnn" (UID: "d098371a-9920-44d2-b918-f21c8142fac2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 16:49:33 crc kubenswrapper[4754]: I0126 16:49:33.649978 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-kh4hh" podStartSLOduration=120.649960776 podStartE2EDuration="2m0.649960776s" podCreationTimestamp="2026-01-26 16:47:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 16:49:33.64782059 +0000 UTC m=+140.172001054" watchObservedRunningTime="2026-01-26 16:49:33.649960776 +0000 UTC m=+140.174141210" Jan 26 16:49:33 crc kubenswrapper[4754]: I0126 16:49:33.650839 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-g4dbx" event={"ID":"693e0080-d2bc-427b-a81b-b09d1ce0e521","Type":"ContainerStarted","Data":"d3279a48cb843fd1cdb0e6cf506fb8e7105dc50fcfbf5cc0caf0d7665d3c15b1"} Jan 26 16:49:33 crc kubenswrapper[4754]: I0126 16:49:33.652832 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-964q6" event={"ID":"593039fd-1af9-41c5-8f1a-bce2d698da75","Type":"ContainerStarted","Data":"7df46f6a8aca91a7a148377dbb29ef5d8f3b5506f12a5ee1661a809c48229f2f"} Jan 26 16:49:33 crc kubenswrapper[4754]: I0126 16:49:33.682870 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-g64xp" event={"ID":"5864769c-0a13-4562-a407-f67e4348e1a6","Type":"ContainerStarted","Data":"4afbb9a26eb5e64c3641316a06f96bc71a34191d8e517ed36dda21e61958c8bc"} Jan 26 16:49:33 crc kubenswrapper[4754]: I0126 16:49:33.682926 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-g64xp" event={"ID":"5864769c-0a13-4562-a407-f67e4348e1a6","Type":"ContainerStarted","Data":"ecf5482c46bddb15d82dc3bc28f5f86f79ea6ecb49d7366e776667d4c97616de"} Jan 26 16:49:33 crc kubenswrapper[4754]: I0126 16:49:33.685271 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-95dbg" event={"ID":"5de8942d-833c-499b-842a-ddb2eeeab601","Type":"ContainerStarted","Data":"5fc26b630ece13fe740e38b9dd397af1a123f15164f856ba78ea6da2d4ab2e16"} Jan 26 16:49:33 crc kubenswrapper[4754]: I0126 16:49:33.720061 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-7xsg2" event={"ID":"f06be1d1-4134-4ad7-9233-b4099687bfe2","Type":"ContainerStarted","Data":"8450fa38eda3b5cba6c4dda738ea2aa9cad3994786cefe2ac842546f0b4269f0"} Jan 26 16:49:33 crc kubenswrapper[4754]: I0126 16:49:33.721137 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-7xsg2" Jan 26 16:49:33 crc kubenswrapper[4754]: I0126 16:49:33.727566 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-z2k6h" event={"ID":"99582129-e50d-49c0-a554-aa703fc86452","Type":"ContainerStarted","Data":"ca1d6124bc57a74df00c338b1bcc83f042d5a528df7398ba562a5fd0603436c2"} Jan 26 16:49:33 crc kubenswrapper[4754]: I0126 16:49:33.730169 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-k9sz7" event={"ID":"2a28f341-5f93-4d6c-bc4b-973d33dfeee1","Type":"ContainerStarted","Data":"a7b1a2116140f01645d1d09ad047f6edc2f84fbb9c15315f990b891c1b623394"} Jan 26 16:49:33 crc kubenswrapper[4754]: I0126 16:49:33.731074 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-k9sz7" Jan 26 16:49:33 crc kubenswrapper[4754]: I0126 16:49:33.738245 4754 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-k9sz7 container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.13:6443/healthz\": dial tcp 10.217.0.13:6443: connect: connection refused" start-of-body= Jan 26 16:49:33 crc kubenswrapper[4754]: I0126 16:49:33.738300 4754 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-k9sz7" podUID="2a28f341-5f93-4d6c-bc4b-973d33dfeee1" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.13:6443/healthz\": dial tcp 10.217.0.13:6443: connect: connection refused" Jan 26 16:49:33 crc kubenswrapper[4754]: I0126 16:49:33.752533 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 16:49:33 crc kubenswrapper[4754]: E0126 16:49:33.754108 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 16:49:34.254055231 +0000 UTC m=+140.778235685 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 16:49:33 crc kubenswrapper[4754]: I0126 16:49:33.756382 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-t4r22" event={"ID":"97dfbf03-7b95-4645-afa1-f080330f4cbe","Type":"ContainerStarted","Data":"057d6c9700ed71366dc10debf3aa85c663a1911880ead2c93f9ac10fe13e0584"} Jan 26 16:49:33 crc kubenswrapper[4754]: I0126 16:49:33.758798 4754 patch_prober.go:28] interesting pod/downloads-7954f5f757-rcrn5 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Jan 26 16:49:33 crc kubenswrapper[4754]: I0126 16:49:33.758867 4754 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-rcrn5" podUID="78242150-7a64-4766-808f-fa460cd84b20" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Jan 26 16:49:33 crc kubenswrapper[4754]: I0126 16:49:33.764069 4754 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-7xsg2 container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.36:8080/healthz\": dial tcp 10.217.0.36:8080: connect: connection refused" start-of-body= Jan 26 16:49:33 crc kubenswrapper[4754]: I0126 16:49:33.764119 4754 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-7xsg2" podUID="f06be1d1-4134-4ad7-9233-b4099687bfe2" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.36:8080/healthz\": dial tcp 10.217.0.36:8080: connect: connection refused" Jan 26 16:49:33 crc kubenswrapper[4754]: I0126 16:49:33.777518 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-g64xp" podStartSLOduration=120.777476813 podStartE2EDuration="2m0.777476813s" podCreationTimestamp="2026-01-26 16:47:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 16:49:33.768976948 +0000 UTC m=+140.293157392" watchObservedRunningTime="2026-01-26 16:49:33.777476813 +0000 UTC m=+140.301657247" Jan 26 16:49:33 crc kubenswrapper[4754]: I0126 16:49:33.779183 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-kjc2g" podStartSLOduration=120.779077229 podStartE2EDuration="2m0.779077229s" podCreationTimestamp="2026-01-26 16:47:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 16:49:33.707903916 +0000 UTC m=+140.232084450" watchObservedRunningTime="2026-01-26 16:49:33.779077229 +0000 UTC m=+140.303257683" Jan 26 16:49:33 crc kubenswrapper[4754]: I0126 16:49:33.824010 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-g4dbx" podStartSLOduration=120.82397225 podStartE2EDuration="2m0.82397225s" podCreationTimestamp="2026-01-26 16:47:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 16:49:33.821886716 +0000 UTC m=+140.346067170" watchObservedRunningTime="2026-01-26 16:49:33.82397225 +0000 UTC m=+140.348152684" Jan 26 16:49:33 crc kubenswrapper[4754]: I0126 16:49:33.855992 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-47pnn\" (UID: \"d098371a-9920-44d2-b918-f21c8142fac2\") " pod="openshift-image-registry/image-registry-697d97f7c8-47pnn" Jan 26 16:49:33 crc kubenswrapper[4754]: E0126 16:49:33.861782 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 16:49:34.361738556 +0000 UTC m=+140.885918990 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-47pnn" (UID: "d098371a-9920-44d2-b918-f21c8142fac2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 16:49:33 crc kubenswrapper[4754]: I0126 16:49:33.876337 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-k9sz7" podStartSLOduration=120.876316789 podStartE2EDuration="2m0.876316789s" podCreationTimestamp="2026-01-26 16:47:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 16:49:33.873774168 +0000 UTC m=+140.397954622" watchObservedRunningTime="2026-01-26 16:49:33.876316789 +0000 UTC m=+140.400497243" Jan 26 16:49:33 crc kubenswrapper[4754]: I0126 16:49:33.965482 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 16:49:33 crc kubenswrapper[4754]: E0126 16:49:33.966209 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 16:49:34.465994917 +0000 UTC m=+140.990175351 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 16:49:33 crc kubenswrapper[4754]: I0126 16:49:33.966390 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-47pnn\" (UID: \"d098371a-9920-44d2-b918-f21c8142fac2\") " pod="openshift-image-registry/image-registry-697d97f7c8-47pnn" Jan 26 16:49:33 crc kubenswrapper[4754]: E0126 16:49:33.967401 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 16:49:34.467391648 +0000 UTC m=+140.991572072 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-47pnn" (UID: "d098371a-9920-44d2-b918-f21c8142fac2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 16:49:34 crc kubenswrapper[4754]: I0126 16:49:34.076045 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 16:49:34 crc kubenswrapper[4754]: E0126 16:49:34.076324 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 16:49:34.576309417 +0000 UTC m=+141.100489851 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 16:49:34 crc kubenswrapper[4754]: I0126 16:49:34.178170 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-47pnn\" (UID: \"d098371a-9920-44d2-b918-f21c8142fac2\") " pod="openshift-image-registry/image-registry-697d97f7c8-47pnn" Jan 26 16:49:34 crc kubenswrapper[4754]: E0126 16:49:34.178589 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 16:49:34.678574946 +0000 UTC m=+141.202755380 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-47pnn" (UID: "d098371a-9920-44d2-b918-f21c8142fac2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 16:49:34 crc kubenswrapper[4754]: I0126 16:49:34.279353 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m2525" Jan 26 16:49:34 crc kubenswrapper[4754]: I0126 16:49:34.279804 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 16:49:34 crc kubenswrapper[4754]: E0126 16:49:34.280191 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 16:49:34.780175912 +0000 UTC m=+141.304356346 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 16:49:34 crc kubenswrapper[4754]: I0126 16:49:34.313047 4754 patch_prober.go:28] interesting pod/router-default-5444994796-jxzst container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 26 16:49:34 crc kubenswrapper[4754]: [-]has-synced failed: reason withheld Jan 26 16:49:34 crc kubenswrapper[4754]: [+]process-running ok Jan 26 16:49:34 crc kubenswrapper[4754]: healthz check failed Jan 26 16:49:34 crc kubenswrapper[4754]: I0126 16:49:34.313110 4754 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-jxzst" podUID="44e9e245-6889-4706-bc76-a22f2e7f6019" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 26 16:49:34 crc kubenswrapper[4754]: I0126 16:49:34.382683 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-47pnn\" (UID: \"d098371a-9920-44d2-b918-f21c8142fac2\") " pod="openshift-image-registry/image-registry-697d97f7c8-47pnn" Jan 26 16:49:34 crc kubenswrapper[4754]: E0126 16:49:34.383087 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 16:49:34.883071355 +0000 UTC m=+141.407251789 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-47pnn" (UID: "d098371a-9920-44d2-b918-f21c8142fac2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 16:49:34 crc kubenswrapper[4754]: I0126 16:49:34.486164 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 16:49:34 crc kubenswrapper[4754]: E0126 16:49:34.487470 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 16:49:34.98744078 +0000 UTC m=+141.511621214 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 16:49:34 crc kubenswrapper[4754]: I0126 16:49:34.598260 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-47pnn\" (UID: \"d098371a-9920-44d2-b918-f21c8142fac2\") " pod="openshift-image-registry/image-registry-697d97f7c8-47pnn" Jan 26 16:49:34 crc kubenswrapper[4754]: E0126 16:49:34.598669 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 16:49:35.098653341 +0000 UTC m=+141.622833775 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-47pnn" (UID: "d098371a-9920-44d2-b918-f21c8142fac2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 16:49:34 crc kubenswrapper[4754]: E0126 16:49:34.631779 4754 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9caa5ea4_1245_4dc7_af31_571f68cec263.slice/crio-conmon-76347e16e31561b0fe18412471d0e2973106191023da82c6009b70858c0dd6f8.scope\": RecentStats: unable to find data in memory cache]" Jan 26 16:49:34 crc kubenswrapper[4754]: I0126 16:49:34.700124 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 16:49:34 crc kubenswrapper[4754]: E0126 16:49:34.700361 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 16:49:35.20034552 +0000 UTC m=+141.724525954 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 16:49:34 crc kubenswrapper[4754]: I0126 16:49:34.734773 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-t4r22" podStartSLOduration=121.734758726 podStartE2EDuration="2m1.734758726s" podCreationTimestamp="2026-01-26 16:47:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 16:49:34.73405422 +0000 UTC m=+141.258234664" watchObservedRunningTime="2026-01-26 16:49:34.734758726 +0000 UTC m=+141.258939160" Jan 26 16:49:34 crc kubenswrapper[4754]: I0126 16:49:34.793138 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-964q6" event={"ID":"593039fd-1af9-41c5-8f1a-bce2d698da75","Type":"ContainerStarted","Data":"89977dcf003b8a218a7843feec3d6b1d4fc509f15e48a3f8875aeee8ba11522c"} Jan 26 16:49:34 crc kubenswrapper[4754]: I0126 16:49:34.805094 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-47pnn\" (UID: \"d098371a-9920-44d2-b918-f21c8142fac2\") " pod="openshift-image-registry/image-registry-697d97f7c8-47pnn" Jan 26 16:49:34 crc kubenswrapper[4754]: E0126 16:49:34.805392 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 16:49:35.30537826 +0000 UTC m=+141.829558684 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-47pnn" (UID: "d098371a-9920-44d2-b918-f21c8142fac2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 16:49:34 crc kubenswrapper[4754]: I0126 16:49:34.810315 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-95dbg" event={"ID":"5de8942d-833c-499b-842a-ddb2eeeab601","Type":"ContainerStarted","Data":"5be9e9379112dfb27f2feb84fd1c2872229838c6d28c8a4d28b378958485e8ab"} Jan 26 16:49:34 crc kubenswrapper[4754]: I0126 16:49:34.815959 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-lxx9x" event={"ID":"a9e3ba79-4e68-41af-823a-be5c4494df86","Type":"ContainerStarted","Data":"36a6a0d8ee54667c2f800e101fdec8b19570f9f2f60acbb541975c905090e94e"} Jan 26 16:49:34 crc kubenswrapper[4754]: I0126 16:49:34.828536 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-472pk" event={"ID":"e89d71d4-488f-4bed-9356-5dc92d5a8eae","Type":"ContainerStarted","Data":"a9d6fa4f6ed8364fe6a41b9b85d7a29188685b560ca74131749f63cb3f53e47e"} Jan 26 16:49:34 crc kubenswrapper[4754]: I0126 16:49:34.844987 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29490765-w8zqq" event={"ID":"f5fabb97-e818-4c00-8023-db5e8477878f","Type":"ContainerStarted","Data":"e4fb9df6906463bd5ee7e3667dc19c4ad04a1cf49b018f0036e4fbd87047d607"} Jan 26 16:49:34 crc kubenswrapper[4754]: I0126 16:49:34.904891 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-7xsg2" podStartSLOduration=121.904859729 podStartE2EDuration="2m1.904859729s" podCreationTimestamp="2026-01-26 16:47:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 16:49:34.808580805 +0000 UTC m=+141.332761249" watchObservedRunningTime="2026-01-26 16:49:34.904859729 +0000 UTC m=+141.429040163" Jan 26 16:49:34 crc kubenswrapper[4754]: I0126 16:49:34.910298 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 16:49:34 crc kubenswrapper[4754]: E0126 16:49:34.910679 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 16:49:35.410639967 +0000 UTC m=+141.934820441 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 16:49:34 crc kubenswrapper[4754]: I0126 16:49:34.937995 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-w7zlr" event={"ID":"a411958e-77f0-45bb-a11c-abd88d756dee","Type":"ContainerStarted","Data":"0d21f1b0bddf0da4242ea8ae5fa0b779e22d644d75e0b68fcc081752076ebaed"} Jan 26 16:49:34 crc kubenswrapper[4754]: I0126 16:49:34.938034 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-w7zlr" event={"ID":"a411958e-77f0-45bb-a11c-abd88d756dee","Type":"ContainerStarted","Data":"f36c8842c2b0767e20510e2c62021c4676bf31e7c64e805d32220b880359ef1c"} Jan 26 16:49:34 crc kubenswrapper[4754]: I0126 16:49:34.957447 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-9shrr" event={"ID":"12e696fe-360e-4ea7-93e6-5ca2d0fc8123","Type":"ContainerStarted","Data":"7f25d1a32effaa6dead7a3ef535540a442e399005989205e00f17d0b8dca7e15"} Jan 26 16:49:34 crc kubenswrapper[4754]: I0126 16:49:34.994148 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-7xsg2" event={"ID":"f06be1d1-4134-4ad7-9233-b4099687bfe2","Type":"ContainerStarted","Data":"089c7bd3b8cd79c825b9c58787fb66cd12628763659360ce5e1b716c6f8fa7d0"} Jan 26 16:49:35 crc kubenswrapper[4754]: I0126 16:49:35.010405 4754 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-7xsg2 container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.36:8080/healthz\": dial tcp 10.217.0.36:8080: connect: connection refused" start-of-body= Jan 26 16:49:35 crc kubenswrapper[4754]: I0126 16:49:35.010826 4754 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-7xsg2" podUID="f06be1d1-4134-4ad7-9233-b4099687bfe2" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.36:8080/healthz\": dial tcp 10.217.0.36:8080: connect: connection refused" Jan 26 16:49:35 crc kubenswrapper[4754]: I0126 16:49:35.012324 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-47pnn\" (UID: \"d098371a-9920-44d2-b918-f21c8142fac2\") " pod="openshift-image-registry/image-registry-697d97f7c8-47pnn" Jan 26 16:49:35 crc kubenswrapper[4754]: E0126 16:49:35.014643 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 16:49:35.514627739 +0000 UTC m=+142.038808173 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-47pnn" (UID: "d098371a-9920-44d2-b918-f21c8142fac2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 16:49:35 crc kubenswrapper[4754]: I0126 16:49:35.017714 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8rhs2" event={"ID":"4d0ec6e9-f4f7-4751-8317-60940c63f320","Type":"ContainerStarted","Data":"44f0d9e218c0c8d884a08dbaabe82b356e6214d233497945ab514dfc36bcb2cf"} Jan 26 16:49:35 crc kubenswrapper[4754]: I0126 16:49:35.063173 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-fbbxr" event={"ID":"2644a03c-ada2-45c0-ac4e-edfd969a7752","Type":"ContainerStarted","Data":"a5569b0cbc2c1891e351d1084e9720a701d03473f3ab170ad73ecc0dae063536"} Jan 26 16:49:35 crc kubenswrapper[4754]: I0126 16:49:35.068088 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-fbbxr" Jan 26 16:49:35 crc kubenswrapper[4754]: I0126 16:49:35.083248 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-m2ktn" event={"ID":"221280be-801c-4f1c-abbc-cb8a02f47838","Type":"ContainerStarted","Data":"c08bff35cf42d985a590c15f582b6447b0e88bb1690b0b83d04a40136110fc02"} Jan 26 16:49:35 crc kubenswrapper[4754]: I0126 16:49:35.088285 4754 csr.go:261] certificate signing request csr-rndr8 is approved, waiting to be issued Jan 26 16:49:35 crc kubenswrapper[4754]: I0126 16:49:35.088931 4754 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-fbbxr container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.43:8443/healthz\": dial tcp 10.217.0.43:8443: connect: connection refused" start-of-body= Jan 26 16:49:35 crc kubenswrapper[4754]: I0126 16:49:35.089042 4754 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-fbbxr" podUID="2644a03c-ada2-45c0-ac4e-edfd969a7752" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.43:8443/healthz\": dial tcp 10.217.0.43:8443: connect: connection refused" Jan 26 16:49:35 crc kubenswrapper[4754]: I0126 16:49:35.113627 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-7tnbd" event={"ID":"432c543c-59d1-49c9-ba43-91be5e113975","Type":"ContainerStarted","Data":"7318c52c7d83cf36e37d6488c432f70d4929190a527a7bd5dffc2c6b2948d80e"} Jan 26 16:49:35 crc kubenswrapper[4754]: I0126 16:49:35.113851 4754 csr.go:257] certificate signing request csr-rndr8 is issued Jan 26 16:49:35 crc kubenswrapper[4754]: I0126 16:49:35.114325 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 16:49:35 crc kubenswrapper[4754]: E0126 16:49:35.115452 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 16:49:35.615437456 +0000 UTC m=+142.139617890 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 16:49:35 crc kubenswrapper[4754]: I0126 16:49:35.136804 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-z2k6h" podStartSLOduration=122.136774142 podStartE2EDuration="2m2.136774142s" podCreationTimestamp="2026-01-26 16:47:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 16:49:35.133062439 +0000 UTC m=+141.657242873" watchObservedRunningTime="2026-01-26 16:49:35.136774142 +0000 UTC m=+141.660954576" Jan 26 16:49:35 crc kubenswrapper[4754]: I0126 16:49:35.139593 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8rjs9" event={"ID":"9caa5ea4-1245-4dc7-af31-571f68cec263","Type":"ContainerStarted","Data":"a1698d010bebe35d2c5d019e1c493aef359ec4335e7c3249e1869bf5dd21b0d9"} Jan 26 16:49:35 crc kubenswrapper[4754]: I0126 16:49:35.157834 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7lgch" event={"ID":"be3e5760-2f70-423b-9dd8-b5eeb0482b80","Type":"ContainerStarted","Data":"a0e9a69b36c71a5724a9e4435f11fd430a61c735d4d266974b91911667adc3d9"} Jan 26 16:49:35 crc kubenswrapper[4754]: I0126 16:49:35.158606 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7lgch" Jan 26 16:49:35 crc kubenswrapper[4754]: I0126 16:49:35.159834 4754 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-7lgch container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.17:5443/healthz\": dial tcp 10.217.0.17:5443: connect: connection refused" start-of-body= Jan 26 16:49:35 crc kubenswrapper[4754]: I0126 16:49:35.159869 4754 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7lgch" podUID="be3e5760-2f70-423b-9dd8-b5eeb0482b80" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.17:5443/healthz\": dial tcp 10.217.0.17:5443: connect: connection refused" Jan 26 16:49:35 crc kubenswrapper[4754]: I0126 16:49:35.183904 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-8647g" event={"ID":"fa9a982c-684d-497d-b8eb-f921e00fe511","Type":"ContainerStarted","Data":"d4f7b5caadcc99caf764fbf5f4441bcef2eadb50696f9660708464fbde921d44"} Jan 26 16:49:35 crc kubenswrapper[4754]: I0126 16:49:35.207719 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-vfldk" event={"ID":"3f0de367-82c0-4548-8f81-8c3482795003","Type":"ContainerStarted","Data":"1941dd9bbb8f03f3e8ac4ebd6fc671dc7e779011e854bb8db5affd9b76525a39"} Jan 26 16:49:35 crc kubenswrapper[4754]: I0126 16:49:35.216519 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-47pnn\" (UID: \"d098371a-9920-44d2-b918-f21c8142fac2\") " pod="openshift-image-registry/image-registry-697d97f7c8-47pnn" Jan 26 16:49:35 crc kubenswrapper[4754]: E0126 16:49:35.217540 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 16:49:35.71752843 +0000 UTC m=+142.241708864 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-47pnn" (UID: "d098371a-9920-44d2-b918-f21c8142fac2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 16:49:35 crc kubenswrapper[4754]: I0126 16:49:35.223575 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-k5sf5" event={"ID":"713fa157-05f9-4c30-b7e2-47a2b47cf7c4","Type":"ContainerStarted","Data":"e872256702e6de9deeb7cb2bc59a9f5adfedf0c47987f3ef9d0b36b75b9e7ff1"} Jan 26 16:49:35 crc kubenswrapper[4754]: I0126 16:49:35.223624 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-k5sf5" event={"ID":"713fa157-05f9-4c30-b7e2-47a2b47cf7c4","Type":"ContainerStarted","Data":"a83904b787cef4f892853cf5bf8697e3cd8a4e6b98f1df011ba5cdf56a9a1e73"} Jan 26 16:49:35 crc kubenswrapper[4754]: I0126 16:49:35.238746 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-ptrdc" event={"ID":"39a0ee2d-0f5b-4231-94a3-d11d517d3517","Type":"ContainerStarted","Data":"694f8bd225d689ab7440824b0328f660fcb41ad8de1c9625d80a44f0f88eebf4"} Jan 26 16:49:35 crc kubenswrapper[4754]: I0126 16:49:35.256974 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-fbbxr" podStartSLOduration=122.256955295 podStartE2EDuration="2m2.256955295s" podCreationTimestamp="2026-01-26 16:47:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 16:49:35.255169571 +0000 UTC m=+141.779350025" watchObservedRunningTime="2026-01-26 16:49:35.256955295 +0000 UTC m=+141.781135729" Jan 26 16:49:35 crc kubenswrapper[4754]: I0126 16:49:35.266536 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-8d467" event={"ID":"10609d5f-d6dd-40a5-acc4-63e62584d18e","Type":"ContainerStarted","Data":"3f8c016723913e675930cac5ab044976c7882e63c798f850ffacbfd28f670c93"} Jan 26 16:49:35 crc kubenswrapper[4754]: I0126 16:49:35.273725 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6mj2d" event={"ID":"bd64c275-cf75-4e42-8643-9178665801cb","Type":"ContainerStarted","Data":"a36b80d9afc150a2b40aaf8bb91fce6ab12998549a313a9cb8518ae35312d7be"} Jan 26 16:49:35 crc kubenswrapper[4754]: I0126 16:49:35.286838 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-4px9m" Jan 26 16:49:35 crc kubenswrapper[4754]: I0126 16:49:35.301459 4754 patch_prober.go:28] interesting pod/router-default-5444994796-jxzst container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 26 16:49:35 crc kubenswrapper[4754]: [-]has-synced failed: reason withheld Jan 26 16:49:35 crc kubenswrapper[4754]: [+]process-running ok Jan 26 16:49:35 crc kubenswrapper[4754]: healthz check failed Jan 26 16:49:35 crc kubenswrapper[4754]: I0126 16:49:35.301526 4754 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-jxzst" podUID="44e9e245-6889-4706-bc76-a22f2e7f6019" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 26 16:49:35 crc kubenswrapper[4754]: I0126 16:49:35.327104 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 16:49:35 crc kubenswrapper[4754]: E0126 16:49:35.327196 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 16:49:35.827175895 +0000 UTC m=+142.351356329 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 16:49:35 crc kubenswrapper[4754]: I0126 16:49:35.328028 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-47pnn\" (UID: \"d098371a-9920-44d2-b918-f21c8142fac2\") " pod="openshift-image-registry/image-registry-697d97f7c8-47pnn" Jan 26 16:49:35 crc kubenswrapper[4754]: E0126 16:49:35.331061 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 16:49:35.831044064 +0000 UTC m=+142.355224488 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-47pnn" (UID: "d098371a-9920-44d2-b918-f21c8142fac2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 16:49:35 crc kubenswrapper[4754]: I0126 16:49:35.380445 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-m2ktn" podStartSLOduration=122.380421906 podStartE2EDuration="2m2.380421906s" podCreationTimestamp="2026-01-26 16:47:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 16:49:35.373211587 +0000 UTC m=+141.897392021" watchObservedRunningTime="2026-01-26 16:49:35.380421906 +0000 UTC m=+141.904602340" Jan 26 16:49:35 crc kubenswrapper[4754]: I0126 16:49:35.381815 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-w7zlr" podStartSLOduration=122.381810206 podStartE2EDuration="2m2.381810206s" podCreationTimestamp="2026-01-26 16:47:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 16:49:35.325293037 +0000 UTC m=+141.849473471" watchObservedRunningTime="2026-01-26 16:49:35.381810206 +0000 UTC m=+141.905990640" Jan 26 16:49:35 crc kubenswrapper[4754]: I0126 16:49:35.437114 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 16:49:35 crc kubenswrapper[4754]: E0126 16:49:35.437489 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 16:49:35.937455892 +0000 UTC m=+142.461636326 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 16:49:35 crc kubenswrapper[4754]: I0126 16:49:35.445092 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-lxx9x" podStartSLOduration=8.445075756 podStartE2EDuration="8.445075756s" podCreationTimestamp="2026-01-26 16:49:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 16:49:35.420028507 +0000 UTC m=+141.944208941" watchObservedRunningTime="2026-01-26 16:49:35.445075756 +0000 UTC m=+141.969256180" Jan 26 16:49:35 crc kubenswrapper[4754]: I0126 16:49:35.516904 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29490765-w8zqq" podStartSLOduration=122.516887223 podStartE2EDuration="2m2.516887223s" podCreationTimestamp="2026-01-26 16:47:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 16:49:35.477258781 +0000 UTC m=+142.001439215" watchObservedRunningTime="2026-01-26 16:49:35.516887223 +0000 UTC m=+142.041067657" Jan 26 16:49:35 crc kubenswrapper[4754]: I0126 16:49:35.541009 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-47pnn\" (UID: \"d098371a-9920-44d2-b918-f21c8142fac2\") " pod="openshift-image-registry/image-registry-697d97f7c8-47pnn" Jan 26 16:49:35 crc kubenswrapper[4754]: E0126 16:49:35.541475 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 16:49:36.041457425 +0000 UTC m=+142.565637859 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-47pnn" (UID: "d098371a-9920-44d2-b918-f21c8142fac2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 16:49:35 crc kubenswrapper[4754]: I0126 16:49:35.558393 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6mj2d" podStartSLOduration=122.558376783 podStartE2EDuration="2m2.558376783s" podCreationTimestamp="2026-01-26 16:47:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 16:49:35.519357492 +0000 UTC m=+142.043537936" watchObservedRunningTime="2026-01-26 16:49:35.558376783 +0000 UTC m=+142.082557217" Jan 26 16:49:35 crc kubenswrapper[4754]: I0126 16:49:35.558653 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8rjs9" podStartSLOduration=122.558646332 podStartE2EDuration="2m2.558646332s" podCreationTimestamp="2026-01-26 16:47:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 16:49:35.557192009 +0000 UTC m=+142.081372453" watchObservedRunningTime="2026-01-26 16:49:35.558646332 +0000 UTC m=+142.082826766" Jan 26 16:49:35 crc kubenswrapper[4754]: I0126 16:49:35.641663 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 16:49:35 crc kubenswrapper[4754]: E0126 16:49:35.642122 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 16:49:36.142087886 +0000 UTC m=+142.666268320 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 16:49:35 crc kubenswrapper[4754]: I0126 16:49:35.647102 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7lgch" podStartSLOduration=122.647083155 podStartE2EDuration="2m2.647083155s" podCreationTimestamp="2026-01-26 16:47:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 16:49:35.616506488 +0000 UTC m=+142.140686932" watchObservedRunningTime="2026-01-26 16:49:35.647083155 +0000 UTC m=+142.171263579" Jan 26 16:49:35 crc kubenswrapper[4754]: I0126 16:49:35.700787 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-ptrdc" podStartSLOduration=122.700771982 podStartE2EDuration="2m2.700771982s" podCreationTimestamp="2026-01-26 16:47:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 16:49:35.699006469 +0000 UTC m=+142.223186903" watchObservedRunningTime="2026-01-26 16:49:35.700771982 +0000 UTC m=+142.224952416" Jan 26 16:49:35 crc kubenswrapper[4754]: I0126 16:49:35.733434 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-vfldk" podStartSLOduration=122.733414793 podStartE2EDuration="2m2.733414793s" podCreationTimestamp="2026-01-26 16:47:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 16:49:35.733187725 +0000 UTC m=+142.257368149" watchObservedRunningTime="2026-01-26 16:49:35.733414793 +0000 UTC m=+142.257595227" Jan 26 16:49:35 crc kubenswrapper[4754]: I0126 16:49:35.742487 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-47pnn\" (UID: \"d098371a-9920-44d2-b918-f21c8142fac2\") " pod="openshift-image-registry/image-registry-697d97f7c8-47pnn" Jan 26 16:49:35 crc kubenswrapper[4754]: E0126 16:49:35.742850 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 16:49:36.242839471 +0000 UTC m=+142.767019905 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-47pnn" (UID: "d098371a-9920-44d2-b918-f21c8142fac2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 16:49:35 crc kubenswrapper[4754]: I0126 16:49:35.844342 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-wjdgm" Jan 26 16:49:35 crc kubenswrapper[4754]: I0126 16:49:35.844396 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-k9sz7" Jan 26 16:49:35 crc kubenswrapper[4754]: I0126 16:49:35.848429 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 16:49:35 crc kubenswrapper[4754]: E0126 16:49:35.848784 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 16:49:36.348756552 +0000 UTC m=+142.872936986 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 16:49:35 crc kubenswrapper[4754]: I0126 16:49:35.849189 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-47pnn\" (UID: \"d098371a-9920-44d2-b918-f21c8142fac2\") " pod="openshift-image-registry/image-registry-697d97f7c8-47pnn" Jan 26 16:49:35 crc kubenswrapper[4754]: E0126 16:49:35.849517 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 16:49:36.349502719 +0000 UTC m=+142.873683153 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-47pnn" (UID: "d098371a-9920-44d2-b918-f21c8142fac2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 16:49:35 crc kubenswrapper[4754]: I0126 16:49:35.876561 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-qzw56"] Jan 26 16:49:35 crc kubenswrapper[4754]: I0126 16:49:35.877562 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qzw56" Jan 26 16:49:35 crc kubenswrapper[4754]: W0126 16:49:35.882000 4754 reflector.go:561] object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g": failed to list *v1.Secret: secrets "certified-operators-dockercfg-4rs5g" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-marketplace": no relationship found between node 'crc' and this object Jan 26 16:49:35 crc kubenswrapper[4754]: E0126 16:49:35.882031 4754 reflector.go:158] "Unhandled Error" err="object-\"openshift-marketplace\"/\"certified-operators-dockercfg-4rs5g\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"certified-operators-dockercfg-4rs5g\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-marketplace\": no relationship found between node 'crc' and this object" logger="UnhandledError" Jan 26 16:49:35 crc kubenswrapper[4754]: I0126 16:49:35.906010 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-qzw56"] Jan 26 16:49:35 crc kubenswrapper[4754]: I0126 16:49:35.953279 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 16:49:35 crc kubenswrapper[4754]: E0126 16:49:35.953474 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 16:49:36.45344982 +0000 UTC m=+142.977630254 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 16:49:35 crc kubenswrapper[4754]: I0126 16:49:35.953539 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-47pnn\" (UID: \"d098371a-9920-44d2-b918-f21c8142fac2\") " pod="openshift-image-registry/image-registry-697d97f7c8-47pnn" Jan 26 16:49:35 crc kubenswrapper[4754]: I0126 16:49:35.953677 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xwf62\" (UniqueName: \"kubernetes.io/projected/0ca0de6e-d004-4b46-aee4-66eae716184f-kube-api-access-xwf62\") pod \"certified-operators-qzw56\" (UID: \"0ca0de6e-d004-4b46-aee4-66eae716184f\") " pod="openshift-marketplace/certified-operators-qzw56" Jan 26 16:49:35 crc kubenswrapper[4754]: I0126 16:49:35.953891 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0ca0de6e-d004-4b46-aee4-66eae716184f-catalog-content\") pod \"certified-operators-qzw56\" (UID: \"0ca0de6e-d004-4b46-aee4-66eae716184f\") " pod="openshift-marketplace/certified-operators-qzw56" Jan 26 16:49:35 crc kubenswrapper[4754]: I0126 16:49:35.953928 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0ca0de6e-d004-4b46-aee4-66eae716184f-utilities\") pod \"certified-operators-qzw56\" (UID: \"0ca0de6e-d004-4b46-aee4-66eae716184f\") " pod="openshift-marketplace/certified-operators-qzw56" Jan 26 16:49:35 crc kubenswrapper[4754]: E0126 16:49:35.954280 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 16:49:36.454272039 +0000 UTC m=+142.978452473 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-47pnn" (UID: "d098371a-9920-44d2-b918-f21c8142fac2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 16:49:36 crc kubenswrapper[4754]: I0126 16:49:36.055263 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 16:49:36 crc kubenswrapper[4754]: I0126 16:49:36.055441 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0ca0de6e-d004-4b46-aee4-66eae716184f-catalog-content\") pod \"certified-operators-qzw56\" (UID: \"0ca0de6e-d004-4b46-aee4-66eae716184f\") " pod="openshift-marketplace/certified-operators-qzw56" Jan 26 16:49:36 crc kubenswrapper[4754]: I0126 16:49:36.055463 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0ca0de6e-d004-4b46-aee4-66eae716184f-utilities\") pod \"certified-operators-qzw56\" (UID: \"0ca0de6e-d004-4b46-aee4-66eae716184f\") " pod="openshift-marketplace/certified-operators-qzw56" Jan 26 16:49:36 crc kubenswrapper[4754]: I0126 16:49:36.055515 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xwf62\" (UniqueName: \"kubernetes.io/projected/0ca0de6e-d004-4b46-aee4-66eae716184f-kube-api-access-xwf62\") pod \"certified-operators-qzw56\" (UID: \"0ca0de6e-d004-4b46-aee4-66eae716184f\") " pod="openshift-marketplace/certified-operators-qzw56" Jan 26 16:49:36 crc kubenswrapper[4754]: E0126 16:49:36.055724 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 16:49:36.555706559 +0000 UTC m=+143.079886993 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 16:49:36 crc kubenswrapper[4754]: I0126 16:49:36.056095 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0ca0de6e-d004-4b46-aee4-66eae716184f-catalog-content\") pod \"certified-operators-qzw56\" (UID: \"0ca0de6e-d004-4b46-aee4-66eae716184f\") " pod="openshift-marketplace/certified-operators-qzw56" Jan 26 16:49:36 crc kubenswrapper[4754]: I0126 16:49:36.057252 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0ca0de6e-d004-4b46-aee4-66eae716184f-utilities\") pod \"certified-operators-qzw56\" (UID: \"0ca0de6e-d004-4b46-aee4-66eae716184f\") " pod="openshift-marketplace/certified-operators-qzw56" Jan 26 16:49:36 crc kubenswrapper[4754]: I0126 16:49:36.072390 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-cgm62"] Jan 26 16:49:36 crc kubenswrapper[4754]: I0126 16:49:36.073237 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cgm62" Jan 26 16:49:36 crc kubenswrapper[4754]: I0126 16:49:36.081358 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Jan 26 16:49:36 crc kubenswrapper[4754]: I0126 16:49:36.093364 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-cgm62"] Jan 26 16:49:36 crc kubenswrapper[4754]: I0126 16:49:36.117074 4754 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2027-01-26 16:44:35 +0000 UTC, rotation deadline is 2026-11-02 11:48:34.41356796 +0000 UTC Jan 26 16:49:36 crc kubenswrapper[4754]: I0126 16:49:36.117119 4754 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 6714h58m58.296451187s for next certificate rotation Jan 26 16:49:36 crc kubenswrapper[4754]: I0126 16:49:36.122734 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xwf62\" (UniqueName: \"kubernetes.io/projected/0ca0de6e-d004-4b46-aee4-66eae716184f-kube-api-access-xwf62\") pod \"certified-operators-qzw56\" (UID: \"0ca0de6e-d004-4b46-aee4-66eae716184f\") " pod="openshift-marketplace/certified-operators-qzw56" Jan 26 16:49:36 crc kubenswrapper[4754]: I0126 16:49:36.157070 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c2962a85-523e-4893-8a19-b7baa1360c01-utilities\") pod \"community-operators-cgm62\" (UID: \"c2962a85-523e-4893-8a19-b7baa1360c01\") " pod="openshift-marketplace/community-operators-cgm62" Jan 26 16:49:36 crc kubenswrapper[4754]: I0126 16:49:36.157145 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-47pnn\" (UID: \"d098371a-9920-44d2-b918-f21c8142fac2\") " pod="openshift-image-registry/image-registry-697d97f7c8-47pnn" Jan 26 16:49:36 crc kubenswrapper[4754]: I0126 16:49:36.157208 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z7pgd\" (UniqueName: \"kubernetes.io/projected/c2962a85-523e-4893-8a19-b7baa1360c01-kube-api-access-z7pgd\") pod \"community-operators-cgm62\" (UID: \"c2962a85-523e-4893-8a19-b7baa1360c01\") " pod="openshift-marketplace/community-operators-cgm62" Jan 26 16:49:36 crc kubenswrapper[4754]: I0126 16:49:36.157232 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c2962a85-523e-4893-8a19-b7baa1360c01-catalog-content\") pod \"community-operators-cgm62\" (UID: \"c2962a85-523e-4893-8a19-b7baa1360c01\") " pod="openshift-marketplace/community-operators-cgm62" Jan 26 16:49:36 crc kubenswrapper[4754]: E0126 16:49:36.157855 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 16:49:36.657842595 +0000 UTC m=+143.182023029 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-47pnn" (UID: "d098371a-9920-44d2-b918-f21c8142fac2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 16:49:36 crc kubenswrapper[4754]: I0126 16:49:36.260405 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 16:49:36 crc kubenswrapper[4754]: I0126 16:49:36.260651 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z7pgd\" (UniqueName: \"kubernetes.io/projected/c2962a85-523e-4893-8a19-b7baa1360c01-kube-api-access-z7pgd\") pod \"community-operators-cgm62\" (UID: \"c2962a85-523e-4893-8a19-b7baa1360c01\") " pod="openshift-marketplace/community-operators-cgm62" Jan 26 16:49:36 crc kubenswrapper[4754]: I0126 16:49:36.260702 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c2962a85-523e-4893-8a19-b7baa1360c01-catalog-content\") pod \"community-operators-cgm62\" (UID: \"c2962a85-523e-4893-8a19-b7baa1360c01\") " pod="openshift-marketplace/community-operators-cgm62" Jan 26 16:49:36 crc kubenswrapper[4754]: I0126 16:49:36.260816 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c2962a85-523e-4893-8a19-b7baa1360c01-utilities\") pod \"community-operators-cgm62\" (UID: \"c2962a85-523e-4893-8a19-b7baa1360c01\") " pod="openshift-marketplace/community-operators-cgm62" Jan 26 16:49:36 crc kubenswrapper[4754]: E0126 16:49:36.260876 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 16:49:36.760854211 +0000 UTC m=+143.285034645 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 16:49:36 crc kubenswrapper[4754]: I0126 16:49:36.260919 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-47pnn\" (UID: \"d098371a-9920-44d2-b918-f21c8142fac2\") " pod="openshift-image-registry/image-registry-697d97f7c8-47pnn" Jan 26 16:49:36 crc kubenswrapper[4754]: E0126 16:49:36.261246 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 16:49:36.761235745 +0000 UTC m=+143.285416179 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-47pnn" (UID: "d098371a-9920-44d2-b918-f21c8142fac2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 16:49:36 crc kubenswrapper[4754]: I0126 16:49:36.261422 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c2962a85-523e-4893-8a19-b7baa1360c01-catalog-content\") pod \"community-operators-cgm62\" (UID: \"c2962a85-523e-4893-8a19-b7baa1360c01\") " pod="openshift-marketplace/community-operators-cgm62" Jan 26 16:49:36 crc kubenswrapper[4754]: I0126 16:49:36.261778 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c2962a85-523e-4893-8a19-b7baa1360c01-utilities\") pod \"community-operators-cgm62\" (UID: \"c2962a85-523e-4893-8a19-b7baa1360c01\") " pod="openshift-marketplace/community-operators-cgm62" Jan 26 16:49:36 crc kubenswrapper[4754]: I0126 16:49:36.274956 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-pdrzp"] Jan 26 16:49:36 crc kubenswrapper[4754]: I0126 16:49:36.276038 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pdrzp" Jan 26 16:49:36 crc kubenswrapper[4754]: I0126 16:49:36.287637 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-pdrzp"] Jan 26 16:49:36 crc kubenswrapper[4754]: I0126 16:49:36.299140 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-95dbg" event={"ID":"5de8942d-833c-499b-842a-ddb2eeeab601","Type":"ContainerStarted","Data":"d4997731399a3d7a3ec8f2eaf42292a589b55c4b44c0dcbd06ceaa576fbf0007"} Jan 26 16:49:36 crc kubenswrapper[4754]: I0126 16:49:36.299735 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-95dbg" Jan 26 16:49:36 crc kubenswrapper[4754]: I0126 16:49:36.301362 4754 patch_prober.go:28] interesting pod/router-default-5444994796-jxzst container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 26 16:49:36 crc kubenswrapper[4754]: [-]has-synced failed: reason withheld Jan 26 16:49:36 crc kubenswrapper[4754]: [+]process-running ok Jan 26 16:49:36 crc kubenswrapper[4754]: healthz check failed Jan 26 16:49:36 crc kubenswrapper[4754]: I0126 16:49:36.301399 4754 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-jxzst" podUID="44e9e245-6889-4706-bc76-a22f2e7f6019" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 26 16:49:36 crc kubenswrapper[4754]: I0126 16:49:36.303390 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z7pgd\" (UniqueName: \"kubernetes.io/projected/c2962a85-523e-4893-8a19-b7baa1360c01-kube-api-access-z7pgd\") pod \"community-operators-cgm62\" (UID: \"c2962a85-523e-4893-8a19-b7baa1360c01\") " pod="openshift-marketplace/community-operators-cgm62" Jan 26 16:49:36 crc kubenswrapper[4754]: I0126 16:49:36.316495 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-k5sf5" event={"ID":"713fa157-05f9-4c30-b7e2-47a2b47cf7c4","Type":"ContainerStarted","Data":"4ce5976a45d895385933922695df67b26af0fcd33197a14f3dda785246d831b7"} Jan 26 16:49:36 crc kubenswrapper[4754]: I0126 16:49:36.336462 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-8d467" event={"ID":"10609d5f-d6dd-40a5-acc4-63e62584d18e","Type":"ContainerStarted","Data":"ba06f3e5183b4acd9afed05a81a01f7e6b1527b2c3e023d44d676ecea3148f42"} Jan 26 16:49:36 crc kubenswrapper[4754]: I0126 16:49:36.353859 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-m2ktn" event={"ID":"221280be-801c-4f1c-abbc-cb8a02f47838","Type":"ContainerStarted","Data":"6873fc67cafe1ae540536fd3e4b41dd5c43d00d6827b1250c29b93699965d900"} Jan 26 16:49:36 crc kubenswrapper[4754]: I0126 16:49:36.360790 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-95dbg" podStartSLOduration=123.360776277 podStartE2EDuration="2m3.360776277s" podCreationTimestamp="2026-01-26 16:47:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 16:49:36.360109663 +0000 UTC m=+142.884290107" watchObservedRunningTime="2026-01-26 16:49:36.360776277 +0000 UTC m=+142.884956711" Jan 26 16:49:36 crc kubenswrapper[4754]: I0126 16:49:36.361728 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 16:49:36 crc kubenswrapper[4754]: E0126 16:49:36.362113 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 16:49:36.862086354 +0000 UTC m=+143.386266788 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 16:49:36 crc kubenswrapper[4754]: I0126 16:49:36.362637 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-47pnn\" (UID: \"d098371a-9920-44d2-b918-f21c8142fac2\") " pod="openshift-image-registry/image-registry-697d97f7c8-47pnn" Jan 26 16:49:36 crc kubenswrapper[4754]: E0126 16:49:36.363647 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 16:49:36.86363355 +0000 UTC m=+143.387813974 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-47pnn" (UID: "d098371a-9920-44d2-b918-f21c8142fac2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 16:49:36 crc kubenswrapper[4754]: I0126 16:49:36.370166 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-472pk" event={"ID":"e89d71d4-488f-4bed-9356-5dc92d5a8eae","Type":"ContainerStarted","Data":"d02f18bb848721900a504b4dedb2fc9677d0629e471cbd85b20dd3140567ed7d"} Jan 26 16:49:36 crc kubenswrapper[4754]: I0126 16:49:36.382788 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8rjs9" event={"ID":"9caa5ea4-1245-4dc7-af31-571f68cec263","Type":"ContainerStarted","Data":"76347e16e31561b0fe18412471d0e2973106191023da82c6009b70858c0dd6f8"} Jan 26 16:49:36 crc kubenswrapper[4754]: I0126 16:49:36.391789 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cgm62" Jan 26 16:49:36 crc kubenswrapper[4754]: I0126 16:49:36.406001 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-8647g" event={"ID":"fa9a982c-684d-497d-b8eb-f921e00fe511","Type":"ContainerStarted","Data":"faa217e2be447502261eab7d616eec1e297330c51fbf8d5e51cab84e2737ec2e"} Jan 26 16:49:36 crc kubenswrapper[4754]: I0126 16:49:36.446572 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-k5sf5" podStartSLOduration=123.446554736 podStartE2EDuration="2m3.446554736s" podCreationTimestamp="2026-01-26 16:47:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 16:49:36.406910272 +0000 UTC m=+142.931090706" watchObservedRunningTime="2026-01-26 16:49:36.446554736 +0000 UTC m=+142.970735170" Jan 26 16:49:36 crc kubenswrapper[4754]: I0126 16:49:36.447183 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-7tnbd" event={"ID":"432c543c-59d1-49c9-ba43-91be5e113975","Type":"ContainerStarted","Data":"4eb78c6a653d1fc11408758414b774f73cf9cd1897cc8a82d536c550c8a322e2"} Jan 26 16:49:36 crc kubenswrapper[4754]: I0126 16:49:36.447234 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-7tnbd" event={"ID":"432c543c-59d1-49c9-ba43-91be5e113975","Type":"ContainerStarted","Data":"2d7812cecda9050c53f9c01bbc03cac24f9a37a8082d4c572bc0ab9af79a7047"} Jan 26 16:49:36 crc kubenswrapper[4754]: I0126 16:49:36.465148 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 16:49:36 crc kubenswrapper[4754]: I0126 16:49:36.465383 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8wwwp\" (UniqueName: \"kubernetes.io/projected/00549dbb-6196-4495-9fbe-4f641333fbe4-kube-api-access-8wwwp\") pod \"certified-operators-pdrzp\" (UID: \"00549dbb-6196-4495-9fbe-4f641333fbe4\") " pod="openshift-marketplace/certified-operators-pdrzp" Jan 26 16:49:36 crc kubenswrapper[4754]: I0126 16:49:36.465419 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/00549dbb-6196-4495-9fbe-4f641333fbe4-utilities\") pod \"certified-operators-pdrzp\" (UID: \"00549dbb-6196-4495-9fbe-4f641333fbe4\") " pod="openshift-marketplace/certified-operators-pdrzp" Jan 26 16:49:36 crc kubenswrapper[4754]: I0126 16:49:36.465449 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/00549dbb-6196-4495-9fbe-4f641333fbe4-catalog-content\") pod \"certified-operators-pdrzp\" (UID: \"00549dbb-6196-4495-9fbe-4f641333fbe4\") " pod="openshift-marketplace/certified-operators-pdrzp" Jan 26 16:49:36 crc kubenswrapper[4754]: E0126 16:49:36.465978 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 16:49:36.965949722 +0000 UTC m=+143.490130206 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 16:49:36 crc kubenswrapper[4754]: I0126 16:49:36.471587 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-8d467" podStartSLOduration=123.471566463 podStartE2EDuration="2m3.471566463s" podCreationTimestamp="2026-01-26 16:47:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 16:49:36.448276567 +0000 UTC m=+142.972457011" watchObservedRunningTime="2026-01-26 16:49:36.471566463 +0000 UTC m=+142.995746897" Jan 26 16:49:36 crc kubenswrapper[4754]: I0126 16:49:36.474626 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-n6gc5"] Jan 26 16:49:36 crc kubenswrapper[4754]: I0126 16:49:36.475784 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-n6gc5" Jan 26 16:49:36 crc kubenswrapper[4754]: I0126 16:49:36.482027 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-9shrr" event={"ID":"12e696fe-360e-4ea7-93e6-5ca2d0fc8123","Type":"ContainerStarted","Data":"d5e307cc55e0e4d5e011bcf672c50fd957104a7fa83bdb6804fcbe4f8f88c149"} Jan 26 16:49:36 crc kubenswrapper[4754]: I0126 16:49:36.485104 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-472pk" podStartSLOduration=123.485083518 podStartE2EDuration="2m3.485083518s" podCreationTimestamp="2026-01-26 16:47:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 16:49:36.476987157 +0000 UTC m=+143.001167591" watchObservedRunningTime="2026-01-26 16:49:36.485083518 +0000 UTC m=+143.009263952" Jan 26 16:49:36 crc kubenswrapper[4754]: I0126 16:49:36.500581 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8rhs2" event={"ID":"4d0ec6e9-f4f7-4751-8317-60940c63f320","Type":"ContainerStarted","Data":"8a3de13ea02520bfda9adfa82b523681f47090583daac1383d355a0e0e981eb1"} Jan 26 16:49:36 crc kubenswrapper[4754]: I0126 16:49:36.500843 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8rhs2" event={"ID":"4d0ec6e9-f4f7-4751-8317-60940c63f320","Type":"ContainerStarted","Data":"9646250af2fd1d7a1a3d840145cb729c7b5b4b235e2a01a9b506e8b33d2a5dc7"} Jan 26 16:49:36 crc kubenswrapper[4754]: I0126 16:49:36.517452 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-964q6" event={"ID":"593039fd-1af9-41c5-8f1a-bce2d698da75","Type":"ContainerStarted","Data":"955956138ca8cd3925f7043669fa6ce983e21438dce28616629813e1eec20e85"} Jan 26 16:49:36 crc kubenswrapper[4754]: I0126 16:49:36.517943 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-964q6" Jan 26 16:49:36 crc kubenswrapper[4754]: I0126 16:49:36.525504 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7lgch" event={"ID":"be3e5760-2f70-423b-9dd8-b5eeb0482b80","Type":"ContainerStarted","Data":"e156a5ba129944886d03dd4a66bc2fe6e0b8b53a2ef42d87bd56c0502045ef9d"} Jan 26 16:49:36 crc kubenswrapper[4754]: I0126 16:49:36.526151 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-n6gc5"] Jan 26 16:49:36 crc kubenswrapper[4754]: I0126 16:49:36.528489 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-7tnbd" podStartSLOduration=123.528479375 podStartE2EDuration="2m3.528479375s" podCreationTimestamp="2026-01-26 16:47:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 16:49:36.526457643 +0000 UTC m=+143.050638077" watchObservedRunningTime="2026-01-26 16:49:36.528479375 +0000 UTC m=+143.052659809" Jan 26 16:49:36 crc kubenswrapper[4754]: I0126 16:49:36.558054 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-vfldk" event={"ID":"3f0de367-82c0-4548-8f81-8c3482795003","Type":"ContainerStarted","Data":"5ac957cd8a1e9df9abbb3e7ff7c419531bbde6b0e304d3e46deba45ba1744cce"} Jan 26 16:49:36 crc kubenswrapper[4754]: I0126 16:49:36.559875 4754 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-7xsg2 container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.36:8080/healthz\": dial tcp 10.217.0.36:8080: connect: connection refused" start-of-body= Jan 26 16:49:36 crc kubenswrapper[4754]: I0126 16:49:36.559923 4754 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-7xsg2" podUID="f06be1d1-4134-4ad7-9233-b4099687bfe2" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.36:8080/healthz\": dial tcp 10.217.0.36:8080: connect: connection refused" Jan 26 16:49:36 crc kubenswrapper[4754]: I0126 16:49:36.582067 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/00549dbb-6196-4495-9fbe-4f641333fbe4-catalog-content\") pod \"certified-operators-pdrzp\" (UID: \"00549dbb-6196-4495-9fbe-4f641333fbe4\") " pod="openshift-marketplace/certified-operators-pdrzp" Jan 26 16:49:36 crc kubenswrapper[4754]: I0126 16:49:36.582279 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-47pnn\" (UID: \"d098371a-9920-44d2-b918-f21c8142fac2\") " pod="openshift-image-registry/image-registry-697d97f7c8-47pnn" Jan 26 16:49:36 crc kubenswrapper[4754]: I0126 16:49:36.582450 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8wwwp\" (UniqueName: \"kubernetes.io/projected/00549dbb-6196-4495-9fbe-4f641333fbe4-kube-api-access-8wwwp\") pod \"certified-operators-pdrzp\" (UID: \"00549dbb-6196-4495-9fbe-4f641333fbe4\") " pod="openshift-marketplace/certified-operators-pdrzp" Jan 26 16:49:36 crc kubenswrapper[4754]: I0126 16:49:36.582588 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/00549dbb-6196-4495-9fbe-4f641333fbe4-utilities\") pod \"certified-operators-pdrzp\" (UID: \"00549dbb-6196-4495-9fbe-4f641333fbe4\") " pod="openshift-marketplace/certified-operators-pdrzp" Jan 26 16:49:36 crc kubenswrapper[4754]: E0126 16:49:36.583029 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 16:49:37.083011862 +0000 UTC m=+143.607192296 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-47pnn" (UID: "d098371a-9920-44d2-b918-f21c8142fac2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 16:49:36 crc kubenswrapper[4754]: I0126 16:49:36.583941 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/00549dbb-6196-4495-9fbe-4f641333fbe4-utilities\") pod \"certified-operators-pdrzp\" (UID: \"00549dbb-6196-4495-9fbe-4f641333fbe4\") " pod="openshift-marketplace/certified-operators-pdrzp" Jan 26 16:49:36 crc kubenswrapper[4754]: I0126 16:49:36.590130 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/00549dbb-6196-4495-9fbe-4f641333fbe4-catalog-content\") pod \"certified-operators-pdrzp\" (UID: \"00549dbb-6196-4495-9fbe-4f641333fbe4\") " pod="openshift-marketplace/certified-operators-pdrzp" Jan 26 16:49:36 crc kubenswrapper[4754]: I0126 16:49:36.604067 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-mj6ns" Jan 26 16:49:36 crc kubenswrapper[4754]: I0126 16:49:36.609433 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-fbbxr" Jan 26 16:49:36 crc kubenswrapper[4754]: I0126 16:49:36.624566 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-8647g" podStartSLOduration=123.624540772 podStartE2EDuration="2m3.624540772s" podCreationTimestamp="2026-01-26 16:47:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 16:49:36.612147268 +0000 UTC m=+143.136327712" watchObservedRunningTime="2026-01-26 16:49:36.624540772 +0000 UTC m=+143.148721206" Jan 26 16:49:36 crc kubenswrapper[4754]: I0126 16:49:36.661783 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8wwwp\" (UniqueName: \"kubernetes.io/projected/00549dbb-6196-4495-9fbe-4f641333fbe4-kube-api-access-8wwwp\") pod \"certified-operators-pdrzp\" (UID: \"00549dbb-6196-4495-9fbe-4f641333fbe4\") " pod="openshift-marketplace/certified-operators-pdrzp" Jan 26 16:49:36 crc kubenswrapper[4754]: I0126 16:49:36.684387 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 16:49:36 crc kubenswrapper[4754]: I0126 16:49:36.684945 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4n544\" (UniqueName: \"kubernetes.io/projected/405f8b06-885b-4a7f-9f85-3af411a5504e-kube-api-access-4n544\") pod \"community-operators-n6gc5\" (UID: \"405f8b06-885b-4a7f-9f85-3af411a5504e\") " pod="openshift-marketplace/community-operators-n6gc5" Jan 26 16:49:36 crc kubenswrapper[4754]: I0126 16:49:36.684993 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/405f8b06-885b-4a7f-9f85-3af411a5504e-utilities\") pod \"community-operators-n6gc5\" (UID: \"405f8b06-885b-4a7f-9f85-3af411a5504e\") " pod="openshift-marketplace/community-operators-n6gc5" Jan 26 16:49:36 crc kubenswrapper[4754]: I0126 16:49:36.685141 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/405f8b06-885b-4a7f-9f85-3af411a5504e-catalog-content\") pod \"community-operators-n6gc5\" (UID: \"405f8b06-885b-4a7f-9f85-3af411a5504e\") " pod="openshift-marketplace/community-operators-n6gc5" Jan 26 16:49:36 crc kubenswrapper[4754]: E0126 16:49:36.686027 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 16:49:37.186012509 +0000 UTC m=+143.710192943 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 16:49:36 crc kubenswrapper[4754]: I0126 16:49:36.745605 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8rhs2" podStartSLOduration=123.745587636 podStartE2EDuration="2m3.745587636s" podCreationTimestamp="2026-01-26 16:47:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 16:49:36.744541499 +0000 UTC m=+143.268721933" watchObservedRunningTime="2026-01-26 16:49:36.745587636 +0000 UTC m=+143.269768070" Jan 26 16:49:36 crc kubenswrapper[4754]: I0126 16:49:36.788478 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/405f8b06-885b-4a7f-9f85-3af411a5504e-catalog-content\") pod \"community-operators-n6gc5\" (UID: \"405f8b06-885b-4a7f-9f85-3af411a5504e\") " pod="openshift-marketplace/community-operators-n6gc5" Jan 26 16:49:36 crc kubenswrapper[4754]: I0126 16:49:36.788625 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4n544\" (UniqueName: \"kubernetes.io/projected/405f8b06-885b-4a7f-9f85-3af411a5504e-kube-api-access-4n544\") pod \"community-operators-n6gc5\" (UID: \"405f8b06-885b-4a7f-9f85-3af411a5504e\") " pod="openshift-marketplace/community-operators-n6gc5" Jan 26 16:49:36 crc kubenswrapper[4754]: I0126 16:49:36.788646 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/405f8b06-885b-4a7f-9f85-3af411a5504e-utilities\") pod \"community-operators-n6gc5\" (UID: \"405f8b06-885b-4a7f-9f85-3af411a5504e\") " pod="openshift-marketplace/community-operators-n6gc5" Jan 26 16:49:36 crc kubenswrapper[4754]: I0126 16:49:36.788672 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-47pnn\" (UID: \"d098371a-9920-44d2-b918-f21c8142fac2\") " pod="openshift-image-registry/image-registry-697d97f7c8-47pnn" Jan 26 16:49:36 crc kubenswrapper[4754]: E0126 16:49:36.788945 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 16:49:37.288932602 +0000 UTC m=+143.813113036 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-47pnn" (UID: "d098371a-9920-44d2-b918-f21c8142fac2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 16:49:36 crc kubenswrapper[4754]: I0126 16:49:36.789611 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/405f8b06-885b-4a7f-9f85-3af411a5504e-catalog-content\") pod \"community-operators-n6gc5\" (UID: \"405f8b06-885b-4a7f-9f85-3af411a5504e\") " pod="openshift-marketplace/community-operators-n6gc5" Jan 26 16:49:36 crc kubenswrapper[4754]: I0126 16:49:36.790143 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/405f8b06-885b-4a7f-9f85-3af411a5504e-utilities\") pod \"community-operators-n6gc5\" (UID: \"405f8b06-885b-4a7f-9f85-3af411a5504e\") " pod="openshift-marketplace/community-operators-n6gc5" Jan 26 16:49:36 crc kubenswrapper[4754]: I0126 16:49:36.849562 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-964q6" podStartSLOduration=9.849540627 podStartE2EDuration="9.849540627s" podCreationTimestamp="2026-01-26 16:49:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 16:49:36.80810292 +0000 UTC m=+143.332283354" watchObservedRunningTime="2026-01-26 16:49:36.849540627 +0000 UTC m=+143.373721061" Jan 26 16:49:36 crc kubenswrapper[4754]: I0126 16:49:36.881998 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Jan 26 16:49:36 crc kubenswrapper[4754]: I0126 16:49:36.882850 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qzw56" Jan 26 16:49:36 crc kubenswrapper[4754]: I0126 16:49:36.883658 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4n544\" (UniqueName: \"kubernetes.io/projected/405f8b06-885b-4a7f-9f85-3af411a5504e-kube-api-access-4n544\") pod \"community-operators-n6gc5\" (UID: \"405f8b06-885b-4a7f-9f85-3af411a5504e\") " pod="openshift-marketplace/community-operators-n6gc5" Jan 26 16:49:36 crc kubenswrapper[4754]: I0126 16:49:36.893373 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 16:49:36 crc kubenswrapper[4754]: E0126 16:49:36.893915 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 16:49:37.393892228 +0000 UTC m=+143.918072662 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 16:49:36 crc kubenswrapper[4754]: I0126 16:49:36.911592 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pdrzp" Jan 26 16:49:37 crc kubenswrapper[4754]: I0126 16:49:37.000660 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-47pnn\" (UID: \"d098371a-9920-44d2-b918-f21c8142fac2\") " pod="openshift-image-registry/image-registry-697d97f7c8-47pnn" Jan 26 16:49:37 crc kubenswrapper[4754]: E0126 16:49:37.001093 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 16:49:37.501077966 +0000 UTC m=+144.025258410 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-47pnn" (UID: "d098371a-9920-44d2-b918-f21c8142fac2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 16:49:37 crc kubenswrapper[4754]: I0126 16:49:37.104237 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 16:49:37 crc kubenswrapper[4754]: E0126 16:49:37.104727 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 16:49:37.604702844 +0000 UTC m=+144.128883288 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 16:49:37 crc kubenswrapper[4754]: I0126 16:49:37.116144 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-n6gc5" Jan 26 16:49:37 crc kubenswrapper[4754]: I0126 16:49:37.135831 4754 patch_prober.go:28] interesting pod/machine-config-daemon-x65wv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 26 16:49:37 crc kubenswrapper[4754]: I0126 16:49:37.135894 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 26 16:49:37 crc kubenswrapper[4754]: I0126 16:49:37.205361 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-47pnn\" (UID: \"d098371a-9920-44d2-b918-f21c8142fac2\") " pod="openshift-image-registry/image-registry-697d97f7c8-47pnn" Jan 26 16:49:37 crc kubenswrapper[4754]: E0126 16:49:37.205724 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 16:49:37.705712349 +0000 UTC m=+144.229892783 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-47pnn" (UID: "d098371a-9920-44d2-b918-f21c8142fac2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 16:49:37 crc kubenswrapper[4754]: I0126 16:49:37.230835 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-cgm62"] Jan 26 16:49:37 crc kubenswrapper[4754]: I0126 16:49:37.307224 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 16:49:37 crc kubenswrapper[4754]: E0126 16:49:37.307897 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 16:49:37.807880575 +0000 UTC m=+144.332061009 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 16:49:37 crc kubenswrapper[4754]: I0126 16:49:37.313859 4754 patch_prober.go:28] interesting pod/router-default-5444994796-jxzst container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 26 16:49:37 crc kubenswrapper[4754]: [-]has-synced failed: reason withheld Jan 26 16:49:37 crc kubenswrapper[4754]: [+]process-running ok Jan 26 16:49:37 crc kubenswrapper[4754]: healthz check failed Jan 26 16:49:37 crc kubenswrapper[4754]: I0126 16:49:37.313943 4754 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-jxzst" podUID="44e9e245-6889-4706-bc76-a22f2e7f6019" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 26 16:49:37 crc kubenswrapper[4754]: I0126 16:49:37.364080 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-pdrzp"] Jan 26 16:49:37 crc kubenswrapper[4754]: I0126 16:49:37.410457 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-47pnn\" (UID: \"d098371a-9920-44d2-b918-f21c8142fac2\") " pod="openshift-image-registry/image-registry-697d97f7c8-47pnn" Jan 26 16:49:37 crc kubenswrapper[4754]: E0126 16:49:37.410766 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 16:49:37.910753697 +0000 UTC m=+144.434934131 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-47pnn" (UID: "d098371a-9920-44d2-b918-f21c8142fac2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 16:49:37 crc kubenswrapper[4754]: I0126 16:49:37.515598 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 16:49:37 crc kubenswrapper[4754]: E0126 16:49:37.518889 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 16:49:38.018867387 +0000 UTC m=+144.543047821 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 16:49:37 crc kubenswrapper[4754]: I0126 16:49:37.524401 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-qzw56"] Jan 26 16:49:37 crc kubenswrapper[4754]: I0126 16:49:37.525805 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-n6gc5"] Jan 26 16:49:37 crc kubenswrapper[4754]: I0126 16:49:37.526825 4754 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-7lgch container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.17:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Jan 26 16:49:37 crc kubenswrapper[4754]: I0126 16:49:37.526893 4754 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7lgch" podUID="be3e5760-2f70-423b-9dd8-b5eeb0482b80" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.17:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Jan 26 16:49:37 crc kubenswrapper[4754]: I0126 16:49:37.564285 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-9shrr" event={"ID":"12e696fe-360e-4ea7-93e6-5ca2d0fc8123","Type":"ContainerStarted","Data":"5d34f699c78132f92acca8d505b7f92f6a65c9a934ab397866cffb2ed833919b"} Jan 26 16:49:37 crc kubenswrapper[4754]: I0126 16:49:37.574835 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pdrzp" event={"ID":"00549dbb-6196-4495-9fbe-4f641333fbe4","Type":"ContainerStarted","Data":"58ac598971e156a2bceaaf859c1344080a3816dafb3d572b49ec804f2ae08890"} Jan 26 16:49:37 crc kubenswrapper[4754]: I0126 16:49:37.582755 4754 generic.go:334] "Generic (PLEG): container finished" podID="c2962a85-523e-4893-8a19-b7baa1360c01" containerID="4df8dfad1bfb4083b96d0aa561d538a881291c90d138da405053526c0dba8ee4" exitCode=0 Jan 26 16:49:37 crc kubenswrapper[4754]: I0126 16:49:37.583532 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cgm62" event={"ID":"c2962a85-523e-4893-8a19-b7baa1360c01","Type":"ContainerDied","Data":"4df8dfad1bfb4083b96d0aa561d538a881291c90d138da405053526c0dba8ee4"} Jan 26 16:49:37 crc kubenswrapper[4754]: I0126 16:49:37.583639 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cgm62" event={"ID":"c2962a85-523e-4893-8a19-b7baa1360c01","Type":"ContainerStarted","Data":"883b049d0221704c56bf963e8d4d31fbcca1c49b4524942c567e53ec46248777"} Jan 26 16:49:37 crc kubenswrapper[4754]: I0126 16:49:37.618182 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-47pnn\" (UID: \"d098371a-9920-44d2-b918-f21c8142fac2\") " pod="openshift-image-registry/image-registry-697d97f7c8-47pnn" Jan 26 16:49:37 crc kubenswrapper[4754]: E0126 16:49:37.618518 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 16:49:38.118506083 +0000 UTC m=+144.642686517 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-47pnn" (UID: "d098371a-9920-44d2-b918-f21c8142fac2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 16:49:37 crc kubenswrapper[4754]: I0126 16:49:37.622597 4754 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 26 16:49:37 crc kubenswrapper[4754]: I0126 16:49:37.719862 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 16:49:37 crc kubenswrapper[4754]: E0126 16:49:37.722697 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 16:49:38.222666311 +0000 UTC m=+144.746846735 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 16:49:37 crc kubenswrapper[4754]: I0126 16:49:37.740381 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7lgch" Jan 26 16:49:37 crc kubenswrapper[4754]: I0126 16:49:37.842721 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-47pnn\" (UID: \"d098371a-9920-44d2-b918-f21c8142fac2\") " pod="openshift-image-registry/image-registry-697d97f7c8-47pnn" Jan 26 16:49:37 crc kubenswrapper[4754]: E0126 16:49:37.847819 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 16:49:38.347793061 +0000 UTC m=+144.871973495 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-47pnn" (UID: "d098371a-9920-44d2-b918-f21c8142fac2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 16:49:37 crc kubenswrapper[4754]: I0126 16:49:37.918675 4754 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Jan 26 16:49:37 crc kubenswrapper[4754]: I0126 16:49:37.950933 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 16:49:37 crc kubenswrapper[4754]: E0126 16:49:37.951076 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 16:49:38.451057677 +0000 UTC m=+144.975238111 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 16:49:37 crc kubenswrapper[4754]: I0126 16:49:37.951185 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-47pnn\" (UID: \"d098371a-9920-44d2-b918-f21c8142fac2\") " pod="openshift-image-registry/image-registry-697d97f7c8-47pnn" Jan 26 16:49:37 crc kubenswrapper[4754]: E0126 16:49:37.951519 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 16:49:38.451511413 +0000 UTC m=+144.975691847 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-47pnn" (UID: "d098371a-9920-44d2-b918-f21c8142fac2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 16:49:38 crc kubenswrapper[4754]: I0126 16:49:38.051649 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 16:49:38 crc kubenswrapper[4754]: E0126 16:49:38.051808 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 16:49:38.551780692 +0000 UTC m=+145.075961126 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 16:49:38 crc kubenswrapper[4754]: I0126 16:49:38.052035 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-47pnn\" (UID: \"d098371a-9920-44d2-b918-f21c8142fac2\") " pod="openshift-image-registry/image-registry-697d97f7c8-47pnn" Jan 26 16:49:38 crc kubenswrapper[4754]: E0126 16:49:38.052356 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 16:49:38.552348372 +0000 UTC m=+145.076528806 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-47pnn" (UID: "d098371a-9920-44d2-b918-f21c8142fac2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 16:49:38 crc kubenswrapper[4754]: I0126 16:49:38.061850 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-hhhn5"] Jan 26 16:49:38 crc kubenswrapper[4754]: I0126 16:49:38.062901 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hhhn5" Jan 26 16:49:38 crc kubenswrapper[4754]: I0126 16:49:38.064922 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Jan 26 16:49:38 crc kubenswrapper[4754]: I0126 16:49:38.076676 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-hhhn5"] Jan 26 16:49:38 crc kubenswrapper[4754]: I0126 16:49:38.152975 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 16:49:38 crc kubenswrapper[4754]: I0126 16:49:38.153450 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d55d7\" (UniqueName: \"kubernetes.io/projected/e36dc2d0-47de-46a6-aeba-52475754867e-kube-api-access-d55d7\") pod \"redhat-marketplace-hhhn5\" (UID: \"e36dc2d0-47de-46a6-aeba-52475754867e\") " pod="openshift-marketplace/redhat-marketplace-hhhn5" Jan 26 16:49:38 crc kubenswrapper[4754]: I0126 16:49:38.153493 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e36dc2d0-47de-46a6-aeba-52475754867e-utilities\") pod \"redhat-marketplace-hhhn5\" (UID: \"e36dc2d0-47de-46a6-aeba-52475754867e\") " pod="openshift-marketplace/redhat-marketplace-hhhn5" Jan 26 16:49:38 crc kubenswrapper[4754]: I0126 16:49:38.153542 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e36dc2d0-47de-46a6-aeba-52475754867e-catalog-content\") pod \"redhat-marketplace-hhhn5\" (UID: \"e36dc2d0-47de-46a6-aeba-52475754867e\") " pod="openshift-marketplace/redhat-marketplace-hhhn5" Jan 26 16:49:38 crc kubenswrapper[4754]: E0126 16:49:38.153647 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 16:49:38.653627366 +0000 UTC m=+145.177807800 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 16:49:38 crc kubenswrapper[4754]: I0126 16:49:38.254458 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d55d7\" (UniqueName: \"kubernetes.io/projected/e36dc2d0-47de-46a6-aeba-52475754867e-kube-api-access-d55d7\") pod \"redhat-marketplace-hhhn5\" (UID: \"e36dc2d0-47de-46a6-aeba-52475754867e\") " pod="openshift-marketplace/redhat-marketplace-hhhn5" Jan 26 16:49:38 crc kubenswrapper[4754]: I0126 16:49:38.254535 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e36dc2d0-47de-46a6-aeba-52475754867e-utilities\") pod \"redhat-marketplace-hhhn5\" (UID: \"e36dc2d0-47de-46a6-aeba-52475754867e\") " pod="openshift-marketplace/redhat-marketplace-hhhn5" Jan 26 16:49:38 crc kubenswrapper[4754]: I0126 16:49:38.254565 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-47pnn\" (UID: \"d098371a-9920-44d2-b918-f21c8142fac2\") " pod="openshift-image-registry/image-registry-697d97f7c8-47pnn" Jan 26 16:49:38 crc kubenswrapper[4754]: I0126 16:49:38.254620 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e36dc2d0-47de-46a6-aeba-52475754867e-catalog-content\") pod \"redhat-marketplace-hhhn5\" (UID: \"e36dc2d0-47de-46a6-aeba-52475754867e\") " pod="openshift-marketplace/redhat-marketplace-hhhn5" Jan 26 16:49:38 crc kubenswrapper[4754]: E0126 16:49:38.255089 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 16:49:38.755070107 +0000 UTC m=+145.279250541 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-47pnn" (UID: "d098371a-9920-44d2-b918-f21c8142fac2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 16:49:38 crc kubenswrapper[4754]: I0126 16:49:38.255124 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e36dc2d0-47de-46a6-aeba-52475754867e-utilities\") pod \"redhat-marketplace-hhhn5\" (UID: \"e36dc2d0-47de-46a6-aeba-52475754867e\") " pod="openshift-marketplace/redhat-marketplace-hhhn5" Jan 26 16:49:38 crc kubenswrapper[4754]: I0126 16:49:38.255155 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e36dc2d0-47de-46a6-aeba-52475754867e-catalog-content\") pod \"redhat-marketplace-hhhn5\" (UID: \"e36dc2d0-47de-46a6-aeba-52475754867e\") " pod="openshift-marketplace/redhat-marketplace-hhhn5" Jan 26 16:49:38 crc kubenswrapper[4754]: I0126 16:49:38.277879 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d55d7\" (UniqueName: \"kubernetes.io/projected/e36dc2d0-47de-46a6-aeba-52475754867e-kube-api-access-d55d7\") pod \"redhat-marketplace-hhhn5\" (UID: \"e36dc2d0-47de-46a6-aeba-52475754867e\") " pod="openshift-marketplace/redhat-marketplace-hhhn5" Jan 26 16:49:38 crc kubenswrapper[4754]: I0126 16:49:38.302065 4754 patch_prober.go:28] interesting pod/router-default-5444994796-jxzst container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 26 16:49:38 crc kubenswrapper[4754]: [-]has-synced failed: reason withheld Jan 26 16:49:38 crc kubenswrapper[4754]: [+]process-running ok Jan 26 16:49:38 crc kubenswrapper[4754]: healthz check failed Jan 26 16:49:38 crc kubenswrapper[4754]: I0126 16:49:38.302130 4754 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-jxzst" podUID="44e9e245-6889-4706-bc76-a22f2e7f6019" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 26 16:49:38 crc kubenswrapper[4754]: I0126 16:49:38.339842 4754 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2026-01-26T16:49:37.918732207Z","Handler":null,"Name":""} Jan 26 16:49:38 crc kubenswrapper[4754]: I0126 16:49:38.342852 4754 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Jan 26 16:49:38 crc kubenswrapper[4754]: I0126 16:49:38.343167 4754 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Jan 26 16:49:38 crc kubenswrapper[4754]: I0126 16:49:38.355842 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 16:49:38 crc kubenswrapper[4754]: I0126 16:49:38.360486 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Jan 26 16:49:38 crc kubenswrapper[4754]: I0126 16:49:38.401309 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hhhn5" Jan 26 16:49:38 crc kubenswrapper[4754]: I0126 16:49:38.457884 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-47pnn\" (UID: \"d098371a-9920-44d2-b918-f21c8142fac2\") " pod="openshift-image-registry/image-registry-697d97f7c8-47pnn" Jan 26 16:49:38 crc kubenswrapper[4754]: I0126 16:49:38.460435 4754 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Jan 26 16:49:38 crc kubenswrapper[4754]: I0126 16:49:38.460477 4754 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-47pnn\" (UID: \"d098371a-9920-44d2-b918-f21c8142fac2\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-47pnn" Jan 26 16:49:38 crc kubenswrapper[4754]: I0126 16:49:38.460820 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-plfzc"] Jan 26 16:49:38 crc kubenswrapper[4754]: I0126 16:49:38.461761 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-plfzc" Jan 26 16:49:38 crc kubenswrapper[4754]: I0126 16:49:38.471801 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-plfzc"] Jan 26 16:49:38 crc kubenswrapper[4754]: I0126 16:49:38.505930 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-47pnn\" (UID: \"d098371a-9920-44d2-b918-f21c8142fac2\") " pod="openshift-image-registry/image-registry-697d97f7c8-47pnn" Jan 26 16:49:38 crc kubenswrapper[4754]: I0126 16:49:38.516530 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-jsbxt" Jan 26 16:49:38 crc kubenswrapper[4754]: I0126 16:49:38.559224 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ffaa8507-8be1-4bdb-b7b5-28104f9cdbd6-catalog-content\") pod \"redhat-marketplace-plfzc\" (UID: \"ffaa8507-8be1-4bdb-b7b5-28104f9cdbd6\") " pod="openshift-marketplace/redhat-marketplace-plfzc" Jan 26 16:49:38 crc kubenswrapper[4754]: I0126 16:49:38.559379 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hts6c\" (UniqueName: \"kubernetes.io/projected/ffaa8507-8be1-4bdb-b7b5-28104f9cdbd6-kube-api-access-hts6c\") pod \"redhat-marketplace-plfzc\" (UID: \"ffaa8507-8be1-4bdb-b7b5-28104f9cdbd6\") " pod="openshift-marketplace/redhat-marketplace-plfzc" Jan 26 16:49:38 crc kubenswrapper[4754]: I0126 16:49:38.559483 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ffaa8507-8be1-4bdb-b7b5-28104f9cdbd6-utilities\") pod \"redhat-marketplace-plfzc\" (UID: \"ffaa8507-8be1-4bdb-b7b5-28104f9cdbd6\") " pod="openshift-marketplace/redhat-marketplace-plfzc" Jan 26 16:49:38 crc kubenswrapper[4754]: I0126 16:49:38.562680 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-47pnn" Jan 26 16:49:38 crc kubenswrapper[4754]: I0126 16:49:38.606083 4754 generic.go:334] "Generic (PLEG): container finished" podID="405f8b06-885b-4a7f-9f85-3af411a5504e" containerID="20a4db1eb833f1df1ecdad9bd783bf791d2a67664cb40a433a5f49bee71cbc7b" exitCode=0 Jan 26 16:49:38 crc kubenswrapper[4754]: I0126 16:49:38.606156 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n6gc5" event={"ID":"405f8b06-885b-4a7f-9f85-3af411a5504e","Type":"ContainerDied","Data":"20a4db1eb833f1df1ecdad9bd783bf791d2a67664cb40a433a5f49bee71cbc7b"} Jan 26 16:49:38 crc kubenswrapper[4754]: I0126 16:49:38.606186 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n6gc5" event={"ID":"405f8b06-885b-4a7f-9f85-3af411a5504e","Type":"ContainerStarted","Data":"bdde98850848c7ec866e6c5c166d541f33a37a6358e679b86327114f43383271"} Jan 26 16:49:38 crc kubenswrapper[4754]: I0126 16:49:38.653084 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-9shrr" event={"ID":"12e696fe-360e-4ea7-93e6-5ca2d0fc8123","Type":"ContainerStarted","Data":"5415036ec813b35cd90ce038d72b431c7b947669614e8b5272958970fec4bbd8"} Jan 26 16:49:38 crc kubenswrapper[4754]: I0126 16:49:38.653393 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-9shrr" event={"ID":"12e696fe-360e-4ea7-93e6-5ca2d0fc8123","Type":"ContainerStarted","Data":"ee79abf67a0b5098d9c94d1f802fbf8ab79bbbbc15575fe502a98bc9f398bc07"} Jan 26 16:49:38 crc kubenswrapper[4754]: I0126 16:49:38.657366 4754 generic.go:334] "Generic (PLEG): container finished" podID="00549dbb-6196-4495-9fbe-4f641333fbe4" containerID="12b0ca264597c9685d69dc4164af214cba7256ed8b758fd7235dab47fb7a5cb8" exitCode=0 Jan 26 16:49:38 crc kubenswrapper[4754]: I0126 16:49:38.657516 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pdrzp" event={"ID":"00549dbb-6196-4495-9fbe-4f641333fbe4","Type":"ContainerDied","Data":"12b0ca264597c9685d69dc4164af214cba7256ed8b758fd7235dab47fb7a5cb8"} Jan 26 16:49:38 crc kubenswrapper[4754]: I0126 16:49:38.660738 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ffaa8507-8be1-4bdb-b7b5-28104f9cdbd6-utilities\") pod \"redhat-marketplace-plfzc\" (UID: \"ffaa8507-8be1-4bdb-b7b5-28104f9cdbd6\") " pod="openshift-marketplace/redhat-marketplace-plfzc" Jan 26 16:49:38 crc kubenswrapper[4754]: I0126 16:49:38.660851 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ffaa8507-8be1-4bdb-b7b5-28104f9cdbd6-catalog-content\") pod \"redhat-marketplace-plfzc\" (UID: \"ffaa8507-8be1-4bdb-b7b5-28104f9cdbd6\") " pod="openshift-marketplace/redhat-marketplace-plfzc" Jan 26 16:49:38 crc kubenswrapper[4754]: I0126 16:49:38.660928 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hts6c\" (UniqueName: \"kubernetes.io/projected/ffaa8507-8be1-4bdb-b7b5-28104f9cdbd6-kube-api-access-hts6c\") pod \"redhat-marketplace-plfzc\" (UID: \"ffaa8507-8be1-4bdb-b7b5-28104f9cdbd6\") " pod="openshift-marketplace/redhat-marketplace-plfzc" Jan 26 16:49:38 crc kubenswrapper[4754]: I0126 16:49:38.662509 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ffaa8507-8be1-4bdb-b7b5-28104f9cdbd6-utilities\") pod \"redhat-marketplace-plfzc\" (UID: \"ffaa8507-8be1-4bdb-b7b5-28104f9cdbd6\") " pod="openshift-marketplace/redhat-marketplace-plfzc" Jan 26 16:49:38 crc kubenswrapper[4754]: I0126 16:49:38.663272 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ffaa8507-8be1-4bdb-b7b5-28104f9cdbd6-catalog-content\") pod \"redhat-marketplace-plfzc\" (UID: \"ffaa8507-8be1-4bdb-b7b5-28104f9cdbd6\") " pod="openshift-marketplace/redhat-marketplace-plfzc" Jan 26 16:49:38 crc kubenswrapper[4754]: I0126 16:49:38.667863 4754 generic.go:334] "Generic (PLEG): container finished" podID="0ca0de6e-d004-4b46-aee4-66eae716184f" containerID="c6ef6f376b3de6a06cb247c19a2dde75e11dca8e32c9db492300d119a97e789b" exitCode=0 Jan 26 16:49:38 crc kubenswrapper[4754]: I0126 16:49:38.669828 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qzw56" event={"ID":"0ca0de6e-d004-4b46-aee4-66eae716184f","Type":"ContainerDied","Data":"c6ef6f376b3de6a06cb247c19a2dde75e11dca8e32c9db492300d119a97e789b"} Jan 26 16:49:38 crc kubenswrapper[4754]: I0126 16:49:38.669903 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qzw56" event={"ID":"0ca0de6e-d004-4b46-aee4-66eae716184f","Type":"ContainerStarted","Data":"cf4854adfd91381464448a0db0431004cb6acb1b31a263fa88df803c4200c819"} Jan 26 16:49:38 crc kubenswrapper[4754]: I0126 16:49:38.694192 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-9shrr" podStartSLOduration=11.694170984 podStartE2EDuration="11.694170984s" podCreationTimestamp="2026-01-26 16:49:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 16:49:38.690447061 +0000 UTC m=+145.214627515" watchObservedRunningTime="2026-01-26 16:49:38.694170984 +0000 UTC m=+145.218351418" Jan 26 16:49:38 crc kubenswrapper[4754]: I0126 16:49:38.711837 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hts6c\" (UniqueName: \"kubernetes.io/projected/ffaa8507-8be1-4bdb-b7b5-28104f9cdbd6-kube-api-access-hts6c\") pod \"redhat-marketplace-plfzc\" (UID: \"ffaa8507-8be1-4bdb-b7b5-28104f9cdbd6\") " pod="openshift-marketplace/redhat-marketplace-plfzc" Jan 26 16:49:38 crc kubenswrapper[4754]: I0126 16:49:38.788759 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-plfzc" Jan 26 16:49:38 crc kubenswrapper[4754]: I0126 16:49:38.882865 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-hhhn5"] Jan 26 16:49:38 crc kubenswrapper[4754]: I0126 16:49:38.980458 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-47pnn"] Jan 26 16:49:38 crc kubenswrapper[4754]: W0126 16:49:38.998205 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd098371a_9920_44d2_b918_f21c8142fac2.slice/crio-537130dd816455c11cb6fc9665a7a41fe83c79a3a56347ddf6eac6e8f83da252 WatchSource:0}: Error finding container 537130dd816455c11cb6fc9665a7a41fe83c79a3a56347ddf6eac6e8f83da252: Status 404 returned error can't find the container with id 537130dd816455c11cb6fc9665a7a41fe83c79a3a56347ddf6eac6e8f83da252 Jan 26 16:49:39 crc kubenswrapper[4754]: I0126 16:49:39.077105 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-plfzc"] Jan 26 16:49:39 crc kubenswrapper[4754]: I0126 16:49:39.085703 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-5rvw2"] Jan 26 16:49:39 crc kubenswrapper[4754]: I0126 16:49:39.087631 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5rvw2" Jan 26 16:49:39 crc kubenswrapper[4754]: I0126 16:49:39.089474 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-5rvw2"] Jan 26 16:49:39 crc kubenswrapper[4754]: I0126 16:49:39.091408 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Jan 26 16:49:39 crc kubenswrapper[4754]: W0126 16:49:39.098749 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podffaa8507_8be1_4bdb_b7b5_28104f9cdbd6.slice/crio-72a9bd98676797e9a665dda6d8c130093e4dd4b22599b054bf0adc9288a02ae4 WatchSource:0}: Error finding container 72a9bd98676797e9a665dda6d8c130093e4dd4b22599b054bf0adc9288a02ae4: Status 404 returned error can't find the container with id 72a9bd98676797e9a665dda6d8c130093e4dd4b22599b054bf0adc9288a02ae4 Jan 26 16:49:39 crc kubenswrapper[4754]: I0126 16:49:39.169443 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b1b51a00-e6fe-428a-85ef-9d31badc8457-catalog-content\") pod \"redhat-operators-5rvw2\" (UID: \"b1b51a00-e6fe-428a-85ef-9d31badc8457\") " pod="openshift-marketplace/redhat-operators-5rvw2" Jan 26 16:49:39 crc kubenswrapper[4754]: I0126 16:49:39.169520 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8j5j7\" (UniqueName: \"kubernetes.io/projected/b1b51a00-e6fe-428a-85ef-9d31badc8457-kube-api-access-8j5j7\") pod \"redhat-operators-5rvw2\" (UID: \"b1b51a00-e6fe-428a-85ef-9d31badc8457\") " pod="openshift-marketplace/redhat-operators-5rvw2" Jan 26 16:49:39 crc kubenswrapper[4754]: I0126 16:49:39.169891 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b1b51a00-e6fe-428a-85ef-9d31badc8457-utilities\") pod \"redhat-operators-5rvw2\" (UID: \"b1b51a00-e6fe-428a-85ef-9d31badc8457\") " pod="openshift-marketplace/redhat-operators-5rvw2" Jan 26 16:49:39 crc kubenswrapper[4754]: I0126 16:49:39.271543 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b1b51a00-e6fe-428a-85ef-9d31badc8457-catalog-content\") pod \"redhat-operators-5rvw2\" (UID: \"b1b51a00-e6fe-428a-85ef-9d31badc8457\") " pod="openshift-marketplace/redhat-operators-5rvw2" Jan 26 16:49:39 crc kubenswrapper[4754]: I0126 16:49:39.271588 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8j5j7\" (UniqueName: \"kubernetes.io/projected/b1b51a00-e6fe-428a-85ef-9d31badc8457-kube-api-access-8j5j7\") pod \"redhat-operators-5rvw2\" (UID: \"b1b51a00-e6fe-428a-85ef-9d31badc8457\") " pod="openshift-marketplace/redhat-operators-5rvw2" Jan 26 16:49:39 crc kubenswrapper[4754]: I0126 16:49:39.271632 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b1b51a00-e6fe-428a-85ef-9d31badc8457-utilities\") pod \"redhat-operators-5rvw2\" (UID: \"b1b51a00-e6fe-428a-85ef-9d31badc8457\") " pod="openshift-marketplace/redhat-operators-5rvw2" Jan 26 16:49:39 crc kubenswrapper[4754]: I0126 16:49:39.272165 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b1b51a00-e6fe-428a-85ef-9d31badc8457-catalog-content\") pod \"redhat-operators-5rvw2\" (UID: \"b1b51a00-e6fe-428a-85ef-9d31badc8457\") " pod="openshift-marketplace/redhat-operators-5rvw2" Jan 26 16:49:39 crc kubenswrapper[4754]: I0126 16:49:39.272311 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b1b51a00-e6fe-428a-85ef-9d31badc8457-utilities\") pod \"redhat-operators-5rvw2\" (UID: \"b1b51a00-e6fe-428a-85ef-9d31badc8457\") " pod="openshift-marketplace/redhat-operators-5rvw2" Jan 26 16:49:39 crc kubenswrapper[4754]: I0126 16:49:39.292969 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8j5j7\" (UniqueName: \"kubernetes.io/projected/b1b51a00-e6fe-428a-85ef-9d31badc8457-kube-api-access-8j5j7\") pod \"redhat-operators-5rvw2\" (UID: \"b1b51a00-e6fe-428a-85ef-9d31badc8457\") " pod="openshift-marketplace/redhat-operators-5rvw2" Jan 26 16:49:39 crc kubenswrapper[4754]: I0126 16:49:39.299100 4754 patch_prober.go:28] interesting pod/router-default-5444994796-jxzst container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 26 16:49:39 crc kubenswrapper[4754]: [-]has-synced failed: reason withheld Jan 26 16:49:39 crc kubenswrapper[4754]: [+]process-running ok Jan 26 16:49:39 crc kubenswrapper[4754]: healthz check failed Jan 26 16:49:39 crc kubenswrapper[4754]: I0126 16:49:39.299218 4754 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-jxzst" podUID="44e9e245-6889-4706-bc76-a22f2e7f6019" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 26 16:49:39 crc kubenswrapper[4754]: I0126 16:49:39.415020 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5rvw2" Jan 26 16:49:39 crc kubenswrapper[4754]: I0126 16:49:39.455546 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-8d467" Jan 26 16:49:39 crc kubenswrapper[4754]: I0126 16:49:39.455611 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-8d467" Jan 26 16:49:39 crc kubenswrapper[4754]: I0126 16:49:39.461944 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-pdj77"] Jan 26 16:49:39 crc kubenswrapper[4754]: I0126 16:49:39.463024 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pdj77" Jan 26 16:49:39 crc kubenswrapper[4754]: I0126 16:49:39.470035 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-8d467" Jan 26 16:49:39 crc kubenswrapper[4754]: I0126 16:49:39.475947 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-pdj77"] Jan 26 16:49:39 crc kubenswrapper[4754]: I0126 16:49:39.487765 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6mj2d" Jan 26 16:49:39 crc kubenswrapper[4754]: I0126 16:49:39.487800 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6mj2d" Jan 26 16:49:39 crc kubenswrapper[4754]: I0126 16:49:39.495023 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-5p6w2" Jan 26 16:49:39 crc kubenswrapper[4754]: I0126 16:49:39.496198 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-5p6w2" Jan 26 16:49:39 crc kubenswrapper[4754]: I0126 16:49:39.503093 4754 patch_prober.go:28] interesting pod/console-f9d7485db-5p6w2 container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.20:8443/health\": dial tcp 10.217.0.20:8443: connect: connection refused" start-of-body= Jan 26 16:49:39 crc kubenswrapper[4754]: I0126 16:49:39.503141 4754 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-5p6w2" podUID="528dc2c4-71cc-46cb-b847-426bb8638188" containerName="console" probeResult="failure" output="Get \"https://10.217.0.20:8443/health\": dial tcp 10.217.0.20:8443: connect: connection refused" Jan 26 16:49:39 crc kubenswrapper[4754]: I0126 16:49:39.504096 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6mj2d" Jan 26 16:49:39 crc kubenswrapper[4754]: I0126 16:49:39.550302 4754 patch_prober.go:28] interesting pod/downloads-7954f5f757-rcrn5 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Jan 26 16:49:39 crc kubenswrapper[4754]: I0126 16:49:39.550319 4754 patch_prober.go:28] interesting pod/downloads-7954f5f757-rcrn5 container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Jan 26 16:49:39 crc kubenswrapper[4754]: I0126 16:49:39.550361 4754 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-rcrn5" podUID="78242150-7a64-4766-808f-fa460cd84b20" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Jan 26 16:49:39 crc kubenswrapper[4754]: I0126 16:49:39.550384 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-rcrn5" podUID="78242150-7a64-4766-808f-fa460cd84b20" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Jan 26 16:49:39 crc kubenswrapper[4754]: I0126 16:49:39.583311 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ppdns\" (UniqueName: \"kubernetes.io/projected/143cd811-0dff-4141-bbb4-f940c83a0bb0-kube-api-access-ppdns\") pod \"redhat-operators-pdj77\" (UID: \"143cd811-0dff-4141-bbb4-f940c83a0bb0\") " pod="openshift-marketplace/redhat-operators-pdj77" Jan 26 16:49:39 crc kubenswrapper[4754]: I0126 16:49:39.583374 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/143cd811-0dff-4141-bbb4-f940c83a0bb0-catalog-content\") pod \"redhat-operators-pdj77\" (UID: \"143cd811-0dff-4141-bbb4-f940c83a0bb0\") " pod="openshift-marketplace/redhat-operators-pdj77" Jan 26 16:49:39 crc kubenswrapper[4754]: I0126 16:49:39.584272 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/143cd811-0dff-4141-bbb4-f940c83a0bb0-utilities\") pod \"redhat-operators-pdj77\" (UID: \"143cd811-0dff-4141-bbb4-f940c83a0bb0\") " pod="openshift-marketplace/redhat-operators-pdj77" Jan 26 16:49:39 crc kubenswrapper[4754]: I0126 16:49:39.682039 4754 generic.go:334] "Generic (PLEG): container finished" podID="f5fabb97-e818-4c00-8023-db5e8477878f" containerID="e4fb9df6906463bd5ee7e3667dc19c4ad04a1cf49b018f0036e4fbd87047d607" exitCode=0 Jan 26 16:49:39 crc kubenswrapper[4754]: I0126 16:49:39.682094 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29490765-w8zqq" event={"ID":"f5fabb97-e818-4c00-8023-db5e8477878f","Type":"ContainerDied","Data":"e4fb9df6906463bd5ee7e3667dc19c4ad04a1cf49b018f0036e4fbd87047d607"} Jan 26 16:49:39 crc kubenswrapper[4754]: I0126 16:49:39.687313 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/143cd811-0dff-4141-bbb4-f940c83a0bb0-utilities\") pod \"redhat-operators-pdj77\" (UID: \"143cd811-0dff-4141-bbb4-f940c83a0bb0\") " pod="openshift-marketplace/redhat-operators-pdj77" Jan 26 16:49:39 crc kubenswrapper[4754]: I0126 16:49:39.687768 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/143cd811-0dff-4141-bbb4-f940c83a0bb0-utilities\") pod \"redhat-operators-pdj77\" (UID: \"143cd811-0dff-4141-bbb4-f940c83a0bb0\") " pod="openshift-marketplace/redhat-operators-pdj77" Jan 26 16:49:39 crc kubenswrapper[4754]: I0126 16:49:39.687972 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ppdns\" (UniqueName: \"kubernetes.io/projected/143cd811-0dff-4141-bbb4-f940c83a0bb0-kube-api-access-ppdns\") pod \"redhat-operators-pdj77\" (UID: \"143cd811-0dff-4141-bbb4-f940c83a0bb0\") " pod="openshift-marketplace/redhat-operators-pdj77" Jan 26 16:49:39 crc kubenswrapper[4754]: I0126 16:49:39.688040 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/143cd811-0dff-4141-bbb4-f940c83a0bb0-catalog-content\") pod \"redhat-operators-pdj77\" (UID: \"143cd811-0dff-4141-bbb4-f940c83a0bb0\") " pod="openshift-marketplace/redhat-operators-pdj77" Jan 26 16:49:39 crc kubenswrapper[4754]: I0126 16:49:39.688302 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/143cd811-0dff-4141-bbb4-f940c83a0bb0-catalog-content\") pod \"redhat-operators-pdj77\" (UID: \"143cd811-0dff-4141-bbb4-f940c83a0bb0\") " pod="openshift-marketplace/redhat-operators-pdj77" Jan 26 16:49:39 crc kubenswrapper[4754]: I0126 16:49:39.692012 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-47pnn" event={"ID":"d098371a-9920-44d2-b918-f21c8142fac2","Type":"ContainerStarted","Data":"d8f5c5806d081e327104a1e858d04ead57c27dca4287c0e64499580aaaf7f79e"} Jan 26 16:49:39 crc kubenswrapper[4754]: I0126 16:49:39.692047 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-47pnn" event={"ID":"d098371a-9920-44d2-b918-f21c8142fac2","Type":"ContainerStarted","Data":"537130dd816455c11cb6fc9665a7a41fe83c79a3a56347ddf6eac6e8f83da252"} Jan 26 16:49:39 crc kubenswrapper[4754]: I0126 16:49:39.692212 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-47pnn" Jan 26 16:49:39 crc kubenswrapper[4754]: I0126 16:49:39.694281 4754 generic.go:334] "Generic (PLEG): container finished" podID="ffaa8507-8be1-4bdb-b7b5-28104f9cdbd6" containerID="f5ad89d1e92c1dc4aa297ec8985614cde8f91e56458d1c6b3e6db70a8efdf4ac" exitCode=0 Jan 26 16:49:39 crc kubenswrapper[4754]: I0126 16:49:39.694344 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-plfzc" event={"ID":"ffaa8507-8be1-4bdb-b7b5-28104f9cdbd6","Type":"ContainerDied","Data":"f5ad89d1e92c1dc4aa297ec8985614cde8f91e56458d1c6b3e6db70a8efdf4ac"} Jan 26 16:49:39 crc kubenswrapper[4754]: I0126 16:49:39.694363 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-plfzc" event={"ID":"ffaa8507-8be1-4bdb-b7b5-28104f9cdbd6","Type":"ContainerStarted","Data":"72a9bd98676797e9a665dda6d8c130093e4dd4b22599b054bf0adc9288a02ae4"} Jan 26 16:49:39 crc kubenswrapper[4754]: I0126 16:49:39.720899 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ppdns\" (UniqueName: \"kubernetes.io/projected/143cd811-0dff-4141-bbb4-f940c83a0bb0-kube-api-access-ppdns\") pod \"redhat-operators-pdj77\" (UID: \"143cd811-0dff-4141-bbb4-f940c83a0bb0\") " pod="openshift-marketplace/redhat-operators-pdj77" Jan 26 16:49:39 crc kubenswrapper[4754]: I0126 16:49:39.734649 4754 generic.go:334] "Generic (PLEG): container finished" podID="e36dc2d0-47de-46a6-aeba-52475754867e" containerID="fc7875a2ec569f0ebeb226ca273f9f8ee7f570b1b8622e77565446d2a859eaa3" exitCode=0 Jan 26 16:49:39 crc kubenswrapper[4754]: I0126 16:49:39.735955 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hhhn5" event={"ID":"e36dc2d0-47de-46a6-aeba-52475754867e","Type":"ContainerDied","Data":"fc7875a2ec569f0ebeb226ca273f9f8ee7f570b1b8622e77565446d2a859eaa3"} Jan 26 16:49:39 crc kubenswrapper[4754]: I0126 16:49:39.735985 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hhhn5" event={"ID":"e36dc2d0-47de-46a6-aeba-52475754867e","Type":"ContainerStarted","Data":"fc6f5c5f1c7db70e9c330d4135008ece5d08f6a86648d7e02cda040ca7905d48"} Jan 26 16:49:39 crc kubenswrapper[4754]: I0126 16:49:39.741168 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Jan 26 16:49:39 crc kubenswrapper[4754]: I0126 16:49:39.742147 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 26 16:49:39 crc kubenswrapper[4754]: I0126 16:49:39.747847 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Jan 26 16:49:39 crc kubenswrapper[4754]: I0126 16:49:39.747916 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Jan 26 16:49:39 crc kubenswrapper[4754]: I0126 16:49:39.759796 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-8d467" Jan 26 16:49:39 crc kubenswrapper[4754]: I0126 16:49:39.763890 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6mj2d" Jan 26 16:49:39 crc kubenswrapper[4754]: I0126 16:49:39.766002 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Jan 26 16:49:39 crc kubenswrapper[4754]: I0126 16:49:39.769240 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-47pnn" podStartSLOduration=126.769225774 podStartE2EDuration="2m6.769225774s" podCreationTimestamp="2026-01-26 16:47:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 16:49:39.758319803 +0000 UTC m=+146.282500237" watchObservedRunningTime="2026-01-26 16:49:39.769225774 +0000 UTC m=+146.293406208" Jan 26 16:49:39 crc kubenswrapper[4754]: I0126 16:49:39.789657 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/71b1f977-1a94-4e7f-bcc1-7a6ec55d043f-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"71b1f977-1a94-4e7f-bcc1-7a6ec55d043f\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 26 16:49:39 crc kubenswrapper[4754]: I0126 16:49:39.789823 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/71b1f977-1a94-4e7f-bcc1-7a6ec55d043f-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"71b1f977-1a94-4e7f-bcc1-7a6ec55d043f\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 26 16:49:39 crc kubenswrapper[4754]: I0126 16:49:39.795640 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pdj77" Jan 26 16:49:39 crc kubenswrapper[4754]: W0126 16:49:39.814025 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb1b51a00_e6fe_428a_85ef_9d31badc8457.slice/crio-d16b4b2bb0d8fe98b530a4f93ec01d1a0c5511b630a321508338e6bb65347d6b WatchSource:0}: Error finding container d16b4b2bb0d8fe98b530a4f93ec01d1a0c5511b630a321508338e6bb65347d6b: Status 404 returned error can't find the container with id d16b4b2bb0d8fe98b530a4f93ec01d1a0c5511b630a321508338e6bb65347d6b Jan 26 16:49:39 crc kubenswrapper[4754]: I0126 16:49:39.816850 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Jan 26 16:49:39 crc kubenswrapper[4754]: I0126 16:49:39.817292 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-5rvw2"] Jan 26 16:49:39 crc kubenswrapper[4754]: I0126 16:49:39.891021 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 16:49:39 crc kubenswrapper[4754]: I0126 16:49:39.891099 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/71b1f977-1a94-4e7f-bcc1-7a6ec55d043f-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"71b1f977-1a94-4e7f-bcc1-7a6ec55d043f\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 26 16:49:39 crc kubenswrapper[4754]: I0126 16:49:39.891127 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 16:49:39 crc kubenswrapper[4754]: I0126 16:49:39.891181 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 16:49:39 crc kubenswrapper[4754]: I0126 16:49:39.891206 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 16:49:39 crc kubenswrapper[4754]: I0126 16:49:39.891263 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/71b1f977-1a94-4e7f-bcc1-7a6ec55d043f-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"71b1f977-1a94-4e7f-bcc1-7a6ec55d043f\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 26 16:49:39 crc kubenswrapper[4754]: I0126 16:49:39.892320 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/71b1f977-1a94-4e7f-bcc1-7a6ec55d043f-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"71b1f977-1a94-4e7f-bcc1-7a6ec55d043f\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 26 16:49:39 crc kubenswrapper[4754]: I0126 16:49:39.892923 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 16:49:39 crc kubenswrapper[4754]: I0126 16:49:39.907398 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 16:49:39 crc kubenswrapper[4754]: I0126 16:49:39.907580 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 16:49:39 crc kubenswrapper[4754]: I0126 16:49:39.911781 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 16:49:39 crc kubenswrapper[4754]: I0126 16:49:39.914000 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 16:49:39 crc kubenswrapper[4754]: I0126 16:49:39.914315 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/71b1f977-1a94-4e7f-bcc1-7a6ec55d043f-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"71b1f977-1a94-4e7f-bcc1-7a6ec55d043f\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 26 16:49:40 crc kubenswrapper[4754]: I0126 16:49:40.118256 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 26 16:49:40 crc kubenswrapper[4754]: I0126 16:49:40.194651 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 16:49:40 crc kubenswrapper[4754]: I0126 16:49:40.205086 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-pdj77"] Jan 26 16:49:40 crc kubenswrapper[4754]: I0126 16:49:40.209272 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 16:49:40 crc kubenswrapper[4754]: I0126 16:49:40.294010 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-jxzst" Jan 26 16:49:40 crc kubenswrapper[4754]: I0126 16:49:40.301094 4754 patch_prober.go:28] interesting pod/router-default-5444994796-jxzst container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 26 16:49:40 crc kubenswrapper[4754]: [-]has-synced failed: reason withheld Jan 26 16:49:40 crc kubenswrapper[4754]: [+]process-running ok Jan 26 16:49:40 crc kubenswrapper[4754]: healthz check failed Jan 26 16:49:40 crc kubenswrapper[4754]: I0126 16:49:40.301139 4754 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-jxzst" podUID="44e9e245-6889-4706-bc76-a22f2e7f6019" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 26 16:49:40 crc kubenswrapper[4754]: I0126 16:49:40.344071 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-7xsg2" Jan 26 16:49:40 crc kubenswrapper[4754]: W0126 16:49:40.415362 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9d751cbb_f2e2_430d_9754_c882a5e924a5.slice/crio-4c6c3311fdb9ac3943304e102fd58c6137281d6d6ccac9f49e3ee57c6a08e5ef WatchSource:0}: Error finding container 4c6c3311fdb9ac3943304e102fd58c6137281d6d6ccac9f49e3ee57c6a08e5ef: Status 404 returned error can't find the container with id 4c6c3311fdb9ac3943304e102fd58c6137281d6d6ccac9f49e3ee57c6a08e5ef Jan 26 16:49:40 crc kubenswrapper[4754]: I0126 16:49:40.747503 4754 generic.go:334] "Generic (PLEG): container finished" podID="b1b51a00-e6fe-428a-85ef-9d31badc8457" containerID="9960a5f602fda10c53a46543fdf587ea8c2512b205a707973e1f2eb83783eb4e" exitCode=0 Jan 26 16:49:40 crc kubenswrapper[4754]: I0126 16:49:40.748638 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5rvw2" event={"ID":"b1b51a00-e6fe-428a-85ef-9d31badc8457","Type":"ContainerDied","Data":"9960a5f602fda10c53a46543fdf587ea8c2512b205a707973e1f2eb83783eb4e"} Jan 26 16:49:40 crc kubenswrapper[4754]: I0126 16:49:40.748671 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5rvw2" event={"ID":"b1b51a00-e6fe-428a-85ef-9d31badc8457","Type":"ContainerStarted","Data":"d16b4b2bb0d8fe98b530a4f93ec01d1a0c5511b630a321508338e6bb65347d6b"} Jan 26 16:49:40 crc kubenswrapper[4754]: I0126 16:49:40.751999 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"ba5611a0352b2399f1a854a0aae0a0ac6e3e9b8776bb0b6f2daf01f1130d24a1"} Jan 26 16:49:40 crc kubenswrapper[4754]: I0126 16:49:40.752036 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"4c6c3311fdb9ac3943304e102fd58c6137281d6d6ccac9f49e3ee57c6a08e5ef"} Jan 26 16:49:40 crc kubenswrapper[4754]: I0126 16:49:40.755174 4754 generic.go:334] "Generic (PLEG): container finished" podID="143cd811-0dff-4141-bbb4-f940c83a0bb0" containerID="2f98d952e095831b8788d75c46f17948c983ef880a4fc6066e33ac5984405790" exitCode=0 Jan 26 16:49:40 crc kubenswrapper[4754]: I0126 16:49:40.756324 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pdj77" event={"ID":"143cd811-0dff-4141-bbb4-f940c83a0bb0","Type":"ContainerDied","Data":"2f98d952e095831b8788d75c46f17948c983ef880a4fc6066e33ac5984405790"} Jan 26 16:49:40 crc kubenswrapper[4754]: I0126 16:49:40.756352 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pdj77" event={"ID":"143cd811-0dff-4141-bbb4-f940c83a0bb0","Type":"ContainerStarted","Data":"00d039400512dc3846805a224b072eca6624c1f0ca90c23bbeb2ddd01b9e2f2f"} Jan 26 16:49:40 crc kubenswrapper[4754]: I0126 16:49:40.833377 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Jan 26 16:49:40 crc kubenswrapper[4754]: W0126 16:49:40.854418 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod71b1f977_1a94_4e7f_bcc1_7a6ec55d043f.slice/crio-dcb7e79580fb5a7b513753b873d93c92764be4fb3ec1a285d182d3445fbfad51 WatchSource:0}: Error finding container dcb7e79580fb5a7b513753b873d93c92764be4fb3ec1a285d182d3445fbfad51: Status 404 returned error can't find the container with id dcb7e79580fb5a7b513753b873d93c92764be4fb3ec1a285d182d3445fbfad51 Jan 26 16:49:40 crc kubenswrapper[4754]: W0126 16:49:40.919235 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5fe485a1_e14f_4c09_b5b9_f252bc42b7e8.slice/crio-74aec8a1fc03e596bb5e28f5cd152de04e67f40bc0ca3a779cb1176007e122ea WatchSource:0}: Error finding container 74aec8a1fc03e596bb5e28f5cd152de04e67f40bc0ca3a779cb1176007e122ea: Status 404 returned error can't find the container with id 74aec8a1fc03e596bb5e28f5cd152de04e67f40bc0ca3a779cb1176007e122ea Jan 26 16:49:40 crc kubenswrapper[4754]: W0126 16:49:40.921955 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3b6479f0_333b_4a96_9adf_2099afdc2447.slice/crio-9a9d44c34877b1518005d45adc779d5097c0700aa1b0f2c0fcc1e40555e432d2 WatchSource:0}: Error finding container 9a9d44c34877b1518005d45adc779d5097c0700aa1b0f2c0fcc1e40555e432d2: Status 404 returned error can't find the container with id 9a9d44c34877b1518005d45adc779d5097c0700aa1b0f2c0fcc1e40555e432d2 Jan 26 16:49:41 crc kubenswrapper[4754]: I0126 16:49:41.019944 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29490765-w8zqq" Jan 26 16:49:41 crc kubenswrapper[4754]: I0126 16:49:41.031611 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f5fabb97-e818-4c00-8023-db5e8477878f-config-volume\") pod \"f5fabb97-e818-4c00-8023-db5e8477878f\" (UID: \"f5fabb97-e818-4c00-8023-db5e8477878f\") " Jan 26 16:49:41 crc kubenswrapper[4754]: I0126 16:49:41.032711 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f5fabb97-e818-4c00-8023-db5e8477878f-config-volume" (OuterVolumeSpecName: "config-volume") pod "f5fabb97-e818-4c00-8023-db5e8477878f" (UID: "f5fabb97-e818-4c00-8023-db5e8477878f"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 16:49:41 crc kubenswrapper[4754]: I0126 16:49:41.132906 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mt7kw\" (UniqueName: \"kubernetes.io/projected/f5fabb97-e818-4c00-8023-db5e8477878f-kube-api-access-mt7kw\") pod \"f5fabb97-e818-4c00-8023-db5e8477878f\" (UID: \"f5fabb97-e818-4c00-8023-db5e8477878f\") " Jan 26 16:49:41 crc kubenswrapper[4754]: I0126 16:49:41.133013 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f5fabb97-e818-4c00-8023-db5e8477878f-secret-volume\") pod \"f5fabb97-e818-4c00-8023-db5e8477878f\" (UID: \"f5fabb97-e818-4c00-8023-db5e8477878f\") " Jan 26 16:49:41 crc kubenswrapper[4754]: I0126 16:49:41.133549 4754 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f5fabb97-e818-4c00-8023-db5e8477878f-config-volume\") on node \"crc\" DevicePath \"\"" Jan 26 16:49:41 crc kubenswrapper[4754]: I0126 16:49:41.137561 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f5fabb97-e818-4c00-8023-db5e8477878f-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "f5fabb97-e818-4c00-8023-db5e8477878f" (UID: "f5fabb97-e818-4c00-8023-db5e8477878f"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 16:49:41 crc kubenswrapper[4754]: I0126 16:49:41.138394 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f5fabb97-e818-4c00-8023-db5e8477878f-kube-api-access-mt7kw" (OuterVolumeSpecName: "kube-api-access-mt7kw") pod "f5fabb97-e818-4c00-8023-db5e8477878f" (UID: "f5fabb97-e818-4c00-8023-db5e8477878f"). InnerVolumeSpecName "kube-api-access-mt7kw". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:49:41 crc kubenswrapper[4754]: I0126 16:49:41.235750 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mt7kw\" (UniqueName: \"kubernetes.io/projected/f5fabb97-e818-4c00-8023-db5e8477878f-kube-api-access-mt7kw\") on node \"crc\" DevicePath \"\"" Jan 26 16:49:41 crc kubenswrapper[4754]: I0126 16:49:41.235792 4754 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f5fabb97-e818-4c00-8023-db5e8477878f-secret-volume\") on node \"crc\" DevicePath \"\"" Jan 26 16:49:41 crc kubenswrapper[4754]: I0126 16:49:41.297128 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-jxzst" Jan 26 16:49:41 crc kubenswrapper[4754]: I0126 16:49:41.299452 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-jxzst" Jan 26 16:49:41 crc kubenswrapper[4754]: I0126 16:49:41.798044 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"0dabd08dbebdc669147abd135e153aa012ffd33bc9785b0264f734efacb2d4b0"} Jan 26 16:49:41 crc kubenswrapper[4754]: I0126 16:49:41.799802 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"9a9d44c34877b1518005d45adc779d5097c0700aa1b0f2c0fcc1e40555e432d2"} Jan 26 16:49:41 crc kubenswrapper[4754]: I0126 16:49:41.800192 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 16:49:41 crc kubenswrapper[4754]: I0126 16:49:41.813423 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"4db2bbe1142c6e858a3423cc08dae4534e8ec2ec3e600e4bad5a85737ad738b7"} Jan 26 16:49:41 crc kubenswrapper[4754]: I0126 16:49:41.813477 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"74aec8a1fc03e596bb5e28f5cd152de04e67f40bc0ca3a779cb1176007e122ea"} Jan 26 16:49:41 crc kubenswrapper[4754]: I0126 16:49:41.818184 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"71b1f977-1a94-4e7f-bcc1-7a6ec55d043f","Type":"ContainerStarted","Data":"dcb7e79580fb5a7b513753b873d93c92764be4fb3ec1a285d182d3445fbfad51"} Jan 26 16:49:41 crc kubenswrapper[4754]: I0126 16:49:41.820120 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29490765-w8zqq" event={"ID":"f5fabb97-e818-4c00-8023-db5e8477878f","Type":"ContainerDied","Data":"a31fd3a98cc51d16587d5922ed23a21a41a4042ccf85c441b26d01ca1ee5c458"} Jan 26 16:49:41 crc kubenswrapper[4754]: I0126 16:49:41.820120 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29490765-w8zqq" Jan 26 16:49:41 crc kubenswrapper[4754]: I0126 16:49:41.820143 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a31fd3a98cc51d16587d5922ed23a21a41a4042ccf85c441b26d01ca1ee5c458" Jan 26 16:49:42 crc kubenswrapper[4754]: I0126 16:49:42.829108 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"71b1f977-1a94-4e7f-bcc1-7a6ec55d043f","Type":"ContainerStarted","Data":"c57cd64469341cd9fef85ce6d10657d7cf4eaef50167316ac09c4d508161e344"} Jan 26 16:49:43 crc kubenswrapper[4754]: I0126 16:49:43.838980 4754 generic.go:334] "Generic (PLEG): container finished" podID="71b1f977-1a94-4e7f-bcc1-7a6ec55d043f" containerID="c57cd64469341cd9fef85ce6d10657d7cf4eaef50167316ac09c4d508161e344" exitCode=0 Jan 26 16:49:43 crc kubenswrapper[4754]: I0126 16:49:43.839021 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"71b1f977-1a94-4e7f-bcc1-7a6ec55d043f","Type":"ContainerDied","Data":"c57cd64469341cd9fef85ce6d10657d7cf4eaef50167316ac09c4d508161e344"} Jan 26 16:49:44 crc kubenswrapper[4754]: I0126 16:49:44.414335 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Jan 26 16:49:44 crc kubenswrapper[4754]: E0126 16:49:44.414910 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5fabb97-e818-4c00-8023-db5e8477878f" containerName="collect-profiles" Jan 26 16:49:44 crc kubenswrapper[4754]: I0126 16:49:44.414933 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5fabb97-e818-4c00-8023-db5e8477878f" containerName="collect-profiles" Jan 26 16:49:44 crc kubenswrapper[4754]: I0126 16:49:44.415068 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="f5fabb97-e818-4c00-8023-db5e8477878f" containerName="collect-profiles" Jan 26 16:49:44 crc kubenswrapper[4754]: I0126 16:49:44.415551 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 26 16:49:44 crc kubenswrapper[4754]: I0126 16:49:44.422574 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Jan 26 16:49:44 crc kubenswrapper[4754]: I0126 16:49:44.424017 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Jan 26 16:49:44 crc kubenswrapper[4754]: I0126 16:49:44.450293 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Jan 26 16:49:44 crc kubenswrapper[4754]: I0126 16:49:44.595182 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/99f61c83-b9c7-4559-9eaf-1c08e38244f0-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"99f61c83-b9c7-4559-9eaf-1c08e38244f0\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 26 16:49:44 crc kubenswrapper[4754]: I0126 16:49:44.595344 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/99f61c83-b9c7-4559-9eaf-1c08e38244f0-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"99f61c83-b9c7-4559-9eaf-1c08e38244f0\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 26 16:49:44 crc kubenswrapper[4754]: I0126 16:49:44.699130 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/99f61c83-b9c7-4559-9eaf-1c08e38244f0-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"99f61c83-b9c7-4559-9eaf-1c08e38244f0\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 26 16:49:44 crc kubenswrapper[4754]: I0126 16:49:44.699311 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/99f61c83-b9c7-4559-9eaf-1c08e38244f0-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"99f61c83-b9c7-4559-9eaf-1c08e38244f0\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 26 16:49:44 crc kubenswrapper[4754]: I0126 16:49:44.699825 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/99f61c83-b9c7-4559-9eaf-1c08e38244f0-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"99f61c83-b9c7-4559-9eaf-1c08e38244f0\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 26 16:49:44 crc kubenswrapper[4754]: I0126 16:49:44.730458 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/99f61c83-b9c7-4559-9eaf-1c08e38244f0-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"99f61c83-b9c7-4559-9eaf-1c08e38244f0\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 26 16:49:44 crc kubenswrapper[4754]: I0126 16:49:44.740100 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 26 16:49:45 crc kubenswrapper[4754]: I0126 16:49:45.376755 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Jan 26 16:49:45 crc kubenswrapper[4754]: I0126 16:49:45.403142 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-964q6" Jan 26 16:49:49 crc kubenswrapper[4754]: I0126 16:49:49.499893 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-5p6w2" Jan 26 16:49:49 crc kubenswrapper[4754]: I0126 16:49:49.504582 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-5p6w2" Jan 26 16:49:49 crc kubenswrapper[4754]: I0126 16:49:49.563701 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-rcrn5" Jan 26 16:49:54 crc kubenswrapper[4754]: I0126 16:49:54.819795 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 26 16:49:54 crc kubenswrapper[4754]: I0126 16:49:54.876488 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/71b1f977-1a94-4e7f-bcc1-7a6ec55d043f-kubelet-dir\") pod \"71b1f977-1a94-4e7f-bcc1-7a6ec55d043f\" (UID: \"71b1f977-1a94-4e7f-bcc1-7a6ec55d043f\") " Jan 26 16:49:54 crc kubenswrapper[4754]: I0126 16:49:54.876593 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/71b1f977-1a94-4e7f-bcc1-7a6ec55d043f-kube-api-access\") pod \"71b1f977-1a94-4e7f-bcc1-7a6ec55d043f\" (UID: \"71b1f977-1a94-4e7f-bcc1-7a6ec55d043f\") " Jan 26 16:49:54 crc kubenswrapper[4754]: I0126 16:49:54.876762 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/71b1f977-1a94-4e7f-bcc1-7a6ec55d043f-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "71b1f977-1a94-4e7f-bcc1-7a6ec55d043f" (UID: "71b1f977-1a94-4e7f-bcc1-7a6ec55d043f"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 16:49:54 crc kubenswrapper[4754]: I0126 16:49:54.877566 4754 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/71b1f977-1a94-4e7f-bcc1-7a6ec55d043f-kubelet-dir\") on node \"crc\" DevicePath \"\"" Jan 26 16:49:54 crc kubenswrapper[4754]: I0126 16:49:54.881473 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/71b1f977-1a94-4e7f-bcc1-7a6ec55d043f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "71b1f977-1a94-4e7f-bcc1-7a6ec55d043f" (UID: "71b1f977-1a94-4e7f-bcc1-7a6ec55d043f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:49:54 crc kubenswrapper[4754]: I0126 16:49:54.972018 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"71b1f977-1a94-4e7f-bcc1-7a6ec55d043f","Type":"ContainerDied","Data":"dcb7e79580fb5a7b513753b873d93c92764be4fb3ec1a285d182d3445fbfad51"} Jan 26 16:49:54 crc kubenswrapper[4754]: I0126 16:49:54.972322 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dcb7e79580fb5a7b513753b873d93c92764be4fb3ec1a285d182d3445fbfad51" Jan 26 16:49:54 crc kubenswrapper[4754]: I0126 16:49:54.972107 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 26 16:49:54 crc kubenswrapper[4754]: I0126 16:49:54.978345 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/71b1f977-1a94-4e7f-bcc1-7a6ec55d043f-kube-api-access\") on node \"crc\" DevicePath \"\"" Jan 26 16:49:56 crc kubenswrapper[4754]: I0126 16:49:56.496573 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d45179f0-9f81-4d77-8805-6b05bdb32f8f-metrics-certs\") pod \"network-metrics-daemon-t4ps6\" (UID: \"d45179f0-9f81-4d77-8805-6b05bdb32f8f\") " pod="openshift-multus/network-metrics-daemon-t4ps6" Jan 26 16:49:56 crc kubenswrapper[4754]: I0126 16:49:56.501166 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d45179f0-9f81-4d77-8805-6b05bdb32f8f-metrics-certs\") pod \"network-metrics-daemon-t4ps6\" (UID: \"d45179f0-9f81-4d77-8805-6b05bdb32f8f\") " pod="openshift-multus/network-metrics-daemon-t4ps6" Jan 26 16:49:56 crc kubenswrapper[4754]: I0126 16:49:56.684364 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-t4ps6" Jan 26 16:49:57 crc kubenswrapper[4754]: W0126 16:49:57.445884 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod99f61c83_b9c7_4559_9eaf_1c08e38244f0.slice/crio-ca937f87082dd69051cd8b845636584e9c95f4902f9570697e4a1d3c964e2f1b WatchSource:0}: Error finding container ca937f87082dd69051cd8b845636584e9c95f4902f9570697e4a1d3c964e2f1b: Status 404 returned error can't find the container with id ca937f87082dd69051cd8b845636584e9c95f4902f9570697e4a1d3c964e2f1b Jan 26 16:49:58 crc kubenswrapper[4754]: I0126 16:49:58.000007 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"99f61c83-b9c7-4559-9eaf-1c08e38244f0","Type":"ContainerStarted","Data":"ca937f87082dd69051cd8b845636584e9c95f4902f9570697e4a1d3c964e2f1b"} Jan 26 16:49:58 crc kubenswrapper[4754]: I0126 16:49:58.568418 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-47pnn" Jan 26 16:50:07 crc kubenswrapper[4754]: I0126 16:50:07.129471 4754 patch_prober.go:28] interesting pod/machine-config-daemon-x65wv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 26 16:50:07 crc kubenswrapper[4754]: I0126 16:50:07.130104 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 26 16:50:10 crc kubenswrapper[4754]: I0126 16:50:10.226509 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-95dbg" Jan 26 16:50:10 crc kubenswrapper[4754]: E0126 16:50:10.727322 4754 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Jan 26 16:50:10 crc kubenswrapper[4754]: E0126 16:50:10.727534 4754 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-8wwwp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-pdrzp_openshift-marketplace(00549dbb-6196-4495-9fbe-4f641333fbe4): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Jan 26 16:50:10 crc kubenswrapper[4754]: E0126 16:50:10.730007 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-pdrzp" podUID="00549dbb-6196-4495-9fbe-4f641333fbe4" Jan 26 16:50:11 crc kubenswrapper[4754]: E0126 16:50:11.698098 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-pdrzp" podUID="00549dbb-6196-4495-9fbe-4f641333fbe4" Jan 26 16:50:11 crc kubenswrapper[4754]: E0126 16:50:11.762019 4754 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Jan 26 16:50:11 crc kubenswrapper[4754]: E0126 16:50:11.762202 4754 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-z7pgd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-cgm62_openshift-marketplace(c2962a85-523e-4893-8a19-b7baa1360c01): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Jan 26 16:50:11 crc kubenswrapper[4754]: E0126 16:50:11.763385 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-cgm62" podUID="c2962a85-523e-4893-8a19-b7baa1360c01" Jan 26 16:50:11 crc kubenswrapper[4754]: I0126 16:50:11.816650 4754 pod_container_manager_linux.go:210] "Failed to delete cgroup paths" cgroupName=["kubepods","burstable","podf5fabb97-e818-4c00-8023-db5e8477878f"] err="unable to destroy cgroup paths for cgroup [kubepods burstable podf5fabb97-e818-4c00-8023-db5e8477878f] : Timed out while waiting for systemd to remove kubepods-burstable-podf5fabb97_e818_4c00_8023_db5e8477878f.slice" Jan 26 16:50:12 crc kubenswrapper[4754]: E0126 16:50:12.893997 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-cgm62" podUID="c2962a85-523e-4893-8a19-b7baa1360c01" Jan 26 16:50:12 crc kubenswrapper[4754]: E0126 16:50:12.958725 4754 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Jan 26 16:50:12 crc kubenswrapper[4754]: E0126 16:50:12.958868 4754 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-d55d7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-hhhn5_openshift-marketplace(e36dc2d0-47de-46a6-aeba-52475754867e): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Jan 26 16:50:12 crc kubenswrapper[4754]: E0126 16:50:12.960039 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-hhhn5" podUID="e36dc2d0-47de-46a6-aeba-52475754867e" Jan 26 16:50:13 crc kubenswrapper[4754]: E0126 16:50:13.007005 4754 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Jan 26 16:50:13 crc kubenswrapper[4754]: E0126 16:50:13.007168 4754 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-xwf62,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-qzw56_openshift-marketplace(0ca0de6e-d004-4b46-aee4-66eae716184f): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Jan 26 16:50:13 crc kubenswrapper[4754]: E0126 16:50:13.008947 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-qzw56" podUID="0ca0de6e-d004-4b46-aee4-66eae716184f" Jan 26 16:50:13 crc kubenswrapper[4754]: E0126 16:50:13.045816 4754 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Jan 26 16:50:13 crc kubenswrapper[4754]: E0126 16:50:13.046139 4754 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-hts6c,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-plfzc_openshift-marketplace(ffaa8507-8be1-4bdb-b7b5-28104f9cdbd6): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Jan 26 16:50:13 crc kubenswrapper[4754]: E0126 16:50:13.047569 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-plfzc" podUID="ffaa8507-8be1-4bdb-b7b5-28104f9cdbd6" Jan 26 16:50:13 crc kubenswrapper[4754]: E0126 16:50:13.067286 4754 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Jan 26 16:50:13 crc kubenswrapper[4754]: E0126 16:50:13.067529 4754 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-4n544,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-n6gc5_openshift-marketplace(405f8b06-885b-4a7f-9f85-3af411a5504e): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Jan 26 16:50:13 crc kubenswrapper[4754]: E0126 16:50:13.068929 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-n6gc5" podUID="405f8b06-885b-4a7f-9f85-3af411a5504e" Jan 26 16:50:16 crc kubenswrapper[4754]: E0126 16:50:16.618641 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-n6gc5" podUID="405f8b06-885b-4a7f-9f85-3af411a5504e" Jan 26 16:50:16 crc kubenswrapper[4754]: E0126 16:50:16.618728 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-qzw56" podUID="0ca0de6e-d004-4b46-aee4-66eae716184f" Jan 26 16:50:16 crc kubenswrapper[4754]: E0126 16:50:16.618714 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-hhhn5" podUID="e36dc2d0-47de-46a6-aeba-52475754867e" Jan 26 16:50:16 crc kubenswrapper[4754]: E0126 16:50:16.618743 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-plfzc" podUID="ffaa8507-8be1-4bdb-b7b5-28104f9cdbd6" Jan 26 16:50:16 crc kubenswrapper[4754]: E0126 16:50:16.689012 4754 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Jan 26 16:50:16 crc kubenswrapper[4754]: E0126 16:50:16.689233 4754 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-8j5j7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-5rvw2_openshift-marketplace(b1b51a00-e6fe-428a-85ef-9d31badc8457): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Jan 26 16:50:16 crc kubenswrapper[4754]: E0126 16:50:16.690652 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-5rvw2" podUID="b1b51a00-e6fe-428a-85ef-9d31badc8457" Jan 26 16:50:16 crc kubenswrapper[4754]: E0126 16:50:16.709487 4754 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Jan 26 16:50:16 crc kubenswrapper[4754]: E0126 16:50:16.709726 4754 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-ppdns,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-pdj77_openshift-marketplace(143cd811-0dff-4141-bbb4-f940c83a0bb0): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Jan 26 16:50:16 crc kubenswrapper[4754]: E0126 16:50:16.711083 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-pdj77" podUID="143cd811-0dff-4141-bbb4-f940c83a0bb0" Jan 26 16:50:17 crc kubenswrapper[4754]: I0126 16:50:17.051398 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-t4ps6"] Jan 26 16:50:17 crc kubenswrapper[4754]: W0126 16:50:17.065007 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd45179f0_9f81_4d77_8805_6b05bdb32f8f.slice/crio-6d98ecad58070b8f7373ecadb576ac9c9ae9d6d4b4cd620fbe609d7b48896670 WatchSource:0}: Error finding container 6d98ecad58070b8f7373ecadb576ac9c9ae9d6d4b4cd620fbe609d7b48896670: Status 404 returned error can't find the container with id 6d98ecad58070b8f7373ecadb576ac9c9ae9d6d4b4cd620fbe609d7b48896670 Jan 26 16:50:17 crc kubenswrapper[4754]: I0126 16:50:17.133762 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-t4ps6" event={"ID":"d45179f0-9f81-4d77-8805-6b05bdb32f8f","Type":"ContainerStarted","Data":"6d98ecad58070b8f7373ecadb576ac9c9ae9d6d4b4cd620fbe609d7b48896670"} Jan 26 16:50:17 crc kubenswrapper[4754]: I0126 16:50:17.135509 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"99f61c83-b9c7-4559-9eaf-1c08e38244f0","Type":"ContainerStarted","Data":"776255005da2d97a30205602a82431003af5c79ef9c4ddb4343fb96cf55d9b01"} Jan 26 16:50:17 crc kubenswrapper[4754]: E0126 16:50:17.137059 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-5rvw2" podUID="b1b51a00-e6fe-428a-85ef-9d31badc8457" Jan 26 16:50:17 crc kubenswrapper[4754]: E0126 16:50:17.137797 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-pdj77" podUID="143cd811-0dff-4141-bbb4-f940c83a0bb0" Jan 26 16:50:17 crc kubenswrapper[4754]: I0126 16:50:17.192380 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-8-crc" podStartSLOduration=33.192354287 podStartE2EDuration="33.192354287s" podCreationTimestamp="2026-01-26 16:49:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 16:50:17.189965502 +0000 UTC m=+183.714145936" watchObservedRunningTime="2026-01-26 16:50:17.192354287 +0000 UTC m=+183.716534731" Jan 26 16:50:18 crc kubenswrapper[4754]: I0126 16:50:18.153893 4754 generic.go:334] "Generic (PLEG): container finished" podID="99f61c83-b9c7-4559-9eaf-1c08e38244f0" containerID="776255005da2d97a30205602a82431003af5c79ef9c4ddb4343fb96cf55d9b01" exitCode=0 Jan 26 16:50:18 crc kubenswrapper[4754]: I0126 16:50:18.153992 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"99f61c83-b9c7-4559-9eaf-1c08e38244f0","Type":"ContainerDied","Data":"776255005da2d97a30205602a82431003af5c79ef9c4ddb4343fb96cf55d9b01"} Jan 26 16:50:18 crc kubenswrapper[4754]: I0126 16:50:18.166772 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-t4ps6" event={"ID":"d45179f0-9f81-4d77-8805-6b05bdb32f8f","Type":"ContainerStarted","Data":"2a86ed1a4cb8130f2c665681fe5e8850c0c9140c47b64f485262fab7b4d7204e"} Jan 26 16:50:18 crc kubenswrapper[4754]: I0126 16:50:18.166822 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-t4ps6" event={"ID":"d45179f0-9f81-4d77-8805-6b05bdb32f8f","Type":"ContainerStarted","Data":"e5fbd849014c92f384892b8876fe6bc3db417174022bab61e153e2592402bc1f"} Jan 26 16:50:18 crc kubenswrapper[4754]: I0126 16:50:18.197388 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-t4ps6" podStartSLOduration=165.197362243 podStartE2EDuration="2m45.197362243s" podCreationTimestamp="2026-01-26 16:47:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 16:50:18.1933636 +0000 UTC m=+184.717544064" watchObservedRunningTime="2026-01-26 16:50:18.197362243 +0000 UTC m=+184.721542677" Jan 26 16:50:19 crc kubenswrapper[4754]: I0126 16:50:19.495077 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 26 16:50:19 crc kubenswrapper[4754]: I0126 16:50:19.533750 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/99f61c83-b9c7-4559-9eaf-1c08e38244f0-kubelet-dir\") pod \"99f61c83-b9c7-4559-9eaf-1c08e38244f0\" (UID: \"99f61c83-b9c7-4559-9eaf-1c08e38244f0\") " Jan 26 16:50:19 crc kubenswrapper[4754]: I0126 16:50:19.533813 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/99f61c83-b9c7-4559-9eaf-1c08e38244f0-kube-api-access\") pod \"99f61c83-b9c7-4559-9eaf-1c08e38244f0\" (UID: \"99f61c83-b9c7-4559-9eaf-1c08e38244f0\") " Jan 26 16:50:19 crc kubenswrapper[4754]: I0126 16:50:19.534114 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/99f61c83-b9c7-4559-9eaf-1c08e38244f0-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "99f61c83-b9c7-4559-9eaf-1c08e38244f0" (UID: "99f61c83-b9c7-4559-9eaf-1c08e38244f0"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 16:50:19 crc kubenswrapper[4754]: I0126 16:50:19.541526 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/99f61c83-b9c7-4559-9eaf-1c08e38244f0-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "99f61c83-b9c7-4559-9eaf-1c08e38244f0" (UID: "99f61c83-b9c7-4559-9eaf-1c08e38244f0"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:50:19 crc kubenswrapper[4754]: I0126 16:50:19.635217 4754 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/99f61c83-b9c7-4559-9eaf-1c08e38244f0-kubelet-dir\") on node \"crc\" DevicePath \"\"" Jan 26 16:50:19 crc kubenswrapper[4754]: I0126 16:50:19.635269 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/99f61c83-b9c7-4559-9eaf-1c08e38244f0-kube-api-access\") on node \"crc\" DevicePath \"\"" Jan 26 16:50:20 crc kubenswrapper[4754]: I0126 16:50:20.182075 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"99f61c83-b9c7-4559-9eaf-1c08e38244f0","Type":"ContainerDied","Data":"ca937f87082dd69051cd8b845636584e9c95f4902f9570697e4a1d3c964e2f1b"} Jan 26 16:50:20 crc kubenswrapper[4754]: I0126 16:50:20.182309 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ca937f87082dd69051cd8b845636584e9c95f4902f9570697e4a1d3c964e2f1b" Jan 26 16:50:20 crc kubenswrapper[4754]: I0126 16:50:20.182165 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 26 16:50:20 crc kubenswrapper[4754]: I0126 16:50:20.229611 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 16:50:21 crc kubenswrapper[4754]: I0126 16:50:21.224469 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Jan 26 16:50:21 crc kubenswrapper[4754]: E0126 16:50:21.225042 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="71b1f977-1a94-4e7f-bcc1-7a6ec55d043f" containerName="pruner" Jan 26 16:50:21 crc kubenswrapper[4754]: I0126 16:50:21.225059 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="71b1f977-1a94-4e7f-bcc1-7a6ec55d043f" containerName="pruner" Jan 26 16:50:21 crc kubenswrapper[4754]: E0126 16:50:21.225072 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99f61c83-b9c7-4559-9eaf-1c08e38244f0" containerName="pruner" Jan 26 16:50:21 crc kubenswrapper[4754]: I0126 16:50:21.225083 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="99f61c83-b9c7-4559-9eaf-1c08e38244f0" containerName="pruner" Jan 26 16:50:21 crc kubenswrapper[4754]: I0126 16:50:21.225214 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="71b1f977-1a94-4e7f-bcc1-7a6ec55d043f" containerName="pruner" Jan 26 16:50:21 crc kubenswrapper[4754]: I0126 16:50:21.225231 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="99f61c83-b9c7-4559-9eaf-1c08e38244f0" containerName="pruner" Jan 26 16:50:21 crc kubenswrapper[4754]: I0126 16:50:21.225738 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 26 16:50:21 crc kubenswrapper[4754]: I0126 16:50:21.227884 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Jan 26 16:50:21 crc kubenswrapper[4754]: I0126 16:50:21.230206 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Jan 26 16:50:21 crc kubenswrapper[4754]: I0126 16:50:21.230381 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Jan 26 16:50:21 crc kubenswrapper[4754]: I0126 16:50:21.264286 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/871951d9-c0af-4f26-9ba9-2390b34d43df-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"871951d9-c0af-4f26-9ba9-2390b34d43df\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 26 16:50:21 crc kubenswrapper[4754]: I0126 16:50:21.264358 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/871951d9-c0af-4f26-9ba9-2390b34d43df-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"871951d9-c0af-4f26-9ba9-2390b34d43df\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 26 16:50:21 crc kubenswrapper[4754]: I0126 16:50:21.365918 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/871951d9-c0af-4f26-9ba9-2390b34d43df-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"871951d9-c0af-4f26-9ba9-2390b34d43df\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 26 16:50:21 crc kubenswrapper[4754]: I0126 16:50:21.366016 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/871951d9-c0af-4f26-9ba9-2390b34d43df-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"871951d9-c0af-4f26-9ba9-2390b34d43df\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 26 16:50:21 crc kubenswrapper[4754]: I0126 16:50:21.366097 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/871951d9-c0af-4f26-9ba9-2390b34d43df-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"871951d9-c0af-4f26-9ba9-2390b34d43df\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 26 16:50:21 crc kubenswrapper[4754]: I0126 16:50:21.387476 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/871951d9-c0af-4f26-9ba9-2390b34d43df-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"871951d9-c0af-4f26-9ba9-2390b34d43df\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 26 16:50:21 crc kubenswrapper[4754]: I0126 16:50:21.658893 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 26 16:50:22 crc kubenswrapper[4754]: I0126 16:50:22.074018 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Jan 26 16:50:22 crc kubenswrapper[4754]: W0126 16:50:22.085826 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod871951d9_c0af_4f26_9ba9_2390b34d43df.slice/crio-007af352e60ef9c583ca39f74b489c965613d02bc2701fed3ad1bc7fe6fd49c7 WatchSource:0}: Error finding container 007af352e60ef9c583ca39f74b489c965613d02bc2701fed3ad1bc7fe6fd49c7: Status 404 returned error can't find the container with id 007af352e60ef9c583ca39f74b489c965613d02bc2701fed3ad1bc7fe6fd49c7 Jan 26 16:50:22 crc kubenswrapper[4754]: I0126 16:50:22.199010 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"871951d9-c0af-4f26-9ba9-2390b34d43df","Type":"ContainerStarted","Data":"007af352e60ef9c583ca39f74b489c965613d02bc2701fed3ad1bc7fe6fd49c7"} Jan 26 16:50:23 crc kubenswrapper[4754]: I0126 16:50:23.216158 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"871951d9-c0af-4f26-9ba9-2390b34d43df","Type":"ContainerStarted","Data":"7065cd551442a92e4a2371ea0f57996732bb161e243da699bca9a79facdc8e1e"} Jan 26 16:50:23 crc kubenswrapper[4754]: I0126 16:50:23.234385 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-9-crc" podStartSLOduration=2.234367722 podStartE2EDuration="2.234367722s" podCreationTimestamp="2026-01-26 16:50:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 16:50:23.231127346 +0000 UTC m=+189.755307780" watchObservedRunningTime="2026-01-26 16:50:23.234367722 +0000 UTC m=+189.758548156" Jan 26 16:50:24 crc kubenswrapper[4754]: I0126 16:50:24.223950 4754 generic.go:334] "Generic (PLEG): container finished" podID="871951d9-c0af-4f26-9ba9-2390b34d43df" containerID="7065cd551442a92e4a2371ea0f57996732bb161e243da699bca9a79facdc8e1e" exitCode=0 Jan 26 16:50:24 crc kubenswrapper[4754]: I0126 16:50:24.224016 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"871951d9-c0af-4f26-9ba9-2390b34d43df","Type":"ContainerDied","Data":"7065cd551442a92e4a2371ea0f57996732bb161e243da699bca9a79facdc8e1e"} Jan 26 16:50:25 crc kubenswrapper[4754]: I0126 16:50:25.550372 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 26 16:50:25 crc kubenswrapper[4754]: I0126 16:50:25.629277 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/871951d9-c0af-4f26-9ba9-2390b34d43df-kubelet-dir\") pod \"871951d9-c0af-4f26-9ba9-2390b34d43df\" (UID: \"871951d9-c0af-4f26-9ba9-2390b34d43df\") " Jan 26 16:50:25 crc kubenswrapper[4754]: I0126 16:50:25.629417 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/871951d9-c0af-4f26-9ba9-2390b34d43df-kube-api-access\") pod \"871951d9-c0af-4f26-9ba9-2390b34d43df\" (UID: \"871951d9-c0af-4f26-9ba9-2390b34d43df\") " Jan 26 16:50:25 crc kubenswrapper[4754]: I0126 16:50:25.629879 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/871951d9-c0af-4f26-9ba9-2390b34d43df-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "871951d9-c0af-4f26-9ba9-2390b34d43df" (UID: "871951d9-c0af-4f26-9ba9-2390b34d43df"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 16:50:25 crc kubenswrapper[4754]: I0126 16:50:25.636041 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/871951d9-c0af-4f26-9ba9-2390b34d43df-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "871951d9-c0af-4f26-9ba9-2390b34d43df" (UID: "871951d9-c0af-4f26-9ba9-2390b34d43df"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:50:25 crc kubenswrapper[4754]: I0126 16:50:25.731299 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/871951d9-c0af-4f26-9ba9-2390b34d43df-kube-api-access\") on node \"crc\" DevicePath \"\"" Jan 26 16:50:25 crc kubenswrapper[4754]: I0126 16:50:25.731615 4754 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/871951d9-c0af-4f26-9ba9-2390b34d43df-kubelet-dir\") on node \"crc\" DevicePath \"\"" Jan 26 16:50:26 crc kubenswrapper[4754]: I0126 16:50:26.234695 4754 generic.go:334] "Generic (PLEG): container finished" podID="c2962a85-523e-4893-8a19-b7baa1360c01" containerID="fb749d40f32892c54ca3fb9072c0e952b9194886ebca7c67687290380d14da66" exitCode=0 Jan 26 16:50:26 crc kubenswrapper[4754]: I0126 16:50:26.234846 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cgm62" event={"ID":"c2962a85-523e-4893-8a19-b7baa1360c01","Type":"ContainerDied","Data":"fb749d40f32892c54ca3fb9072c0e952b9194886ebca7c67687290380d14da66"} Jan 26 16:50:26 crc kubenswrapper[4754]: I0126 16:50:26.237240 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"871951d9-c0af-4f26-9ba9-2390b34d43df","Type":"ContainerDied","Data":"007af352e60ef9c583ca39f74b489c965613d02bc2701fed3ad1bc7fe6fd49c7"} Jan 26 16:50:26 crc kubenswrapper[4754]: I0126 16:50:26.237276 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="007af352e60ef9c583ca39f74b489c965613d02bc2701fed3ad1bc7fe6fd49c7" Jan 26 16:50:26 crc kubenswrapper[4754]: I0126 16:50:26.237303 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 26 16:50:26 crc kubenswrapper[4754]: I0126 16:50:26.809980 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Jan 26 16:50:26 crc kubenswrapper[4754]: E0126 16:50:26.812088 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="871951d9-c0af-4f26-9ba9-2390b34d43df" containerName="pruner" Jan 26 16:50:26 crc kubenswrapper[4754]: I0126 16:50:26.812112 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="871951d9-c0af-4f26-9ba9-2390b34d43df" containerName="pruner" Jan 26 16:50:26 crc kubenswrapper[4754]: I0126 16:50:26.812250 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="871951d9-c0af-4f26-9ba9-2390b34d43df" containerName="pruner" Jan 26 16:50:26 crc kubenswrapper[4754]: I0126 16:50:26.812721 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Jan 26 16:50:26 crc kubenswrapper[4754]: I0126 16:50:26.815222 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Jan 26 16:50:26 crc kubenswrapper[4754]: I0126 16:50:26.815832 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Jan 26 16:50:26 crc kubenswrapper[4754]: I0126 16:50:26.816068 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Jan 26 16:50:26 crc kubenswrapper[4754]: I0126 16:50:26.948549 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ec0484bb-a715-4082-b6a3-c24200aea47d-kubelet-dir\") pod \"installer-9-crc\" (UID: \"ec0484bb-a715-4082-b6a3-c24200aea47d\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 26 16:50:26 crc kubenswrapper[4754]: I0126 16:50:26.948963 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/ec0484bb-a715-4082-b6a3-c24200aea47d-var-lock\") pod \"installer-9-crc\" (UID: \"ec0484bb-a715-4082-b6a3-c24200aea47d\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 26 16:50:26 crc kubenswrapper[4754]: I0126 16:50:26.949462 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ec0484bb-a715-4082-b6a3-c24200aea47d-kube-api-access\") pod \"installer-9-crc\" (UID: \"ec0484bb-a715-4082-b6a3-c24200aea47d\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 26 16:50:27 crc kubenswrapper[4754]: I0126 16:50:27.050825 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ec0484bb-a715-4082-b6a3-c24200aea47d-kubelet-dir\") pod \"installer-9-crc\" (UID: \"ec0484bb-a715-4082-b6a3-c24200aea47d\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 26 16:50:27 crc kubenswrapper[4754]: I0126 16:50:27.050885 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/ec0484bb-a715-4082-b6a3-c24200aea47d-var-lock\") pod \"installer-9-crc\" (UID: \"ec0484bb-a715-4082-b6a3-c24200aea47d\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 26 16:50:27 crc kubenswrapper[4754]: I0126 16:50:27.050928 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ec0484bb-a715-4082-b6a3-c24200aea47d-kube-api-access\") pod \"installer-9-crc\" (UID: \"ec0484bb-a715-4082-b6a3-c24200aea47d\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 26 16:50:27 crc kubenswrapper[4754]: I0126 16:50:27.050948 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ec0484bb-a715-4082-b6a3-c24200aea47d-kubelet-dir\") pod \"installer-9-crc\" (UID: \"ec0484bb-a715-4082-b6a3-c24200aea47d\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 26 16:50:27 crc kubenswrapper[4754]: I0126 16:50:27.051008 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/ec0484bb-a715-4082-b6a3-c24200aea47d-var-lock\") pod \"installer-9-crc\" (UID: \"ec0484bb-a715-4082-b6a3-c24200aea47d\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 26 16:50:27 crc kubenswrapper[4754]: I0126 16:50:27.070926 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ec0484bb-a715-4082-b6a3-c24200aea47d-kube-api-access\") pod \"installer-9-crc\" (UID: \"ec0484bb-a715-4082-b6a3-c24200aea47d\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 26 16:50:27 crc kubenswrapper[4754]: I0126 16:50:27.128738 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Jan 26 16:50:27 crc kubenswrapper[4754]: I0126 16:50:27.249050 4754 generic.go:334] "Generic (PLEG): container finished" podID="00549dbb-6196-4495-9fbe-4f641333fbe4" containerID="f73529f4bca7fcc98bd8cbf97eb04f827a2485653d45f302b0c037e6088d5c58" exitCode=0 Jan 26 16:50:27 crc kubenswrapper[4754]: I0126 16:50:27.249137 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pdrzp" event={"ID":"00549dbb-6196-4495-9fbe-4f641333fbe4","Type":"ContainerDied","Data":"f73529f4bca7fcc98bd8cbf97eb04f827a2485653d45f302b0c037e6088d5c58"} Jan 26 16:50:27 crc kubenswrapper[4754]: I0126 16:50:27.276211 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cgm62" event={"ID":"c2962a85-523e-4893-8a19-b7baa1360c01","Type":"ContainerStarted","Data":"3533e22f9c0731ad76ebd37e68266f9c0db47db9605ad038aa9b1e1df2d19b8f"} Jan 26 16:50:27 crc kubenswrapper[4754]: I0126 16:50:27.296949 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-cgm62" podStartSLOduration=2.263140924 podStartE2EDuration="51.296928331s" podCreationTimestamp="2026-01-26 16:49:36 +0000 UTC" firstStartedPulling="2026-01-26 16:49:37.622323339 +0000 UTC m=+144.146503763" lastFinishedPulling="2026-01-26 16:50:26.656110736 +0000 UTC m=+193.180291170" observedRunningTime="2026-01-26 16:50:27.293518229 +0000 UTC m=+193.817698673" watchObservedRunningTime="2026-01-26 16:50:27.296928331 +0000 UTC m=+193.821108765" Jan 26 16:50:27 crc kubenswrapper[4754]: I0126 16:50:27.340096 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Jan 26 16:50:27 crc kubenswrapper[4754]: W0126 16:50:27.344809 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-podec0484bb_a715_4082_b6a3_c24200aea47d.slice/crio-f762e0bd95df0403f8f43c124cbcc52c698973710afadcb3ccb8761efb4d9735 WatchSource:0}: Error finding container f762e0bd95df0403f8f43c124cbcc52c698973710afadcb3ccb8761efb4d9735: Status 404 returned error can't find the container with id f762e0bd95df0403f8f43c124cbcc52c698973710afadcb3ccb8761efb4d9735 Jan 26 16:50:28 crc kubenswrapper[4754]: I0126 16:50:28.290534 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pdrzp" event={"ID":"00549dbb-6196-4495-9fbe-4f641333fbe4","Type":"ContainerStarted","Data":"a3e066e210157a65550078911c5eb6bce921318006acd79c094df66473ee03ed"} Jan 26 16:50:28 crc kubenswrapper[4754]: I0126 16:50:28.293057 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"ec0484bb-a715-4082-b6a3-c24200aea47d","Type":"ContainerStarted","Data":"49eaa3ee0155d60c0b3e7836c370d4011c2bc047a24f9c4f96cdb785cc23e842"} Jan 26 16:50:28 crc kubenswrapper[4754]: I0126 16:50:28.293097 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"ec0484bb-a715-4082-b6a3-c24200aea47d","Type":"ContainerStarted","Data":"f762e0bd95df0403f8f43c124cbcc52c698973710afadcb3ccb8761efb4d9735"} Jan 26 16:50:28 crc kubenswrapper[4754]: I0126 16:50:28.303230 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-k9sz7"] Jan 26 16:50:28 crc kubenswrapper[4754]: I0126 16:50:28.321643 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-pdrzp" podStartSLOduration=3.081527284 podStartE2EDuration="52.321601913s" podCreationTimestamp="2026-01-26 16:49:36 +0000 UTC" firstStartedPulling="2026-01-26 16:49:38.659913645 +0000 UTC m=+145.184094089" lastFinishedPulling="2026-01-26 16:50:27.899988284 +0000 UTC m=+194.424168718" observedRunningTime="2026-01-26 16:50:28.317990164 +0000 UTC m=+194.842170608" watchObservedRunningTime="2026-01-26 16:50:28.321601913 +0000 UTC m=+194.845782347" Jan 26 16:50:28 crc kubenswrapper[4754]: I0126 16:50:28.342360 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-9-crc" podStartSLOduration=2.342332507 podStartE2EDuration="2.342332507s" podCreationTimestamp="2026-01-26 16:50:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 16:50:28.339337819 +0000 UTC m=+194.863518263" watchObservedRunningTime="2026-01-26 16:50:28.342332507 +0000 UTC m=+194.866512951" Jan 26 16:50:29 crc kubenswrapper[4754]: I0126 16:50:29.311374 4754 generic.go:334] "Generic (PLEG): container finished" podID="405f8b06-885b-4a7f-9f85-3af411a5504e" containerID="8d1a15d616cd22a1407e01237ef38de0b9959b08b683f259372a140ae7c25711" exitCode=0 Jan 26 16:50:29 crc kubenswrapper[4754]: I0126 16:50:29.311462 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n6gc5" event={"ID":"405f8b06-885b-4a7f-9f85-3af411a5504e","Type":"ContainerDied","Data":"8d1a15d616cd22a1407e01237ef38de0b9959b08b683f259372a140ae7c25711"} Jan 26 16:50:31 crc kubenswrapper[4754]: I0126 16:50:31.336867 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5rvw2" event={"ID":"b1b51a00-e6fe-428a-85ef-9d31badc8457","Type":"ContainerStarted","Data":"48dce29db1605144274c78c071b5a77b3561d6c5d566b2d0b375a641b11eeb33"} Jan 26 16:50:31 crc kubenswrapper[4754]: I0126 16:50:31.340347 4754 generic.go:334] "Generic (PLEG): container finished" podID="ffaa8507-8be1-4bdb-b7b5-28104f9cdbd6" containerID="981bdf1a08dce3f8d632fe59f6c775cfc3fe3d9320be331166b12859d53c5d76" exitCode=0 Jan 26 16:50:31 crc kubenswrapper[4754]: I0126 16:50:31.340435 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-plfzc" event={"ID":"ffaa8507-8be1-4bdb-b7b5-28104f9cdbd6","Type":"ContainerDied","Data":"981bdf1a08dce3f8d632fe59f6c775cfc3fe3d9320be331166b12859d53c5d76"} Jan 26 16:50:31 crc kubenswrapper[4754]: I0126 16:50:31.345127 4754 generic.go:334] "Generic (PLEG): container finished" podID="0ca0de6e-d004-4b46-aee4-66eae716184f" containerID="3773bafe0080ca3c474a93c1b34b4f446d36eb3f9c3e6cda5f3da21f5babf026" exitCode=0 Jan 26 16:50:31 crc kubenswrapper[4754]: I0126 16:50:31.345252 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qzw56" event={"ID":"0ca0de6e-d004-4b46-aee4-66eae716184f","Type":"ContainerDied","Data":"3773bafe0080ca3c474a93c1b34b4f446d36eb3f9c3e6cda5f3da21f5babf026"} Jan 26 16:50:31 crc kubenswrapper[4754]: I0126 16:50:31.348843 4754 generic.go:334] "Generic (PLEG): container finished" podID="e36dc2d0-47de-46a6-aeba-52475754867e" containerID="d20015cf16e678be973812f50ecb7e568826bfca9eeb39c0d06deeda2df3a367" exitCode=0 Jan 26 16:50:31 crc kubenswrapper[4754]: I0126 16:50:31.348936 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hhhn5" event={"ID":"e36dc2d0-47de-46a6-aeba-52475754867e","Type":"ContainerDied","Data":"d20015cf16e678be973812f50ecb7e568826bfca9eeb39c0d06deeda2df3a367"} Jan 26 16:50:31 crc kubenswrapper[4754]: I0126 16:50:31.351853 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n6gc5" event={"ID":"405f8b06-885b-4a7f-9f85-3af411a5504e","Type":"ContainerStarted","Data":"611ca687d126090816b4a9fa2bb55e52aea89046cf32da59b9488ddc03890c1f"} Jan 26 16:50:31 crc kubenswrapper[4754]: I0126 16:50:31.370975 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pdj77" event={"ID":"143cd811-0dff-4141-bbb4-f940c83a0bb0","Type":"ContainerStarted","Data":"68a86560e6569c8a47cba91060d966af8846227eef9b2e2217fbf2fee2618da0"} Jan 26 16:50:31 crc kubenswrapper[4754]: I0126 16:50:31.448487 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-n6gc5" podStartSLOduration=3.744624516 podStartE2EDuration="55.448471361s" podCreationTimestamp="2026-01-26 16:49:36 +0000 UTC" firstStartedPulling="2026-01-26 16:49:38.60796113 +0000 UTC m=+145.132141564" lastFinishedPulling="2026-01-26 16:50:30.311807975 +0000 UTC m=+196.835988409" observedRunningTime="2026-01-26 16:50:31.446310148 +0000 UTC m=+197.970490582" watchObservedRunningTime="2026-01-26 16:50:31.448471361 +0000 UTC m=+197.972651795" Jan 26 16:50:33 crc kubenswrapper[4754]: I0126 16:50:33.426651 4754 generic.go:334] "Generic (PLEG): container finished" podID="143cd811-0dff-4141-bbb4-f940c83a0bb0" containerID="68a86560e6569c8a47cba91060d966af8846227eef9b2e2217fbf2fee2618da0" exitCode=0 Jan 26 16:50:33 crc kubenswrapper[4754]: I0126 16:50:33.426759 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pdj77" event={"ID":"143cd811-0dff-4141-bbb4-f940c83a0bb0","Type":"ContainerDied","Data":"68a86560e6569c8a47cba91060d966af8846227eef9b2e2217fbf2fee2618da0"} Jan 26 16:50:33 crc kubenswrapper[4754]: I0126 16:50:33.430414 4754 generic.go:334] "Generic (PLEG): container finished" podID="b1b51a00-e6fe-428a-85ef-9d31badc8457" containerID="48dce29db1605144274c78c071b5a77b3561d6c5d566b2d0b375a641b11eeb33" exitCode=0 Jan 26 16:50:33 crc kubenswrapper[4754]: I0126 16:50:33.430463 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5rvw2" event={"ID":"b1b51a00-e6fe-428a-85ef-9d31badc8457","Type":"ContainerDied","Data":"48dce29db1605144274c78c071b5a77b3561d6c5d566b2d0b375a641b11eeb33"} Jan 26 16:50:36 crc kubenswrapper[4754]: I0126 16:50:36.392317 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-cgm62" Jan 26 16:50:36 crc kubenswrapper[4754]: I0126 16:50:36.392910 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-cgm62" Jan 26 16:50:36 crc kubenswrapper[4754]: I0126 16:50:36.452239 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-plfzc" event={"ID":"ffaa8507-8be1-4bdb-b7b5-28104f9cdbd6","Type":"ContainerStarted","Data":"6a36d67ad9d8f7a38badd7a4f0db61b7b45a38e46e7229756fa9846805f51abe"} Jan 26 16:50:36 crc kubenswrapper[4754]: I0126 16:50:36.912163 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-pdrzp" Jan 26 16:50:36 crc kubenswrapper[4754]: I0126 16:50:36.912242 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-pdrzp" Jan 26 16:50:37 crc kubenswrapper[4754]: I0126 16:50:37.118749 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-n6gc5" Jan 26 16:50:37 crc kubenswrapper[4754]: I0126 16:50:37.118791 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-n6gc5" Jan 26 16:50:37 crc kubenswrapper[4754]: I0126 16:50:37.129735 4754 patch_prober.go:28] interesting pod/machine-config-daemon-x65wv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 26 16:50:37 crc kubenswrapper[4754]: I0126 16:50:37.129785 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 26 16:50:37 crc kubenswrapper[4754]: I0126 16:50:37.129818 4754 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" Jan 26 16:50:37 crc kubenswrapper[4754]: I0126 16:50:37.130188 4754 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"0fffc03b536bfee3c7bc1bddd500e834e34b391896477a1501e3e06201a19374"} pod="openshift-machine-config-operator/machine-config-daemon-x65wv" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 26 16:50:37 crc kubenswrapper[4754]: I0126 16:50:37.130283 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" containerName="machine-config-daemon" containerID="cri-o://0fffc03b536bfee3c7bc1bddd500e834e34b391896477a1501e3e06201a19374" gracePeriod=600 Jan 26 16:50:37 crc kubenswrapper[4754]: I0126 16:50:37.259184 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-cgm62" Jan 26 16:50:37 crc kubenswrapper[4754]: I0126 16:50:37.259793 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-n6gc5" Jan 26 16:50:37 crc kubenswrapper[4754]: I0126 16:50:37.264279 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-pdrzp" Jan 26 16:50:37 crc kubenswrapper[4754]: I0126 16:50:37.296933 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-cgm62" Jan 26 16:50:37 crc kubenswrapper[4754]: I0126 16:50:37.459837 4754 generic.go:334] "Generic (PLEG): container finished" podID="8c3718a4-f354-4284-92e0-fdfb45a692bd" containerID="0fffc03b536bfee3c7bc1bddd500e834e34b391896477a1501e3e06201a19374" exitCode=0 Jan 26 16:50:37 crc kubenswrapper[4754]: I0126 16:50:37.459949 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" event={"ID":"8c3718a4-f354-4284-92e0-fdfb45a692bd","Type":"ContainerDied","Data":"0fffc03b536bfee3c7bc1bddd500e834e34b391896477a1501e3e06201a19374"} Jan 26 16:50:37 crc kubenswrapper[4754]: I0126 16:50:37.478150 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-plfzc" podStartSLOduration=4.170336246 podStartE2EDuration="59.478134203s" podCreationTimestamp="2026-01-26 16:49:38 +0000 UTC" firstStartedPulling="2026-01-26 16:49:39.695610802 +0000 UTC m=+146.219791236" lastFinishedPulling="2026-01-26 16:50:35.003408759 +0000 UTC m=+201.527589193" observedRunningTime="2026-01-26 16:50:37.477192495 +0000 UTC m=+204.001372929" watchObservedRunningTime="2026-01-26 16:50:37.478134203 +0000 UTC m=+204.002314637" Jan 26 16:50:37 crc kubenswrapper[4754]: I0126 16:50:37.506069 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-pdrzp" Jan 26 16:50:37 crc kubenswrapper[4754]: I0126 16:50:37.510606 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-n6gc5" Jan 26 16:50:38 crc kubenswrapper[4754]: I0126 16:50:38.602723 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-pdrzp"] Jan 26 16:50:38 crc kubenswrapper[4754]: I0126 16:50:38.789984 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-plfzc" Jan 26 16:50:38 crc kubenswrapper[4754]: I0126 16:50:38.790036 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-plfzc" Jan 26 16:50:38 crc kubenswrapper[4754]: I0126 16:50:38.827890 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-plfzc" Jan 26 16:50:39 crc kubenswrapper[4754]: I0126 16:50:39.469571 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-pdrzp" podUID="00549dbb-6196-4495-9fbe-4f641333fbe4" containerName="registry-server" containerID="cri-o://a3e066e210157a65550078911c5eb6bce921318006acd79c094df66473ee03ed" gracePeriod=2 Jan 26 16:50:39 crc kubenswrapper[4754]: I0126 16:50:39.603946 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-n6gc5"] Jan 26 16:50:39 crc kubenswrapper[4754]: I0126 16:50:39.604561 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-n6gc5" podUID="405f8b06-885b-4a7f-9f85-3af411a5504e" containerName="registry-server" containerID="cri-o://611ca687d126090816b4a9fa2bb55e52aea89046cf32da59b9488ddc03890c1f" gracePeriod=2 Jan 26 16:50:40 crc kubenswrapper[4754]: I0126 16:50:40.481406 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" event={"ID":"8c3718a4-f354-4284-92e0-fdfb45a692bd","Type":"ContainerStarted","Data":"e4aaaab9b6e326f553465f494a4b68086742f0af0a269e13ef1c99f9395124a5"} Jan 26 16:50:40 crc kubenswrapper[4754]: I0126 16:50:40.485093 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hhhn5" event={"ID":"e36dc2d0-47de-46a6-aeba-52475754867e","Type":"ContainerStarted","Data":"2e25b7bf0c7d3598416895619dfd7bf1e691d76eb5e80aee438983164baf5bf0"} Jan 26 16:50:40 crc kubenswrapper[4754]: I0126 16:50:40.487681 4754 generic.go:334] "Generic (PLEG): container finished" podID="405f8b06-885b-4a7f-9f85-3af411a5504e" containerID="611ca687d126090816b4a9fa2bb55e52aea89046cf32da59b9488ddc03890c1f" exitCode=0 Jan 26 16:50:40 crc kubenswrapper[4754]: I0126 16:50:40.487774 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n6gc5" event={"ID":"405f8b06-885b-4a7f-9f85-3af411a5504e","Type":"ContainerDied","Data":"611ca687d126090816b4a9fa2bb55e52aea89046cf32da59b9488ddc03890c1f"} Jan 26 16:50:40 crc kubenswrapper[4754]: I0126 16:50:40.490392 4754 generic.go:334] "Generic (PLEG): container finished" podID="00549dbb-6196-4495-9fbe-4f641333fbe4" containerID="a3e066e210157a65550078911c5eb6bce921318006acd79c094df66473ee03ed" exitCode=0 Jan 26 16:50:40 crc kubenswrapper[4754]: I0126 16:50:40.490441 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pdrzp" event={"ID":"00549dbb-6196-4495-9fbe-4f641333fbe4","Type":"ContainerDied","Data":"a3e066e210157a65550078911c5eb6bce921318006acd79c094df66473ee03ed"} Jan 26 16:50:40 crc kubenswrapper[4754]: I0126 16:50:40.514001 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-hhhn5" podStartSLOduration=3.069387733 podStartE2EDuration="1m2.513982072s" podCreationTimestamp="2026-01-26 16:49:38 +0000 UTC" firstStartedPulling="2026-01-26 16:49:39.737876789 +0000 UTC m=+146.262057223" lastFinishedPulling="2026-01-26 16:50:39.182471128 +0000 UTC m=+205.706651562" observedRunningTime="2026-01-26 16:50:40.512097648 +0000 UTC m=+207.036278082" watchObservedRunningTime="2026-01-26 16:50:40.513982072 +0000 UTC m=+207.038162506" Jan 26 16:50:40 crc kubenswrapper[4754]: I0126 16:50:40.689109 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pdrzp" Jan 26 16:50:40 crc kubenswrapper[4754]: I0126 16:50:40.751941 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/00549dbb-6196-4495-9fbe-4f641333fbe4-catalog-content\") pod \"00549dbb-6196-4495-9fbe-4f641333fbe4\" (UID: \"00549dbb-6196-4495-9fbe-4f641333fbe4\") " Jan 26 16:50:40 crc kubenswrapper[4754]: I0126 16:50:40.752139 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/00549dbb-6196-4495-9fbe-4f641333fbe4-utilities\") pod \"00549dbb-6196-4495-9fbe-4f641333fbe4\" (UID: \"00549dbb-6196-4495-9fbe-4f641333fbe4\") " Jan 26 16:50:40 crc kubenswrapper[4754]: I0126 16:50:40.752188 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8wwwp\" (UniqueName: \"kubernetes.io/projected/00549dbb-6196-4495-9fbe-4f641333fbe4-kube-api-access-8wwwp\") pod \"00549dbb-6196-4495-9fbe-4f641333fbe4\" (UID: \"00549dbb-6196-4495-9fbe-4f641333fbe4\") " Jan 26 16:50:40 crc kubenswrapper[4754]: I0126 16:50:40.753264 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/00549dbb-6196-4495-9fbe-4f641333fbe4-utilities" (OuterVolumeSpecName: "utilities") pod "00549dbb-6196-4495-9fbe-4f641333fbe4" (UID: "00549dbb-6196-4495-9fbe-4f641333fbe4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 16:50:40 crc kubenswrapper[4754]: I0126 16:50:40.762766 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/00549dbb-6196-4495-9fbe-4f641333fbe4-kube-api-access-8wwwp" (OuterVolumeSpecName: "kube-api-access-8wwwp") pod "00549dbb-6196-4495-9fbe-4f641333fbe4" (UID: "00549dbb-6196-4495-9fbe-4f641333fbe4"). InnerVolumeSpecName "kube-api-access-8wwwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:50:40 crc kubenswrapper[4754]: I0126 16:50:40.784473 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-n6gc5" Jan 26 16:50:40 crc kubenswrapper[4754]: I0126 16:50:40.853886 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/405f8b06-885b-4a7f-9f85-3af411a5504e-catalog-content\") pod \"405f8b06-885b-4a7f-9f85-3af411a5504e\" (UID: \"405f8b06-885b-4a7f-9f85-3af411a5504e\") " Jan 26 16:50:40 crc kubenswrapper[4754]: I0126 16:50:40.854031 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4n544\" (UniqueName: \"kubernetes.io/projected/405f8b06-885b-4a7f-9f85-3af411a5504e-kube-api-access-4n544\") pod \"405f8b06-885b-4a7f-9f85-3af411a5504e\" (UID: \"405f8b06-885b-4a7f-9f85-3af411a5504e\") " Jan 26 16:50:40 crc kubenswrapper[4754]: I0126 16:50:40.854167 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/405f8b06-885b-4a7f-9f85-3af411a5504e-utilities\") pod \"405f8b06-885b-4a7f-9f85-3af411a5504e\" (UID: \"405f8b06-885b-4a7f-9f85-3af411a5504e\") " Jan 26 16:50:40 crc kubenswrapper[4754]: I0126 16:50:40.854605 4754 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/00549dbb-6196-4495-9fbe-4f641333fbe4-utilities\") on node \"crc\" DevicePath \"\"" Jan 26 16:50:40 crc kubenswrapper[4754]: I0126 16:50:40.854628 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8wwwp\" (UniqueName: \"kubernetes.io/projected/00549dbb-6196-4495-9fbe-4f641333fbe4-kube-api-access-8wwwp\") on node \"crc\" DevicePath \"\"" Jan 26 16:50:40 crc kubenswrapper[4754]: I0126 16:50:40.857731 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/405f8b06-885b-4a7f-9f85-3af411a5504e-utilities" (OuterVolumeSpecName: "utilities") pod "405f8b06-885b-4a7f-9f85-3af411a5504e" (UID: "405f8b06-885b-4a7f-9f85-3af411a5504e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 16:50:40 crc kubenswrapper[4754]: I0126 16:50:40.859981 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/405f8b06-885b-4a7f-9f85-3af411a5504e-kube-api-access-4n544" (OuterVolumeSpecName: "kube-api-access-4n544") pod "405f8b06-885b-4a7f-9f85-3af411a5504e" (UID: "405f8b06-885b-4a7f-9f85-3af411a5504e"). InnerVolumeSpecName "kube-api-access-4n544". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:50:40 crc kubenswrapper[4754]: I0126 16:50:40.904841 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/405f8b06-885b-4a7f-9f85-3af411a5504e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "405f8b06-885b-4a7f-9f85-3af411a5504e" (UID: "405f8b06-885b-4a7f-9f85-3af411a5504e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 16:50:40 crc kubenswrapper[4754]: I0126 16:50:40.955396 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4n544\" (UniqueName: \"kubernetes.io/projected/405f8b06-885b-4a7f-9f85-3af411a5504e-kube-api-access-4n544\") on node \"crc\" DevicePath \"\"" Jan 26 16:50:40 crc kubenswrapper[4754]: I0126 16:50:40.955436 4754 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/405f8b06-885b-4a7f-9f85-3af411a5504e-utilities\") on node \"crc\" DevicePath \"\"" Jan 26 16:50:40 crc kubenswrapper[4754]: I0126 16:50:40.955447 4754 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/405f8b06-885b-4a7f-9f85-3af411a5504e-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 26 16:50:41 crc kubenswrapper[4754]: I0126 16:50:41.333224 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/00549dbb-6196-4495-9fbe-4f641333fbe4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "00549dbb-6196-4495-9fbe-4f641333fbe4" (UID: "00549dbb-6196-4495-9fbe-4f641333fbe4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 16:50:41 crc kubenswrapper[4754]: I0126 16:50:41.360584 4754 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/00549dbb-6196-4495-9fbe-4f641333fbe4-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 26 16:50:41 crc kubenswrapper[4754]: I0126 16:50:41.502552 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-n6gc5" Jan 26 16:50:41 crc kubenswrapper[4754]: I0126 16:50:41.502843 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n6gc5" event={"ID":"405f8b06-885b-4a7f-9f85-3af411a5504e","Type":"ContainerDied","Data":"bdde98850848c7ec866e6c5c166d541f33a37a6358e679b86327114f43383271"} Jan 26 16:50:41 crc kubenswrapper[4754]: I0126 16:50:41.502895 4754 scope.go:117] "RemoveContainer" containerID="611ca687d126090816b4a9fa2bb55e52aea89046cf32da59b9488ddc03890c1f" Jan 26 16:50:41 crc kubenswrapper[4754]: I0126 16:50:41.507709 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pdrzp" event={"ID":"00549dbb-6196-4495-9fbe-4f641333fbe4","Type":"ContainerDied","Data":"58ac598971e156a2bceaaf859c1344080a3816dafb3d572b49ec804f2ae08890"} Jan 26 16:50:41 crc kubenswrapper[4754]: I0126 16:50:41.507728 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pdrzp" Jan 26 16:50:41 crc kubenswrapper[4754]: I0126 16:50:41.512345 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5rvw2" event={"ID":"b1b51a00-e6fe-428a-85ef-9d31badc8457","Type":"ContainerStarted","Data":"bcf0497d70be7f24baa4f82c8fe6e79184c60b7185ce37f72e6a06cb1f133a33"} Jan 26 16:50:41 crc kubenswrapper[4754]: I0126 16:50:41.531886 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-5rvw2" podStartSLOduration=2.937323312 podStartE2EDuration="1m2.531866661s" podCreationTimestamp="2026-01-26 16:49:39 +0000 UTC" firstStartedPulling="2026-01-26 16:49:40.750583071 +0000 UTC m=+147.274763505" lastFinishedPulling="2026-01-26 16:50:40.34512642 +0000 UTC m=+206.869306854" observedRunningTime="2026-01-26 16:50:41.528584376 +0000 UTC m=+208.052764810" watchObservedRunningTime="2026-01-26 16:50:41.531866661 +0000 UTC m=+208.056047105" Jan 26 16:50:41 crc kubenswrapper[4754]: I0126 16:50:41.545883 4754 scope.go:117] "RemoveContainer" containerID="8d1a15d616cd22a1407e01237ef38de0b9959b08b683f259372a140ae7c25711" Jan 26 16:50:41 crc kubenswrapper[4754]: I0126 16:50:41.562348 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-n6gc5"] Jan 26 16:50:41 crc kubenswrapper[4754]: I0126 16:50:41.572221 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-n6gc5"] Jan 26 16:50:41 crc kubenswrapper[4754]: I0126 16:50:41.576605 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-pdrzp"] Jan 26 16:50:41 crc kubenswrapper[4754]: I0126 16:50:41.580526 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-pdrzp"] Jan 26 16:50:41 crc kubenswrapper[4754]: I0126 16:50:41.587646 4754 scope.go:117] "RemoveContainer" containerID="20a4db1eb833f1df1ecdad9bd783bf791d2a67664cb40a433a5f49bee71cbc7b" Jan 26 16:50:41 crc kubenswrapper[4754]: I0126 16:50:41.616295 4754 scope.go:117] "RemoveContainer" containerID="a3e066e210157a65550078911c5eb6bce921318006acd79c094df66473ee03ed" Jan 26 16:50:41 crc kubenswrapper[4754]: I0126 16:50:41.635002 4754 scope.go:117] "RemoveContainer" containerID="f73529f4bca7fcc98bd8cbf97eb04f827a2485653d45f302b0c037e6088d5c58" Jan 26 16:50:41 crc kubenswrapper[4754]: I0126 16:50:41.649161 4754 scope.go:117] "RemoveContainer" containerID="12b0ca264597c9685d69dc4164af214cba7256ed8b758fd7235dab47fb7a5cb8" Jan 26 16:50:41 crc kubenswrapper[4754]: I0126 16:50:41.780572 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="00549dbb-6196-4495-9fbe-4f641333fbe4" path="/var/lib/kubelet/pods/00549dbb-6196-4495-9fbe-4f641333fbe4/volumes" Jan 26 16:50:41 crc kubenswrapper[4754]: I0126 16:50:41.781220 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="405f8b06-885b-4a7f-9f85-3af411a5504e" path="/var/lib/kubelet/pods/405f8b06-885b-4a7f-9f85-3af411a5504e/volumes" Jan 26 16:50:42 crc kubenswrapper[4754]: I0126 16:50:42.521788 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pdj77" event={"ID":"143cd811-0dff-4141-bbb4-f940c83a0bb0","Type":"ContainerStarted","Data":"94f4e35dd7f774e1448690af9ee5b301d4530c4808c7492fe0bad4897dc9af36"} Jan 26 16:50:42 crc kubenswrapper[4754]: I0126 16:50:42.523372 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qzw56" event={"ID":"0ca0de6e-d004-4b46-aee4-66eae716184f","Type":"ContainerStarted","Data":"c87ab9b579140b0a54235a7b7812977815f23e9829ef9fadcf2f4775443d3c8f"} Jan 26 16:50:42 crc kubenswrapper[4754]: I0126 16:50:42.548133 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-pdj77" podStartSLOduration=2.878737402 podStartE2EDuration="1m3.548112491s" podCreationTimestamp="2026-01-26 16:49:39 +0000 UTC" firstStartedPulling="2026-01-26 16:49:40.75752125 +0000 UTC m=+147.281701684" lastFinishedPulling="2026-01-26 16:50:41.426896339 +0000 UTC m=+207.951076773" observedRunningTime="2026-01-26 16:50:42.543437126 +0000 UTC m=+209.067617560" watchObservedRunningTime="2026-01-26 16:50:42.548112491 +0000 UTC m=+209.072292935" Jan 26 16:50:42 crc kubenswrapper[4754]: I0126 16:50:42.563344 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-qzw56" podStartSLOduration=4.811096611 podStartE2EDuration="1m7.563324502s" podCreationTimestamp="2026-01-26 16:49:35 +0000 UTC" firstStartedPulling="2026-01-26 16:49:38.673814614 +0000 UTC m=+145.197995048" lastFinishedPulling="2026-01-26 16:50:41.426042505 +0000 UTC m=+207.950222939" observedRunningTime="2026-01-26 16:50:42.560562902 +0000 UTC m=+209.084743336" watchObservedRunningTime="2026-01-26 16:50:42.563324502 +0000 UTC m=+209.087504946" Jan 26 16:50:46 crc kubenswrapper[4754]: I0126 16:50:46.884359 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-qzw56" Jan 26 16:50:46 crc kubenswrapper[4754]: I0126 16:50:46.884694 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-qzw56" Jan 26 16:50:46 crc kubenswrapper[4754]: I0126 16:50:46.925894 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-qzw56" Jan 26 16:50:47 crc kubenswrapper[4754]: I0126 16:50:47.596007 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-qzw56" Jan 26 16:50:48 crc kubenswrapper[4754]: I0126 16:50:48.402111 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-hhhn5" Jan 26 16:50:48 crc kubenswrapper[4754]: I0126 16:50:48.402391 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-hhhn5" Jan 26 16:50:48 crc kubenswrapper[4754]: I0126 16:50:48.446423 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-hhhn5" Jan 26 16:50:48 crc kubenswrapper[4754]: I0126 16:50:48.593689 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-hhhn5" Jan 26 16:50:48 crc kubenswrapper[4754]: I0126 16:50:48.849373 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-plfzc" Jan 26 16:50:49 crc kubenswrapper[4754]: I0126 16:50:49.406407 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-plfzc"] Jan 26 16:50:49 crc kubenswrapper[4754]: I0126 16:50:49.415869 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-5rvw2" Jan 26 16:50:49 crc kubenswrapper[4754]: I0126 16:50:49.415999 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-5rvw2" Jan 26 16:50:49 crc kubenswrapper[4754]: I0126 16:50:49.454689 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-5rvw2" Jan 26 16:50:49 crc kubenswrapper[4754]: I0126 16:50:49.560715 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-plfzc" podUID="ffaa8507-8be1-4bdb-b7b5-28104f9cdbd6" containerName="registry-server" containerID="cri-o://6a36d67ad9d8f7a38badd7a4f0db61b7b45a38e46e7229756fa9846805f51abe" gracePeriod=2 Jan 26 16:50:49 crc kubenswrapper[4754]: I0126 16:50:49.692355 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-5rvw2" Jan 26 16:50:49 crc kubenswrapper[4754]: I0126 16:50:49.796460 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-pdj77" Jan 26 16:50:49 crc kubenswrapper[4754]: I0126 16:50:49.796507 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-pdj77" Jan 26 16:50:49 crc kubenswrapper[4754]: I0126 16:50:49.856389 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-pdj77" Jan 26 16:50:50 crc kubenswrapper[4754]: I0126 16:50:50.567724 4754 generic.go:334] "Generic (PLEG): container finished" podID="ffaa8507-8be1-4bdb-b7b5-28104f9cdbd6" containerID="6a36d67ad9d8f7a38badd7a4f0db61b7b45a38e46e7229756fa9846805f51abe" exitCode=0 Jan 26 16:50:50 crc kubenswrapper[4754]: I0126 16:50:50.567803 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-plfzc" event={"ID":"ffaa8507-8be1-4bdb-b7b5-28104f9cdbd6","Type":"ContainerDied","Data":"6a36d67ad9d8f7a38badd7a4f0db61b7b45a38e46e7229756fa9846805f51abe"} Jan 26 16:50:50 crc kubenswrapper[4754]: I0126 16:50:50.610748 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-pdj77" Jan 26 16:50:51 crc kubenswrapper[4754]: I0126 16:50:51.054997 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-plfzc" Jan 26 16:50:51 crc kubenswrapper[4754]: I0126 16:50:51.181589 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ffaa8507-8be1-4bdb-b7b5-28104f9cdbd6-catalog-content\") pod \"ffaa8507-8be1-4bdb-b7b5-28104f9cdbd6\" (UID: \"ffaa8507-8be1-4bdb-b7b5-28104f9cdbd6\") " Jan 26 16:50:51 crc kubenswrapper[4754]: I0126 16:50:51.181701 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hts6c\" (UniqueName: \"kubernetes.io/projected/ffaa8507-8be1-4bdb-b7b5-28104f9cdbd6-kube-api-access-hts6c\") pod \"ffaa8507-8be1-4bdb-b7b5-28104f9cdbd6\" (UID: \"ffaa8507-8be1-4bdb-b7b5-28104f9cdbd6\") " Jan 26 16:50:51 crc kubenswrapper[4754]: I0126 16:50:51.181884 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ffaa8507-8be1-4bdb-b7b5-28104f9cdbd6-utilities\") pod \"ffaa8507-8be1-4bdb-b7b5-28104f9cdbd6\" (UID: \"ffaa8507-8be1-4bdb-b7b5-28104f9cdbd6\") " Jan 26 16:50:51 crc kubenswrapper[4754]: I0126 16:50:51.182589 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ffaa8507-8be1-4bdb-b7b5-28104f9cdbd6-utilities" (OuterVolumeSpecName: "utilities") pod "ffaa8507-8be1-4bdb-b7b5-28104f9cdbd6" (UID: "ffaa8507-8be1-4bdb-b7b5-28104f9cdbd6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 16:50:51 crc kubenswrapper[4754]: I0126 16:50:51.189930 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ffaa8507-8be1-4bdb-b7b5-28104f9cdbd6-kube-api-access-hts6c" (OuterVolumeSpecName: "kube-api-access-hts6c") pod "ffaa8507-8be1-4bdb-b7b5-28104f9cdbd6" (UID: "ffaa8507-8be1-4bdb-b7b5-28104f9cdbd6"). InnerVolumeSpecName "kube-api-access-hts6c". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:50:51 crc kubenswrapper[4754]: I0126 16:50:51.224456 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ffaa8507-8be1-4bdb-b7b5-28104f9cdbd6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ffaa8507-8be1-4bdb-b7b5-28104f9cdbd6" (UID: "ffaa8507-8be1-4bdb-b7b5-28104f9cdbd6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 16:50:51 crc kubenswrapper[4754]: I0126 16:50:51.283590 4754 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ffaa8507-8be1-4bdb-b7b5-28104f9cdbd6-utilities\") on node \"crc\" DevicePath \"\"" Jan 26 16:50:51 crc kubenswrapper[4754]: I0126 16:50:51.283628 4754 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ffaa8507-8be1-4bdb-b7b5-28104f9cdbd6-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 26 16:50:51 crc kubenswrapper[4754]: I0126 16:50:51.283643 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hts6c\" (UniqueName: \"kubernetes.io/projected/ffaa8507-8be1-4bdb-b7b5-28104f9cdbd6-kube-api-access-hts6c\") on node \"crc\" DevicePath \"\"" Jan 26 16:50:51 crc kubenswrapper[4754]: I0126 16:50:51.579252 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-plfzc" event={"ID":"ffaa8507-8be1-4bdb-b7b5-28104f9cdbd6","Type":"ContainerDied","Data":"72a9bd98676797e9a665dda6d8c130093e4dd4b22599b054bf0adc9288a02ae4"} Jan 26 16:50:51 crc kubenswrapper[4754]: I0126 16:50:51.579311 4754 scope.go:117] "RemoveContainer" containerID="6a36d67ad9d8f7a38badd7a4f0db61b7b45a38e46e7229756fa9846805f51abe" Jan 26 16:50:51 crc kubenswrapper[4754]: I0126 16:50:51.579508 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-plfzc" Jan 26 16:50:51 crc kubenswrapper[4754]: I0126 16:50:51.599889 4754 scope.go:117] "RemoveContainer" containerID="981bdf1a08dce3f8d632fe59f6c775cfc3fe3d9320be331166b12859d53c5d76" Jan 26 16:50:51 crc kubenswrapper[4754]: I0126 16:50:51.608703 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-plfzc"] Jan 26 16:50:51 crc kubenswrapper[4754]: I0126 16:50:51.613172 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-plfzc"] Jan 26 16:50:51 crc kubenswrapper[4754]: I0126 16:50:51.635956 4754 scope.go:117] "RemoveContainer" containerID="f5ad89d1e92c1dc4aa297ec8985614cde8f91e56458d1c6b3e6db70a8efdf4ac" Jan 26 16:50:51 crc kubenswrapper[4754]: I0126 16:50:51.774761 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ffaa8507-8be1-4bdb-b7b5-28104f9cdbd6" path="/var/lib/kubelet/pods/ffaa8507-8be1-4bdb-b7b5-28104f9cdbd6/volumes" Jan 26 16:50:51 crc kubenswrapper[4754]: I0126 16:50:51.806724 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-pdj77"] Jan 26 16:50:52 crc kubenswrapper[4754]: I0126 16:50:52.585349 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-pdj77" podUID="143cd811-0dff-4141-bbb4-f940c83a0bb0" containerName="registry-server" containerID="cri-o://94f4e35dd7f774e1448690af9ee5b301d4530c4808c7492fe0bad4897dc9af36" gracePeriod=2 Jan 26 16:50:53 crc kubenswrapper[4754]: I0126 16:50:53.353316 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-k9sz7" podUID="2a28f341-5f93-4d6c-bc4b-973d33dfeee1" containerName="oauth-openshift" containerID="cri-o://a7b1a2116140f01645d1d09ad047f6edc2f84fbb9c15315f990b891c1b623394" gracePeriod=15 Jan 26 16:50:56 crc kubenswrapper[4754]: I0126 16:50:56.613987 4754 generic.go:334] "Generic (PLEG): container finished" podID="2a28f341-5f93-4d6c-bc4b-973d33dfeee1" containerID="a7b1a2116140f01645d1d09ad047f6edc2f84fbb9c15315f990b891c1b623394" exitCode=0 Jan 26 16:50:56 crc kubenswrapper[4754]: I0126 16:50:56.614059 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-k9sz7" event={"ID":"2a28f341-5f93-4d6c-bc4b-973d33dfeee1","Type":"ContainerDied","Data":"a7b1a2116140f01645d1d09ad047f6edc2f84fbb9c15315f990b891c1b623394"} Jan 26 16:50:56 crc kubenswrapper[4754]: I0126 16:50:56.616251 4754 generic.go:334] "Generic (PLEG): container finished" podID="143cd811-0dff-4141-bbb4-f940c83a0bb0" containerID="94f4e35dd7f774e1448690af9ee5b301d4530c4808c7492fe0bad4897dc9af36" exitCode=0 Jan 26 16:50:56 crc kubenswrapper[4754]: I0126 16:50:56.616300 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pdj77" event={"ID":"143cd811-0dff-4141-bbb4-f940c83a0bb0","Type":"ContainerDied","Data":"94f4e35dd7f774e1448690af9ee5b301d4530c4808c7492fe0bad4897dc9af36"} Jan 26 16:50:56 crc kubenswrapper[4754]: I0126 16:50:56.828916 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pdj77" Jan 26 16:50:56 crc kubenswrapper[4754]: I0126 16:50:56.856041 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ppdns\" (UniqueName: \"kubernetes.io/projected/143cd811-0dff-4141-bbb4-f940c83a0bb0-kube-api-access-ppdns\") pod \"143cd811-0dff-4141-bbb4-f940c83a0bb0\" (UID: \"143cd811-0dff-4141-bbb4-f940c83a0bb0\") " Jan 26 16:50:56 crc kubenswrapper[4754]: I0126 16:50:56.856182 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/143cd811-0dff-4141-bbb4-f940c83a0bb0-utilities\") pod \"143cd811-0dff-4141-bbb4-f940c83a0bb0\" (UID: \"143cd811-0dff-4141-bbb4-f940c83a0bb0\") " Jan 26 16:50:56 crc kubenswrapper[4754]: I0126 16:50:56.856295 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/143cd811-0dff-4141-bbb4-f940c83a0bb0-catalog-content\") pod \"143cd811-0dff-4141-bbb4-f940c83a0bb0\" (UID: \"143cd811-0dff-4141-bbb4-f940c83a0bb0\") " Jan 26 16:50:56 crc kubenswrapper[4754]: I0126 16:50:56.859114 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/143cd811-0dff-4141-bbb4-f940c83a0bb0-utilities" (OuterVolumeSpecName: "utilities") pod "143cd811-0dff-4141-bbb4-f940c83a0bb0" (UID: "143cd811-0dff-4141-bbb4-f940c83a0bb0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 16:50:56 crc kubenswrapper[4754]: I0126 16:50:56.863952 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/143cd811-0dff-4141-bbb4-f940c83a0bb0-kube-api-access-ppdns" (OuterVolumeSpecName: "kube-api-access-ppdns") pod "143cd811-0dff-4141-bbb4-f940c83a0bb0" (UID: "143cd811-0dff-4141-bbb4-f940c83a0bb0"). InnerVolumeSpecName "kube-api-access-ppdns". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:50:56 crc kubenswrapper[4754]: I0126 16:50:56.958573 4754 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/143cd811-0dff-4141-bbb4-f940c83a0bb0-utilities\") on node \"crc\" DevicePath \"\"" Jan 26 16:50:56 crc kubenswrapper[4754]: I0126 16:50:56.958817 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ppdns\" (UniqueName: \"kubernetes.io/projected/143cd811-0dff-4141-bbb4-f940c83a0bb0-kube-api-access-ppdns\") on node \"crc\" DevicePath \"\"" Jan 26 16:50:56 crc kubenswrapper[4754]: I0126 16:50:56.976502 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/143cd811-0dff-4141-bbb4-f940c83a0bb0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "143cd811-0dff-4141-bbb4-f940c83a0bb0" (UID: "143cd811-0dff-4141-bbb4-f940c83a0bb0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 16:50:57 crc kubenswrapper[4754]: I0126 16:50:57.060050 4754 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/143cd811-0dff-4141-bbb4-f940c83a0bb0-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 26 16:50:57 crc kubenswrapper[4754]: I0126 16:50:57.494641 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-k9sz7" Jan 26 16:50:57 crc kubenswrapper[4754]: I0126 16:50:57.564968 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/2a28f341-5f93-4d6c-bc4b-973d33dfeee1-v4-0-config-system-ocp-branding-template\") pod \"2a28f341-5f93-4d6c-bc4b-973d33dfeee1\" (UID: \"2a28f341-5f93-4d6c-bc4b-973d33dfeee1\") " Jan 26 16:50:57 crc kubenswrapper[4754]: I0126 16:50:57.565055 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/2a28f341-5f93-4d6c-bc4b-973d33dfeee1-v4-0-config-system-service-ca\") pod \"2a28f341-5f93-4d6c-bc4b-973d33dfeee1\" (UID: \"2a28f341-5f93-4d6c-bc4b-973d33dfeee1\") " Jan 26 16:50:57 crc kubenswrapper[4754]: I0126 16:50:57.565136 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/2a28f341-5f93-4d6c-bc4b-973d33dfeee1-v4-0-config-user-idp-0-file-data\") pod \"2a28f341-5f93-4d6c-bc4b-973d33dfeee1\" (UID: \"2a28f341-5f93-4d6c-bc4b-973d33dfeee1\") " Jan 26 16:50:57 crc kubenswrapper[4754]: I0126 16:50:57.565165 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/2a28f341-5f93-4d6c-bc4b-973d33dfeee1-v4-0-config-system-session\") pod \"2a28f341-5f93-4d6c-bc4b-973d33dfeee1\" (UID: \"2a28f341-5f93-4d6c-bc4b-973d33dfeee1\") " Jan 26 16:50:57 crc kubenswrapper[4754]: I0126 16:50:57.565201 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/2a28f341-5f93-4d6c-bc4b-973d33dfeee1-audit-policies\") pod \"2a28f341-5f93-4d6c-bc4b-973d33dfeee1\" (UID: \"2a28f341-5f93-4d6c-bc4b-973d33dfeee1\") " Jan 26 16:50:57 crc kubenswrapper[4754]: I0126 16:50:57.565242 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/2a28f341-5f93-4d6c-bc4b-973d33dfeee1-v4-0-config-user-template-login\") pod \"2a28f341-5f93-4d6c-bc4b-973d33dfeee1\" (UID: \"2a28f341-5f93-4d6c-bc4b-973d33dfeee1\") " Jan 26 16:50:57 crc kubenswrapper[4754]: I0126 16:50:57.565287 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/2a28f341-5f93-4d6c-bc4b-973d33dfeee1-v4-0-config-user-template-provider-selection\") pod \"2a28f341-5f93-4d6c-bc4b-973d33dfeee1\" (UID: \"2a28f341-5f93-4d6c-bc4b-973d33dfeee1\") " Jan 26 16:50:57 crc kubenswrapper[4754]: I0126 16:50:57.565309 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/2a28f341-5f93-4d6c-bc4b-973d33dfeee1-v4-0-config-system-router-certs\") pod \"2a28f341-5f93-4d6c-bc4b-973d33dfeee1\" (UID: \"2a28f341-5f93-4d6c-bc4b-973d33dfeee1\") " Jan 26 16:50:57 crc kubenswrapper[4754]: I0126 16:50:57.565380 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6mzfw\" (UniqueName: \"kubernetes.io/projected/2a28f341-5f93-4d6c-bc4b-973d33dfeee1-kube-api-access-6mzfw\") pod \"2a28f341-5f93-4d6c-bc4b-973d33dfeee1\" (UID: \"2a28f341-5f93-4d6c-bc4b-973d33dfeee1\") " Jan 26 16:50:57 crc kubenswrapper[4754]: I0126 16:50:57.565425 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/2a28f341-5f93-4d6c-bc4b-973d33dfeee1-v4-0-config-user-template-error\") pod \"2a28f341-5f93-4d6c-bc4b-973d33dfeee1\" (UID: \"2a28f341-5f93-4d6c-bc4b-973d33dfeee1\") " Jan 26 16:50:57 crc kubenswrapper[4754]: I0126 16:50:57.565483 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/2a28f341-5f93-4d6c-bc4b-973d33dfeee1-audit-dir\") pod \"2a28f341-5f93-4d6c-bc4b-973d33dfeee1\" (UID: \"2a28f341-5f93-4d6c-bc4b-973d33dfeee1\") " Jan 26 16:50:57 crc kubenswrapper[4754]: I0126 16:50:57.565521 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/2a28f341-5f93-4d6c-bc4b-973d33dfeee1-v4-0-config-system-serving-cert\") pod \"2a28f341-5f93-4d6c-bc4b-973d33dfeee1\" (UID: \"2a28f341-5f93-4d6c-bc4b-973d33dfeee1\") " Jan 26 16:50:57 crc kubenswrapper[4754]: I0126 16:50:57.565551 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/2a28f341-5f93-4d6c-bc4b-973d33dfeee1-v4-0-config-system-cliconfig\") pod \"2a28f341-5f93-4d6c-bc4b-973d33dfeee1\" (UID: \"2a28f341-5f93-4d6c-bc4b-973d33dfeee1\") " Jan 26 16:50:57 crc kubenswrapper[4754]: I0126 16:50:57.565601 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2a28f341-5f93-4d6c-bc4b-973d33dfeee1-v4-0-config-system-trusted-ca-bundle\") pod \"2a28f341-5f93-4d6c-bc4b-973d33dfeee1\" (UID: \"2a28f341-5f93-4d6c-bc4b-973d33dfeee1\") " Jan 26 16:50:57 crc kubenswrapper[4754]: I0126 16:50:57.566109 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2a28f341-5f93-4d6c-bc4b-973d33dfeee1-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "2a28f341-5f93-4d6c-bc4b-973d33dfeee1" (UID: "2a28f341-5f93-4d6c-bc4b-973d33dfeee1"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 16:50:57 crc kubenswrapper[4754]: I0126 16:50:57.566689 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2a28f341-5f93-4d6c-bc4b-973d33dfeee1-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "2a28f341-5f93-4d6c-bc4b-973d33dfeee1" (UID: "2a28f341-5f93-4d6c-bc4b-973d33dfeee1"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 16:50:57 crc kubenswrapper[4754]: I0126 16:50:57.566804 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2a28f341-5f93-4d6c-bc4b-973d33dfeee1-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "2a28f341-5f93-4d6c-bc4b-973d33dfeee1" (UID: "2a28f341-5f93-4d6c-bc4b-973d33dfeee1"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 16:50:57 crc kubenswrapper[4754]: I0126 16:50:57.567195 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2a28f341-5f93-4d6c-bc4b-973d33dfeee1-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "2a28f341-5f93-4d6c-bc4b-973d33dfeee1" (UID: "2a28f341-5f93-4d6c-bc4b-973d33dfeee1"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 16:50:57 crc kubenswrapper[4754]: I0126 16:50:57.569708 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2a28f341-5f93-4d6c-bc4b-973d33dfeee1-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "2a28f341-5f93-4d6c-bc4b-973d33dfeee1" (UID: "2a28f341-5f93-4d6c-bc4b-973d33dfeee1"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 16:50:57 crc kubenswrapper[4754]: I0126 16:50:57.570004 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2a28f341-5f93-4d6c-bc4b-973d33dfeee1-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "2a28f341-5f93-4d6c-bc4b-973d33dfeee1" (UID: "2a28f341-5f93-4d6c-bc4b-973d33dfeee1"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 16:50:57 crc kubenswrapper[4754]: I0126 16:50:57.570459 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2a28f341-5f93-4d6c-bc4b-973d33dfeee1-kube-api-access-6mzfw" (OuterVolumeSpecName: "kube-api-access-6mzfw") pod "2a28f341-5f93-4d6c-bc4b-973d33dfeee1" (UID: "2a28f341-5f93-4d6c-bc4b-973d33dfeee1"). InnerVolumeSpecName "kube-api-access-6mzfw". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:50:57 crc kubenswrapper[4754]: I0126 16:50:57.570545 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2a28f341-5f93-4d6c-bc4b-973d33dfeee1-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "2a28f341-5f93-4d6c-bc4b-973d33dfeee1" (UID: "2a28f341-5f93-4d6c-bc4b-973d33dfeee1"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 16:50:57 crc kubenswrapper[4754]: I0126 16:50:57.570548 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2a28f341-5f93-4d6c-bc4b-973d33dfeee1-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "2a28f341-5f93-4d6c-bc4b-973d33dfeee1" (UID: "2a28f341-5f93-4d6c-bc4b-973d33dfeee1"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 16:50:57 crc kubenswrapper[4754]: I0126 16:50:57.570567 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2a28f341-5f93-4d6c-bc4b-973d33dfeee1-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "2a28f341-5f93-4d6c-bc4b-973d33dfeee1" (UID: "2a28f341-5f93-4d6c-bc4b-973d33dfeee1"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 16:50:57 crc kubenswrapper[4754]: I0126 16:50:57.570737 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2a28f341-5f93-4d6c-bc4b-973d33dfeee1-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "2a28f341-5f93-4d6c-bc4b-973d33dfeee1" (UID: "2a28f341-5f93-4d6c-bc4b-973d33dfeee1"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 16:50:57 crc kubenswrapper[4754]: I0126 16:50:57.570887 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2a28f341-5f93-4d6c-bc4b-973d33dfeee1-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "2a28f341-5f93-4d6c-bc4b-973d33dfeee1" (UID: "2a28f341-5f93-4d6c-bc4b-973d33dfeee1"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 16:50:57 crc kubenswrapper[4754]: I0126 16:50:57.570956 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2a28f341-5f93-4d6c-bc4b-973d33dfeee1-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "2a28f341-5f93-4d6c-bc4b-973d33dfeee1" (UID: "2a28f341-5f93-4d6c-bc4b-973d33dfeee1"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 16:50:57 crc kubenswrapper[4754]: I0126 16:50:57.571088 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2a28f341-5f93-4d6c-bc4b-973d33dfeee1-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "2a28f341-5f93-4d6c-bc4b-973d33dfeee1" (UID: "2a28f341-5f93-4d6c-bc4b-973d33dfeee1"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 16:50:57 crc kubenswrapper[4754]: I0126 16:50:57.623557 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-k9sz7" event={"ID":"2a28f341-5f93-4d6c-bc4b-973d33dfeee1","Type":"ContainerDied","Data":"6547b7c8ba1bb7c76eff300312d4989db2181deca6d0c6b53251d128f9854ded"} Jan 26 16:50:57 crc kubenswrapper[4754]: I0126 16:50:57.623618 4754 scope.go:117] "RemoveContainer" containerID="a7b1a2116140f01645d1d09ad047f6edc2f84fbb9c15315f990b891c1b623394" Jan 26 16:50:57 crc kubenswrapper[4754]: I0126 16:50:57.623743 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-k9sz7" Jan 26 16:50:57 crc kubenswrapper[4754]: I0126 16:50:57.629062 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pdj77" event={"ID":"143cd811-0dff-4141-bbb4-f940c83a0bb0","Type":"ContainerDied","Data":"00d039400512dc3846805a224b072eca6624c1f0ca90c23bbeb2ddd01b9e2f2f"} Jan 26 16:50:57 crc kubenswrapper[4754]: I0126 16:50:57.629119 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pdj77" Jan 26 16:50:57 crc kubenswrapper[4754]: I0126 16:50:57.643146 4754 scope.go:117] "RemoveContainer" containerID="94f4e35dd7f774e1448690af9ee5b301d4530c4808c7492fe0bad4897dc9af36" Jan 26 16:50:57 crc kubenswrapper[4754]: I0126 16:50:57.653477 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-k9sz7"] Jan 26 16:50:57 crc kubenswrapper[4754]: I0126 16:50:57.656394 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-k9sz7"] Jan 26 16:50:57 crc kubenswrapper[4754]: I0126 16:50:57.665390 4754 scope.go:117] "RemoveContainer" containerID="68a86560e6569c8a47cba91060d966af8846227eef9b2e2217fbf2fee2618da0" Jan 26 16:50:57 crc kubenswrapper[4754]: I0126 16:50:57.670398 4754 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/2a28f341-5f93-4d6c-bc4b-973d33dfeee1-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Jan 26 16:50:57 crc kubenswrapper[4754]: I0126 16:50:57.670438 4754 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/2a28f341-5f93-4d6c-bc4b-973d33dfeee1-audit-dir\") on node \"crc\" DevicePath \"\"" Jan 26 16:50:57 crc kubenswrapper[4754]: I0126 16:50:57.670449 4754 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/2a28f341-5f93-4d6c-bc4b-973d33dfeee1-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 26 16:50:57 crc kubenswrapper[4754]: I0126 16:50:57.670459 4754 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/2a28f341-5f93-4d6c-bc4b-973d33dfeee1-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Jan 26 16:50:57 crc kubenswrapper[4754]: I0126 16:50:57.670471 4754 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2a28f341-5f93-4d6c-bc4b-973d33dfeee1-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 26 16:50:57 crc kubenswrapper[4754]: I0126 16:50:57.670481 4754 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/2a28f341-5f93-4d6c-bc4b-973d33dfeee1-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Jan 26 16:50:57 crc kubenswrapper[4754]: I0126 16:50:57.670489 4754 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/2a28f341-5f93-4d6c-bc4b-973d33dfeee1-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Jan 26 16:50:57 crc kubenswrapper[4754]: I0126 16:50:57.670499 4754 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/2a28f341-5f93-4d6c-bc4b-973d33dfeee1-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Jan 26 16:50:57 crc kubenswrapper[4754]: I0126 16:50:57.670509 4754 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/2a28f341-5f93-4d6c-bc4b-973d33dfeee1-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Jan 26 16:50:57 crc kubenswrapper[4754]: I0126 16:50:57.670520 4754 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/2a28f341-5f93-4d6c-bc4b-973d33dfeee1-audit-policies\") on node \"crc\" DevicePath \"\"" Jan 26 16:50:57 crc kubenswrapper[4754]: I0126 16:50:57.670531 4754 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/2a28f341-5f93-4d6c-bc4b-973d33dfeee1-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Jan 26 16:50:57 crc kubenswrapper[4754]: I0126 16:50:57.672983 4754 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/2a28f341-5f93-4d6c-bc4b-973d33dfeee1-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Jan 26 16:50:57 crc kubenswrapper[4754]: I0126 16:50:57.673016 4754 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/2a28f341-5f93-4d6c-bc4b-973d33dfeee1-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Jan 26 16:50:57 crc kubenswrapper[4754]: I0126 16:50:57.673036 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6mzfw\" (UniqueName: \"kubernetes.io/projected/2a28f341-5f93-4d6c-bc4b-973d33dfeee1-kube-api-access-6mzfw\") on node \"crc\" DevicePath \"\"" Jan 26 16:50:57 crc kubenswrapper[4754]: I0126 16:50:57.677429 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-pdj77"] Jan 26 16:50:57 crc kubenswrapper[4754]: I0126 16:50:57.680304 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-pdj77"] Jan 26 16:50:57 crc kubenswrapper[4754]: I0126 16:50:57.685961 4754 scope.go:117] "RemoveContainer" containerID="2f98d952e095831b8788d75c46f17948c983ef880a4fc6066e33ac5984405790" Jan 26 16:50:57 crc kubenswrapper[4754]: I0126 16:50:57.774295 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="143cd811-0dff-4141-bbb4-f940c83a0bb0" path="/var/lib/kubelet/pods/143cd811-0dff-4141-bbb4-f940c83a0bb0/volumes" Jan 26 16:50:57 crc kubenswrapper[4754]: I0126 16:50:57.775128 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2a28f341-5f93-4d6c-bc4b-973d33dfeee1" path="/var/lib/kubelet/pods/2a28f341-5f93-4d6c-bc4b-973d33dfeee1/volumes" Jan 26 16:50:59 crc kubenswrapper[4754]: I0126 16:50:59.787127 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-76cccff864-bhsgv"] Jan 26 16:50:59 crc kubenswrapper[4754]: E0126 16:50:59.787361 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a28f341-5f93-4d6c-bc4b-973d33dfeee1" containerName="oauth-openshift" Jan 26 16:50:59 crc kubenswrapper[4754]: I0126 16:50:59.787377 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a28f341-5f93-4d6c-bc4b-973d33dfeee1" containerName="oauth-openshift" Jan 26 16:50:59 crc kubenswrapper[4754]: E0126 16:50:59.787391 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="405f8b06-885b-4a7f-9f85-3af411a5504e" containerName="registry-server" Jan 26 16:50:59 crc kubenswrapper[4754]: I0126 16:50:59.787399 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="405f8b06-885b-4a7f-9f85-3af411a5504e" containerName="registry-server" Jan 26 16:50:59 crc kubenswrapper[4754]: E0126 16:50:59.787407 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="143cd811-0dff-4141-bbb4-f940c83a0bb0" containerName="extract-utilities" Jan 26 16:50:59 crc kubenswrapper[4754]: I0126 16:50:59.787416 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="143cd811-0dff-4141-bbb4-f940c83a0bb0" containerName="extract-utilities" Jan 26 16:50:59 crc kubenswrapper[4754]: E0126 16:50:59.787428 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="405f8b06-885b-4a7f-9f85-3af411a5504e" containerName="extract-content" Jan 26 16:50:59 crc kubenswrapper[4754]: I0126 16:50:59.787438 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="405f8b06-885b-4a7f-9f85-3af411a5504e" containerName="extract-content" Jan 26 16:50:59 crc kubenswrapper[4754]: E0126 16:50:59.787450 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00549dbb-6196-4495-9fbe-4f641333fbe4" containerName="registry-server" Jan 26 16:50:59 crc kubenswrapper[4754]: I0126 16:50:59.787457 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="00549dbb-6196-4495-9fbe-4f641333fbe4" containerName="registry-server" Jan 26 16:50:59 crc kubenswrapper[4754]: E0126 16:50:59.787467 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="405f8b06-885b-4a7f-9f85-3af411a5504e" containerName="extract-utilities" Jan 26 16:50:59 crc kubenswrapper[4754]: I0126 16:50:59.787473 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="405f8b06-885b-4a7f-9f85-3af411a5504e" containerName="extract-utilities" Jan 26 16:50:59 crc kubenswrapper[4754]: E0126 16:50:59.787486 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00549dbb-6196-4495-9fbe-4f641333fbe4" containerName="extract-utilities" Jan 26 16:50:59 crc kubenswrapper[4754]: I0126 16:50:59.787493 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="00549dbb-6196-4495-9fbe-4f641333fbe4" containerName="extract-utilities" Jan 26 16:50:59 crc kubenswrapper[4754]: E0126 16:50:59.787504 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ffaa8507-8be1-4bdb-b7b5-28104f9cdbd6" containerName="extract-utilities" Jan 26 16:50:59 crc kubenswrapper[4754]: I0126 16:50:59.787511 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="ffaa8507-8be1-4bdb-b7b5-28104f9cdbd6" containerName="extract-utilities" Jan 26 16:50:59 crc kubenswrapper[4754]: E0126 16:50:59.787521 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ffaa8507-8be1-4bdb-b7b5-28104f9cdbd6" containerName="registry-server" Jan 26 16:50:59 crc kubenswrapper[4754]: I0126 16:50:59.787528 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="ffaa8507-8be1-4bdb-b7b5-28104f9cdbd6" containerName="registry-server" Jan 26 16:50:59 crc kubenswrapper[4754]: E0126 16:50:59.787541 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ffaa8507-8be1-4bdb-b7b5-28104f9cdbd6" containerName="extract-content" Jan 26 16:50:59 crc kubenswrapper[4754]: I0126 16:50:59.787547 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="ffaa8507-8be1-4bdb-b7b5-28104f9cdbd6" containerName="extract-content" Jan 26 16:50:59 crc kubenswrapper[4754]: E0126 16:50:59.787557 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00549dbb-6196-4495-9fbe-4f641333fbe4" containerName="extract-content" Jan 26 16:50:59 crc kubenswrapper[4754]: I0126 16:50:59.787564 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="00549dbb-6196-4495-9fbe-4f641333fbe4" containerName="extract-content" Jan 26 16:50:59 crc kubenswrapper[4754]: E0126 16:50:59.787574 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="143cd811-0dff-4141-bbb4-f940c83a0bb0" containerName="extract-content" Jan 26 16:50:59 crc kubenswrapper[4754]: I0126 16:50:59.787581 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="143cd811-0dff-4141-bbb4-f940c83a0bb0" containerName="extract-content" Jan 26 16:50:59 crc kubenswrapper[4754]: E0126 16:50:59.787592 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="143cd811-0dff-4141-bbb4-f940c83a0bb0" containerName="registry-server" Jan 26 16:50:59 crc kubenswrapper[4754]: I0126 16:50:59.787599 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="143cd811-0dff-4141-bbb4-f940c83a0bb0" containerName="registry-server" Jan 26 16:50:59 crc kubenswrapper[4754]: I0126 16:50:59.787750 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="00549dbb-6196-4495-9fbe-4f641333fbe4" containerName="registry-server" Jan 26 16:50:59 crc kubenswrapper[4754]: I0126 16:50:59.787768 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="143cd811-0dff-4141-bbb4-f940c83a0bb0" containerName="registry-server" Jan 26 16:50:59 crc kubenswrapper[4754]: I0126 16:50:59.787778 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="2a28f341-5f93-4d6c-bc4b-973d33dfeee1" containerName="oauth-openshift" Jan 26 16:50:59 crc kubenswrapper[4754]: I0126 16:50:59.787792 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="405f8b06-885b-4a7f-9f85-3af411a5504e" containerName="registry-server" Jan 26 16:50:59 crc kubenswrapper[4754]: I0126 16:50:59.787799 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="ffaa8507-8be1-4bdb-b7b5-28104f9cdbd6" containerName="registry-server" Jan 26 16:50:59 crc kubenswrapper[4754]: I0126 16:50:59.788144 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-76cccff864-bhsgv" Jan 26 16:50:59 crc kubenswrapper[4754]: I0126 16:50:59.791321 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Jan 26 16:50:59 crc kubenswrapper[4754]: I0126 16:50:59.791571 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Jan 26 16:50:59 crc kubenswrapper[4754]: I0126 16:50:59.791770 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Jan 26 16:50:59 crc kubenswrapper[4754]: I0126 16:50:59.792755 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Jan 26 16:50:59 crc kubenswrapper[4754]: I0126 16:50:59.792942 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Jan 26 16:50:59 crc kubenswrapper[4754]: I0126 16:50:59.793087 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Jan 26 16:50:59 crc kubenswrapper[4754]: I0126 16:50:59.793614 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Jan 26 16:50:59 crc kubenswrapper[4754]: I0126 16:50:59.793859 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Jan 26 16:50:59 crc kubenswrapper[4754]: I0126 16:50:59.794019 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Jan 26 16:50:59 crc kubenswrapper[4754]: I0126 16:50:59.794149 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Jan 26 16:50:59 crc kubenswrapper[4754]: I0126 16:50:59.794282 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Jan 26 16:50:59 crc kubenswrapper[4754]: I0126 16:50:59.794415 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Jan 26 16:50:59 crc kubenswrapper[4754]: I0126 16:50:59.799453 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Jan 26 16:50:59 crc kubenswrapper[4754]: I0126 16:50:59.804982 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Jan 26 16:50:59 crc kubenswrapper[4754]: I0126 16:50:59.815541 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Jan 26 16:50:59 crc kubenswrapper[4754]: I0126 16:50:59.822213 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-76cccff864-bhsgv"] Jan 26 16:50:59 crc kubenswrapper[4754]: I0126 16:50:59.898163 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/f0dc5066-2467-42f8-a27f-af57b9d12b4f-v4-0-config-system-serving-cert\") pod \"oauth-openshift-76cccff864-bhsgv\" (UID: \"f0dc5066-2467-42f8-a27f-af57b9d12b4f\") " pod="openshift-authentication/oauth-openshift-76cccff864-bhsgv" Jan 26 16:50:59 crc kubenswrapper[4754]: I0126 16:50:59.898247 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f0dc5066-2467-42f8-a27f-af57b9d12b4f-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-76cccff864-bhsgv\" (UID: \"f0dc5066-2467-42f8-a27f-af57b9d12b4f\") " pod="openshift-authentication/oauth-openshift-76cccff864-bhsgv" Jan 26 16:50:59 crc kubenswrapper[4754]: I0126 16:50:59.898284 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/f0dc5066-2467-42f8-a27f-af57b9d12b4f-v4-0-config-system-session\") pod \"oauth-openshift-76cccff864-bhsgv\" (UID: \"f0dc5066-2467-42f8-a27f-af57b9d12b4f\") " pod="openshift-authentication/oauth-openshift-76cccff864-bhsgv" Jan 26 16:50:59 crc kubenswrapper[4754]: I0126 16:50:59.898310 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f0dc5066-2467-42f8-a27f-af57b9d12b4f-audit-dir\") pod \"oauth-openshift-76cccff864-bhsgv\" (UID: \"f0dc5066-2467-42f8-a27f-af57b9d12b4f\") " pod="openshift-authentication/oauth-openshift-76cccff864-bhsgv" Jan 26 16:50:59 crc kubenswrapper[4754]: I0126 16:50:59.898334 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/f0dc5066-2467-42f8-a27f-af57b9d12b4f-v4-0-config-user-template-login\") pod \"oauth-openshift-76cccff864-bhsgv\" (UID: \"f0dc5066-2467-42f8-a27f-af57b9d12b4f\") " pod="openshift-authentication/oauth-openshift-76cccff864-bhsgv" Jan 26 16:50:59 crc kubenswrapper[4754]: I0126 16:50:59.898380 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/f0dc5066-2467-42f8-a27f-af57b9d12b4f-v4-0-config-user-template-error\") pod \"oauth-openshift-76cccff864-bhsgv\" (UID: \"f0dc5066-2467-42f8-a27f-af57b9d12b4f\") " pod="openshift-authentication/oauth-openshift-76cccff864-bhsgv" Jan 26 16:50:59 crc kubenswrapper[4754]: I0126 16:50:59.898576 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/f0dc5066-2467-42f8-a27f-af57b9d12b4f-v4-0-config-system-cliconfig\") pod \"oauth-openshift-76cccff864-bhsgv\" (UID: \"f0dc5066-2467-42f8-a27f-af57b9d12b4f\") " pod="openshift-authentication/oauth-openshift-76cccff864-bhsgv" Jan 26 16:50:59 crc kubenswrapper[4754]: I0126 16:50:59.898626 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/f0dc5066-2467-42f8-a27f-af57b9d12b4f-v4-0-config-system-service-ca\") pod \"oauth-openshift-76cccff864-bhsgv\" (UID: \"f0dc5066-2467-42f8-a27f-af57b9d12b4f\") " pod="openshift-authentication/oauth-openshift-76cccff864-bhsgv" Jan 26 16:50:59 crc kubenswrapper[4754]: I0126 16:50:59.898844 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/f0dc5066-2467-42f8-a27f-af57b9d12b4f-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-76cccff864-bhsgv\" (UID: \"f0dc5066-2467-42f8-a27f-af57b9d12b4f\") " pod="openshift-authentication/oauth-openshift-76cccff864-bhsgv" Jan 26 16:50:59 crc kubenswrapper[4754]: I0126 16:50:59.898939 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/f0dc5066-2467-42f8-a27f-af57b9d12b4f-v4-0-config-system-router-certs\") pod \"oauth-openshift-76cccff864-bhsgv\" (UID: \"f0dc5066-2467-42f8-a27f-af57b9d12b4f\") " pod="openshift-authentication/oauth-openshift-76cccff864-bhsgv" Jan 26 16:50:59 crc kubenswrapper[4754]: I0126 16:50:59.899072 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/f0dc5066-2467-42f8-a27f-af57b9d12b4f-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-76cccff864-bhsgv\" (UID: \"f0dc5066-2467-42f8-a27f-af57b9d12b4f\") " pod="openshift-authentication/oauth-openshift-76cccff864-bhsgv" Jan 26 16:50:59 crc kubenswrapper[4754]: I0126 16:50:59.899710 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/f0dc5066-2467-42f8-a27f-af57b9d12b4f-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-76cccff864-bhsgv\" (UID: \"f0dc5066-2467-42f8-a27f-af57b9d12b4f\") " pod="openshift-authentication/oauth-openshift-76cccff864-bhsgv" Jan 26 16:50:59 crc kubenswrapper[4754]: I0126 16:50:59.899767 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dzpnn\" (UniqueName: \"kubernetes.io/projected/f0dc5066-2467-42f8-a27f-af57b9d12b4f-kube-api-access-dzpnn\") pod \"oauth-openshift-76cccff864-bhsgv\" (UID: \"f0dc5066-2467-42f8-a27f-af57b9d12b4f\") " pod="openshift-authentication/oauth-openshift-76cccff864-bhsgv" Jan 26 16:50:59 crc kubenswrapper[4754]: I0126 16:50:59.899877 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/f0dc5066-2467-42f8-a27f-af57b9d12b4f-audit-policies\") pod \"oauth-openshift-76cccff864-bhsgv\" (UID: \"f0dc5066-2467-42f8-a27f-af57b9d12b4f\") " pod="openshift-authentication/oauth-openshift-76cccff864-bhsgv" Jan 26 16:51:00 crc kubenswrapper[4754]: I0126 16:51:00.001769 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/f0dc5066-2467-42f8-a27f-af57b9d12b4f-audit-policies\") pod \"oauth-openshift-76cccff864-bhsgv\" (UID: \"f0dc5066-2467-42f8-a27f-af57b9d12b4f\") " pod="openshift-authentication/oauth-openshift-76cccff864-bhsgv" Jan 26 16:51:00 crc kubenswrapper[4754]: I0126 16:51:00.002119 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/f0dc5066-2467-42f8-a27f-af57b9d12b4f-v4-0-config-system-serving-cert\") pod \"oauth-openshift-76cccff864-bhsgv\" (UID: \"f0dc5066-2467-42f8-a27f-af57b9d12b4f\") " pod="openshift-authentication/oauth-openshift-76cccff864-bhsgv" Jan 26 16:51:00 crc kubenswrapper[4754]: I0126 16:51:00.002165 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f0dc5066-2467-42f8-a27f-af57b9d12b4f-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-76cccff864-bhsgv\" (UID: \"f0dc5066-2467-42f8-a27f-af57b9d12b4f\") " pod="openshift-authentication/oauth-openshift-76cccff864-bhsgv" Jan 26 16:51:00 crc kubenswrapper[4754]: I0126 16:51:00.002200 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/f0dc5066-2467-42f8-a27f-af57b9d12b4f-v4-0-config-system-session\") pod \"oauth-openshift-76cccff864-bhsgv\" (UID: \"f0dc5066-2467-42f8-a27f-af57b9d12b4f\") " pod="openshift-authentication/oauth-openshift-76cccff864-bhsgv" Jan 26 16:51:00 crc kubenswrapper[4754]: I0126 16:51:00.002507 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f0dc5066-2467-42f8-a27f-af57b9d12b4f-audit-dir\") pod \"oauth-openshift-76cccff864-bhsgv\" (UID: \"f0dc5066-2467-42f8-a27f-af57b9d12b4f\") " pod="openshift-authentication/oauth-openshift-76cccff864-bhsgv" Jan 26 16:51:00 crc kubenswrapper[4754]: I0126 16:51:00.002571 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/f0dc5066-2467-42f8-a27f-af57b9d12b4f-audit-policies\") pod \"oauth-openshift-76cccff864-bhsgv\" (UID: \"f0dc5066-2467-42f8-a27f-af57b9d12b4f\") " pod="openshift-authentication/oauth-openshift-76cccff864-bhsgv" Jan 26 16:51:00 crc kubenswrapper[4754]: I0126 16:51:00.002828 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f0dc5066-2467-42f8-a27f-af57b9d12b4f-audit-dir\") pod \"oauth-openshift-76cccff864-bhsgv\" (UID: \"f0dc5066-2467-42f8-a27f-af57b9d12b4f\") " pod="openshift-authentication/oauth-openshift-76cccff864-bhsgv" Jan 26 16:51:00 crc kubenswrapper[4754]: I0126 16:51:00.002906 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/f0dc5066-2467-42f8-a27f-af57b9d12b4f-v4-0-config-user-template-login\") pod \"oauth-openshift-76cccff864-bhsgv\" (UID: \"f0dc5066-2467-42f8-a27f-af57b9d12b4f\") " pod="openshift-authentication/oauth-openshift-76cccff864-bhsgv" Jan 26 16:51:00 crc kubenswrapper[4754]: I0126 16:51:00.002936 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/f0dc5066-2467-42f8-a27f-af57b9d12b4f-v4-0-config-user-template-error\") pod \"oauth-openshift-76cccff864-bhsgv\" (UID: \"f0dc5066-2467-42f8-a27f-af57b9d12b4f\") " pod="openshift-authentication/oauth-openshift-76cccff864-bhsgv" Jan 26 16:51:00 crc kubenswrapper[4754]: I0126 16:51:00.002976 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/f0dc5066-2467-42f8-a27f-af57b9d12b4f-v4-0-config-system-cliconfig\") pod \"oauth-openshift-76cccff864-bhsgv\" (UID: \"f0dc5066-2467-42f8-a27f-af57b9d12b4f\") " pod="openshift-authentication/oauth-openshift-76cccff864-bhsgv" Jan 26 16:51:00 crc kubenswrapper[4754]: I0126 16:51:00.002995 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/f0dc5066-2467-42f8-a27f-af57b9d12b4f-v4-0-config-system-service-ca\") pod \"oauth-openshift-76cccff864-bhsgv\" (UID: \"f0dc5066-2467-42f8-a27f-af57b9d12b4f\") " pod="openshift-authentication/oauth-openshift-76cccff864-bhsgv" Jan 26 16:51:00 crc kubenswrapper[4754]: I0126 16:51:00.003067 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/f0dc5066-2467-42f8-a27f-af57b9d12b4f-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-76cccff864-bhsgv\" (UID: \"f0dc5066-2467-42f8-a27f-af57b9d12b4f\") " pod="openshift-authentication/oauth-openshift-76cccff864-bhsgv" Jan 26 16:51:00 crc kubenswrapper[4754]: I0126 16:51:00.003092 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/f0dc5066-2467-42f8-a27f-af57b9d12b4f-v4-0-config-system-router-certs\") pod \"oauth-openshift-76cccff864-bhsgv\" (UID: \"f0dc5066-2467-42f8-a27f-af57b9d12b4f\") " pod="openshift-authentication/oauth-openshift-76cccff864-bhsgv" Jan 26 16:51:00 crc kubenswrapper[4754]: I0126 16:51:00.003123 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/f0dc5066-2467-42f8-a27f-af57b9d12b4f-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-76cccff864-bhsgv\" (UID: \"f0dc5066-2467-42f8-a27f-af57b9d12b4f\") " pod="openshift-authentication/oauth-openshift-76cccff864-bhsgv" Jan 26 16:51:00 crc kubenswrapper[4754]: I0126 16:51:00.003184 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/f0dc5066-2467-42f8-a27f-af57b9d12b4f-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-76cccff864-bhsgv\" (UID: \"f0dc5066-2467-42f8-a27f-af57b9d12b4f\") " pod="openshift-authentication/oauth-openshift-76cccff864-bhsgv" Jan 26 16:51:00 crc kubenswrapper[4754]: I0126 16:51:00.003206 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dzpnn\" (UniqueName: \"kubernetes.io/projected/f0dc5066-2467-42f8-a27f-af57b9d12b4f-kube-api-access-dzpnn\") pod \"oauth-openshift-76cccff864-bhsgv\" (UID: \"f0dc5066-2467-42f8-a27f-af57b9d12b4f\") " pod="openshift-authentication/oauth-openshift-76cccff864-bhsgv" Jan 26 16:51:00 crc kubenswrapper[4754]: I0126 16:51:00.003478 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f0dc5066-2467-42f8-a27f-af57b9d12b4f-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-76cccff864-bhsgv\" (UID: \"f0dc5066-2467-42f8-a27f-af57b9d12b4f\") " pod="openshift-authentication/oauth-openshift-76cccff864-bhsgv" Jan 26 16:51:00 crc kubenswrapper[4754]: I0126 16:51:00.004193 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/f0dc5066-2467-42f8-a27f-af57b9d12b4f-v4-0-config-system-cliconfig\") pod \"oauth-openshift-76cccff864-bhsgv\" (UID: \"f0dc5066-2467-42f8-a27f-af57b9d12b4f\") " pod="openshift-authentication/oauth-openshift-76cccff864-bhsgv" Jan 26 16:51:00 crc kubenswrapper[4754]: I0126 16:51:00.004819 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/f0dc5066-2467-42f8-a27f-af57b9d12b4f-v4-0-config-system-service-ca\") pod \"oauth-openshift-76cccff864-bhsgv\" (UID: \"f0dc5066-2467-42f8-a27f-af57b9d12b4f\") " pod="openshift-authentication/oauth-openshift-76cccff864-bhsgv" Jan 26 16:51:00 crc kubenswrapper[4754]: I0126 16:51:00.007692 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/f0dc5066-2467-42f8-a27f-af57b9d12b4f-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-76cccff864-bhsgv\" (UID: \"f0dc5066-2467-42f8-a27f-af57b9d12b4f\") " pod="openshift-authentication/oauth-openshift-76cccff864-bhsgv" Jan 26 16:51:00 crc kubenswrapper[4754]: I0126 16:51:00.008931 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/f0dc5066-2467-42f8-a27f-af57b9d12b4f-v4-0-config-system-serving-cert\") pod \"oauth-openshift-76cccff864-bhsgv\" (UID: \"f0dc5066-2467-42f8-a27f-af57b9d12b4f\") " pod="openshift-authentication/oauth-openshift-76cccff864-bhsgv" Jan 26 16:51:00 crc kubenswrapper[4754]: I0126 16:51:00.009713 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/f0dc5066-2467-42f8-a27f-af57b9d12b4f-v4-0-config-system-session\") pod \"oauth-openshift-76cccff864-bhsgv\" (UID: \"f0dc5066-2467-42f8-a27f-af57b9d12b4f\") " pod="openshift-authentication/oauth-openshift-76cccff864-bhsgv" Jan 26 16:51:00 crc kubenswrapper[4754]: I0126 16:51:00.009957 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/f0dc5066-2467-42f8-a27f-af57b9d12b4f-v4-0-config-user-template-error\") pod \"oauth-openshift-76cccff864-bhsgv\" (UID: \"f0dc5066-2467-42f8-a27f-af57b9d12b4f\") " pod="openshift-authentication/oauth-openshift-76cccff864-bhsgv" Jan 26 16:51:00 crc kubenswrapper[4754]: I0126 16:51:00.010004 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/f0dc5066-2467-42f8-a27f-af57b9d12b4f-v4-0-config-system-router-certs\") pod \"oauth-openshift-76cccff864-bhsgv\" (UID: \"f0dc5066-2467-42f8-a27f-af57b9d12b4f\") " pod="openshift-authentication/oauth-openshift-76cccff864-bhsgv" Jan 26 16:51:00 crc kubenswrapper[4754]: I0126 16:51:00.010614 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/f0dc5066-2467-42f8-a27f-af57b9d12b4f-v4-0-config-user-template-login\") pod \"oauth-openshift-76cccff864-bhsgv\" (UID: \"f0dc5066-2467-42f8-a27f-af57b9d12b4f\") " pod="openshift-authentication/oauth-openshift-76cccff864-bhsgv" Jan 26 16:51:00 crc kubenswrapper[4754]: I0126 16:51:00.010959 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/f0dc5066-2467-42f8-a27f-af57b9d12b4f-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-76cccff864-bhsgv\" (UID: \"f0dc5066-2467-42f8-a27f-af57b9d12b4f\") " pod="openshift-authentication/oauth-openshift-76cccff864-bhsgv" Jan 26 16:51:00 crc kubenswrapper[4754]: I0126 16:51:00.012724 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/f0dc5066-2467-42f8-a27f-af57b9d12b4f-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-76cccff864-bhsgv\" (UID: \"f0dc5066-2467-42f8-a27f-af57b9d12b4f\") " pod="openshift-authentication/oauth-openshift-76cccff864-bhsgv" Jan 26 16:51:00 crc kubenswrapper[4754]: I0126 16:51:00.021340 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dzpnn\" (UniqueName: \"kubernetes.io/projected/f0dc5066-2467-42f8-a27f-af57b9d12b4f-kube-api-access-dzpnn\") pod \"oauth-openshift-76cccff864-bhsgv\" (UID: \"f0dc5066-2467-42f8-a27f-af57b9d12b4f\") " pod="openshift-authentication/oauth-openshift-76cccff864-bhsgv" Jan 26 16:51:00 crc kubenswrapper[4754]: I0126 16:51:00.119930 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-76cccff864-bhsgv" Jan 26 16:51:00 crc kubenswrapper[4754]: I0126 16:51:00.506277 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-76cccff864-bhsgv"] Jan 26 16:51:00 crc kubenswrapper[4754]: W0126 16:51:00.512698 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf0dc5066_2467_42f8_a27f_af57b9d12b4f.slice/crio-472823974b5cbd6d0d1a18de6c04fd7ece8348dca20af7b75a07bf6bca8603a7 WatchSource:0}: Error finding container 472823974b5cbd6d0d1a18de6c04fd7ece8348dca20af7b75a07bf6bca8603a7: Status 404 returned error can't find the container with id 472823974b5cbd6d0d1a18de6c04fd7ece8348dca20af7b75a07bf6bca8603a7 Jan 26 16:51:00 crc kubenswrapper[4754]: I0126 16:51:00.647605 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-76cccff864-bhsgv" event={"ID":"f0dc5066-2467-42f8-a27f-af57b9d12b4f","Type":"ContainerStarted","Data":"472823974b5cbd6d0d1a18de6c04fd7ece8348dca20af7b75a07bf6bca8603a7"} Jan 26 16:51:01 crc kubenswrapper[4754]: I0126 16:51:01.657878 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-76cccff864-bhsgv" event={"ID":"f0dc5066-2467-42f8-a27f-af57b9d12b4f","Type":"ContainerStarted","Data":"58fc0cadf85c929244907c07c8b856401785e15b9d2082180650cca4bf79be12"} Jan 26 16:51:01 crc kubenswrapper[4754]: I0126 16:51:01.659905 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-76cccff864-bhsgv" Jan 26 16:51:01 crc kubenswrapper[4754]: I0126 16:51:01.665966 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-76cccff864-bhsgv" Jan 26 16:51:01 crc kubenswrapper[4754]: I0126 16:51:01.680802 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-76cccff864-bhsgv" podStartSLOduration=33.6807836 podStartE2EDuration="33.6807836s" podCreationTimestamp="2026-01-26 16:50:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 16:51:01.678996779 +0000 UTC m=+228.203177213" watchObservedRunningTime="2026-01-26 16:51:01.6807836 +0000 UTC m=+228.204964034" Jan 26 16:51:05 crc kubenswrapper[4754]: I0126 16:51:05.338315 4754 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Jan 26 16:51:05 crc kubenswrapper[4754]: I0126 16:51:05.339933 4754 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Jan 26 16:51:05 crc kubenswrapper[4754]: I0126 16:51:05.340114 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 26 16:51:05 crc kubenswrapper[4754]: I0126 16:51:05.340176 4754 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Jan 26 16:51:05 crc kubenswrapper[4754]: I0126 16:51:05.340390 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" containerID="cri-o://b9c5f40188fda14b14128514fe84bddfe8dcede3d6c560283ccb2ac0c66895fc" gracePeriod=15 Jan 26 16:51:05 crc kubenswrapper[4754]: I0126 16:51:05.340484 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://27b74e2e65a547e0e28fddd4f15feacb17f4d57596eeb8da82df9d39243c85e0" gracePeriod=15 Jan 26 16:51:05 crc kubenswrapper[4754]: E0126 16:51:05.340593 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Jan 26 16:51:05 crc kubenswrapper[4754]: I0126 16:51:05.340610 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Jan 26 16:51:05 crc kubenswrapper[4754]: I0126 16:51:05.340608 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" containerID="cri-o://085fa7b8d4fef8b2722fac8c346107699b5a6bd33fbfc74ed35831cb8fdf9dad" gracePeriod=15 Jan 26 16:51:05 crc kubenswrapper[4754]: E0126 16:51:05.340621 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Jan 26 16:51:05 crc kubenswrapper[4754]: I0126 16:51:05.340628 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Jan 26 16:51:05 crc kubenswrapper[4754]: I0126 16:51:05.340630 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" containerID="cri-o://4f45445cbf1d180d6d4dd9588e85bf678f043bfce8cff0812b78ca3510904f9a" gracePeriod=15 Jan 26 16:51:05 crc kubenswrapper[4754]: E0126 16:51:05.340640 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Jan 26 16:51:05 crc kubenswrapper[4754]: I0126 16:51:05.340807 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Jan 26 16:51:05 crc kubenswrapper[4754]: E0126 16:51:05.340825 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Jan 26 16:51:05 crc kubenswrapper[4754]: I0126 16:51:05.340515 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" containerID="cri-o://8c40e018a0979035827963efb9efc48a878a55b9df68ec1a9242d7cff64fe480" gracePeriod=15 Jan 26 16:51:05 crc kubenswrapper[4754]: I0126 16:51:05.340844 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Jan 26 16:51:05 crc kubenswrapper[4754]: E0126 16:51:05.340852 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Jan 26 16:51:05 crc kubenswrapper[4754]: I0126 16:51:05.340860 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Jan 26 16:51:05 crc kubenswrapper[4754]: E0126 16:51:05.340873 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Jan 26 16:51:05 crc kubenswrapper[4754]: I0126 16:51:05.340880 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Jan 26 16:51:05 crc kubenswrapper[4754]: E0126 16:51:05.340900 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Jan 26 16:51:05 crc kubenswrapper[4754]: I0126 16:51:05.340907 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Jan 26 16:51:05 crc kubenswrapper[4754]: I0126 16:51:05.341042 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Jan 26 16:51:05 crc kubenswrapper[4754]: I0126 16:51:05.341061 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Jan 26 16:51:05 crc kubenswrapper[4754]: I0126 16:51:05.341074 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Jan 26 16:51:05 crc kubenswrapper[4754]: I0126 16:51:05.341085 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Jan 26 16:51:05 crc kubenswrapper[4754]: I0126 16:51:05.341097 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Jan 26 16:51:05 crc kubenswrapper[4754]: I0126 16:51:05.341106 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Jan 26 16:51:05 crc kubenswrapper[4754]: I0126 16:51:05.345802 4754 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="f4b27818a5e8e43d0dc095d08835c792" podUID="71bb4a3aecc4ba5b26c4b7318770ce13" Jan 26 16:51:05 crc kubenswrapper[4754]: I0126 16:51:05.376242 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Jan 26 16:51:05 crc kubenswrapper[4754]: I0126 16:51:05.473905 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 26 16:51:05 crc kubenswrapper[4754]: I0126 16:51:05.474150 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 26 16:51:05 crc kubenswrapper[4754]: I0126 16:51:05.474187 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 26 16:51:05 crc kubenswrapper[4754]: I0126 16:51:05.474214 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 26 16:51:05 crc kubenswrapper[4754]: I0126 16:51:05.474251 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 26 16:51:05 crc kubenswrapper[4754]: I0126 16:51:05.474287 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 26 16:51:05 crc kubenswrapper[4754]: I0126 16:51:05.474314 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 26 16:51:05 crc kubenswrapper[4754]: I0126 16:51:05.474357 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 26 16:51:05 crc kubenswrapper[4754]: I0126 16:51:05.575953 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 26 16:51:05 crc kubenswrapper[4754]: I0126 16:51:05.576314 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 26 16:51:05 crc kubenswrapper[4754]: I0126 16:51:05.576345 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 26 16:51:05 crc kubenswrapper[4754]: I0126 16:51:05.576375 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 26 16:51:05 crc kubenswrapper[4754]: I0126 16:51:05.576398 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 26 16:51:05 crc kubenswrapper[4754]: I0126 16:51:05.576419 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 26 16:51:05 crc kubenswrapper[4754]: I0126 16:51:05.576452 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 26 16:51:05 crc kubenswrapper[4754]: I0126 16:51:05.576487 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 26 16:51:05 crc kubenswrapper[4754]: I0126 16:51:05.576572 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 26 16:51:05 crc kubenswrapper[4754]: I0126 16:51:05.576069 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 26 16:51:05 crc kubenswrapper[4754]: I0126 16:51:05.576629 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 26 16:51:05 crc kubenswrapper[4754]: I0126 16:51:05.576658 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 26 16:51:05 crc kubenswrapper[4754]: I0126 16:51:05.576725 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 26 16:51:05 crc kubenswrapper[4754]: I0126 16:51:05.576758 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 26 16:51:05 crc kubenswrapper[4754]: I0126 16:51:05.576772 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 26 16:51:05 crc kubenswrapper[4754]: I0126 16:51:05.576854 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 26 16:51:05 crc kubenswrapper[4754]: I0126 16:51:05.671138 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 26 16:51:05 crc kubenswrapper[4754]: I0126 16:51:05.680469 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Jan 26 16:51:05 crc kubenswrapper[4754]: I0126 16:51:05.681678 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Jan 26 16:51:05 crc kubenswrapper[4754]: I0126 16:51:05.682336 4754 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="8c40e018a0979035827963efb9efc48a878a55b9df68ec1a9242d7cff64fe480" exitCode=0 Jan 26 16:51:05 crc kubenswrapper[4754]: I0126 16:51:05.682364 4754 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="27b74e2e65a547e0e28fddd4f15feacb17f4d57596eeb8da82df9d39243c85e0" exitCode=0 Jan 26 16:51:05 crc kubenswrapper[4754]: I0126 16:51:05.682375 4754 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="4f45445cbf1d180d6d4dd9588e85bf678f043bfce8cff0812b78ca3510904f9a" exitCode=0 Jan 26 16:51:05 crc kubenswrapper[4754]: I0126 16:51:05.682383 4754 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="085fa7b8d4fef8b2722fac8c346107699b5a6bd33fbfc74ed35831cb8fdf9dad" exitCode=2 Jan 26 16:51:05 crc kubenswrapper[4754]: I0126 16:51:05.682433 4754 scope.go:117] "RemoveContainer" containerID="8356f76fbc3a4e636c12d1da5455d3f6767d9b83ce2f2925a11435d7d7261e9f" Jan 26 16:51:05 crc kubenswrapper[4754]: I0126 16:51:05.685276 4754 generic.go:334] "Generic (PLEG): container finished" podID="ec0484bb-a715-4082-b6a3-c24200aea47d" containerID="49eaa3ee0155d60c0b3e7836c370d4011c2bc047a24f9c4f96cdb785cc23e842" exitCode=0 Jan 26 16:51:05 crc kubenswrapper[4754]: I0126 16:51:05.685329 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"ec0484bb-a715-4082-b6a3-c24200aea47d","Type":"ContainerDied","Data":"49eaa3ee0155d60c0b3e7836c370d4011c2bc047a24f9c4f96cdb785cc23e842"} Jan 26 16:51:05 crc kubenswrapper[4754]: I0126 16:51:05.686261 4754 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.236:6443: connect: connection refused" Jan 26 16:51:05 crc kubenswrapper[4754]: I0126 16:51:05.686752 4754 status_manager.go:851] "Failed to get status for pod" podUID="ec0484bb-a715-4082-b6a3-c24200aea47d" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.236:6443: connect: connection refused" Jan 26 16:51:05 crc kubenswrapper[4754]: E0126 16:51:05.694801 4754 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.236:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.188e5600f3d902c6 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-01-26 16:51:05.694257862 +0000 UTC m=+232.218438296,LastTimestamp:2026-01-26 16:51:05.694257862 +0000 UTC m=+232.218438296,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Jan 26 16:51:06 crc kubenswrapper[4754]: I0126 16:51:06.693539 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Jan 26 16:51:06 crc kubenswrapper[4754]: I0126 16:51:06.696689 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"5d9edf9c246049d79a08dffa69380855cc2d0279daea7432bf86307a01b371c3"} Jan 26 16:51:06 crc kubenswrapper[4754]: I0126 16:51:06.696750 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"b22d92432cf0d0794e6c68982f9e2fad37372eb21861a104f83062379a6c148b"} Jan 26 16:51:06 crc kubenswrapper[4754]: I0126 16:51:06.698127 4754 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.236:6443: connect: connection refused" Jan 26 16:51:06 crc kubenswrapper[4754]: I0126 16:51:06.698722 4754 status_manager.go:851] "Failed to get status for pod" podUID="ec0484bb-a715-4082-b6a3-c24200aea47d" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.236:6443: connect: connection refused" Jan 26 16:51:06 crc kubenswrapper[4754]: E0126 16:51:06.846985 4754 desired_state_of_world_populator.go:312] "Error processing volume" err="error processing PVC openshift-image-registry/crc-image-registry-storage: failed to fetch PVC from API server: Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-image-registry/persistentvolumeclaims/crc-image-registry-storage\": dial tcp 38.102.83.236:6443: connect: connection refused" pod="openshift-image-registry/image-registry-697d97f7c8-47pnn" volumeName="registry-storage" Jan 26 16:51:06 crc kubenswrapper[4754]: I0126 16:51:06.950846 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Jan 26 16:51:06 crc kubenswrapper[4754]: I0126 16:51:06.951498 4754 status_manager.go:851] "Failed to get status for pod" podUID="ec0484bb-a715-4082-b6a3-c24200aea47d" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.236:6443: connect: connection refused" Jan 26 16:51:06 crc kubenswrapper[4754]: I0126 16:51:06.951861 4754 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.236:6443: connect: connection refused" Jan 26 16:51:07 crc kubenswrapper[4754]: I0126 16:51:07.093996 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ec0484bb-a715-4082-b6a3-c24200aea47d-kubelet-dir\") pod \"ec0484bb-a715-4082-b6a3-c24200aea47d\" (UID: \"ec0484bb-a715-4082-b6a3-c24200aea47d\") " Jan 26 16:51:07 crc kubenswrapper[4754]: I0126 16:51:07.094053 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ec0484bb-a715-4082-b6a3-c24200aea47d-kube-api-access\") pod \"ec0484bb-a715-4082-b6a3-c24200aea47d\" (UID: \"ec0484bb-a715-4082-b6a3-c24200aea47d\") " Jan 26 16:51:07 crc kubenswrapper[4754]: I0126 16:51:07.094256 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ec0484bb-a715-4082-b6a3-c24200aea47d-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "ec0484bb-a715-4082-b6a3-c24200aea47d" (UID: "ec0484bb-a715-4082-b6a3-c24200aea47d"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 16:51:07 crc kubenswrapper[4754]: I0126 16:51:07.094075 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/ec0484bb-a715-4082-b6a3-c24200aea47d-var-lock\") pod \"ec0484bb-a715-4082-b6a3-c24200aea47d\" (UID: \"ec0484bb-a715-4082-b6a3-c24200aea47d\") " Jan 26 16:51:07 crc kubenswrapper[4754]: I0126 16:51:07.094634 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ec0484bb-a715-4082-b6a3-c24200aea47d-var-lock" (OuterVolumeSpecName: "var-lock") pod "ec0484bb-a715-4082-b6a3-c24200aea47d" (UID: "ec0484bb-a715-4082-b6a3-c24200aea47d"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 16:51:07 crc kubenswrapper[4754]: I0126 16:51:07.094755 4754 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ec0484bb-a715-4082-b6a3-c24200aea47d-kubelet-dir\") on node \"crc\" DevicePath \"\"" Jan 26 16:51:07 crc kubenswrapper[4754]: I0126 16:51:07.094771 4754 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/ec0484bb-a715-4082-b6a3-c24200aea47d-var-lock\") on node \"crc\" DevicePath \"\"" Jan 26 16:51:07 crc kubenswrapper[4754]: I0126 16:51:07.098881 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ec0484bb-a715-4082-b6a3-c24200aea47d-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "ec0484bb-a715-4082-b6a3-c24200aea47d" (UID: "ec0484bb-a715-4082-b6a3-c24200aea47d"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:51:07 crc kubenswrapper[4754]: I0126 16:51:07.195459 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ec0484bb-a715-4082-b6a3-c24200aea47d-kube-api-access\") on node \"crc\" DevicePath \"\"" Jan 26 16:51:07 crc kubenswrapper[4754]: I0126 16:51:07.706583 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Jan 26 16:51:07 crc kubenswrapper[4754]: I0126 16:51:07.707796 4754 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="b9c5f40188fda14b14128514fe84bddfe8dcede3d6c560283ccb2ac0c66895fc" exitCode=0 Jan 26 16:51:07 crc kubenswrapper[4754]: I0126 16:51:07.707885 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="027dc212fe30a734bd5807ade87bb99437982732e20a409d945d845f8c208a4d" Jan 26 16:51:07 crc kubenswrapper[4754]: I0126 16:51:07.709712 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"ec0484bb-a715-4082-b6a3-c24200aea47d","Type":"ContainerDied","Data":"f762e0bd95df0403f8f43c124cbcc52c698973710afadcb3ccb8761efb4d9735"} Jan 26 16:51:07 crc kubenswrapper[4754]: I0126 16:51:07.709771 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f762e0bd95df0403f8f43c124cbcc52c698973710afadcb3ccb8761efb4d9735" Jan 26 16:51:07 crc kubenswrapper[4754]: I0126 16:51:07.709733 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Jan 26 16:51:07 crc kubenswrapper[4754]: I0126 16:51:07.752680 4754 status_manager.go:851] "Failed to get status for pod" podUID="ec0484bb-a715-4082-b6a3-c24200aea47d" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.236:6443: connect: connection refused" Jan 26 16:51:07 crc kubenswrapper[4754]: I0126 16:51:07.753259 4754 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.236:6443: connect: connection refused" Jan 26 16:51:07 crc kubenswrapper[4754]: I0126 16:51:07.755043 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Jan 26 16:51:07 crc kubenswrapper[4754]: I0126 16:51:07.755716 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 26 16:51:07 crc kubenswrapper[4754]: I0126 16:51:07.756187 4754 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.236:6443: connect: connection refused" Jan 26 16:51:07 crc kubenswrapper[4754]: I0126 16:51:07.756443 4754 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.236:6443: connect: connection refused" Jan 26 16:51:07 crc kubenswrapper[4754]: I0126 16:51:07.756727 4754 status_manager.go:851] "Failed to get status for pod" podUID="ec0484bb-a715-4082-b6a3-c24200aea47d" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.236:6443: connect: connection refused" Jan 26 16:51:07 crc kubenswrapper[4754]: I0126 16:51:07.802165 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Jan 26 16:51:07 crc kubenswrapper[4754]: I0126 16:51:07.802281 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 16:51:07 crc kubenswrapper[4754]: I0126 16:51:07.802309 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Jan 26 16:51:07 crc kubenswrapper[4754]: I0126 16:51:07.802329 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Jan 26 16:51:07 crc kubenswrapper[4754]: I0126 16:51:07.802386 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 16:51:07 crc kubenswrapper[4754]: I0126 16:51:07.802475 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 16:51:07 crc kubenswrapper[4754]: I0126 16:51:07.802550 4754 reconciler_common.go:293] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") on node \"crc\" DevicePath \"\"" Jan 26 16:51:07 crc kubenswrapper[4754]: I0126 16:51:07.802561 4754 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") on node \"crc\" DevicePath \"\"" Jan 26 16:51:07 crc kubenswrapper[4754]: I0126 16:51:07.802569 4754 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") on node \"crc\" DevicePath \"\"" Jan 26 16:51:08 crc kubenswrapper[4754]: E0126 16:51:08.023199 4754 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.236:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.188e5600f3d902c6 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-01-26 16:51:05.694257862 +0000 UTC m=+232.218438296,LastTimestamp:2026-01-26 16:51:05.694257862 +0000 UTC m=+232.218438296,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Jan 26 16:51:08 crc kubenswrapper[4754]: E0126 16:51:08.582578 4754 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.236:6443: connect: connection refused" Jan 26 16:51:08 crc kubenswrapper[4754]: E0126 16:51:08.583225 4754 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.236:6443: connect: connection refused" Jan 26 16:51:08 crc kubenswrapper[4754]: E0126 16:51:08.583541 4754 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.236:6443: connect: connection refused" Jan 26 16:51:08 crc kubenswrapper[4754]: E0126 16:51:08.583887 4754 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.236:6443: connect: connection refused" Jan 26 16:51:08 crc kubenswrapper[4754]: E0126 16:51:08.584114 4754 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.236:6443: connect: connection refused" Jan 26 16:51:08 crc kubenswrapper[4754]: I0126 16:51:08.584138 4754 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Jan 26 16:51:08 crc kubenswrapper[4754]: E0126 16:51:08.584346 4754 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.236:6443: connect: connection refused" interval="200ms" Jan 26 16:51:08 crc kubenswrapper[4754]: I0126 16:51:08.716281 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 26 16:51:08 crc kubenswrapper[4754]: I0126 16:51:08.717151 4754 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.236:6443: connect: connection refused" Jan 26 16:51:08 crc kubenswrapper[4754]: I0126 16:51:08.717803 4754 status_manager.go:851] "Failed to get status for pod" podUID="ec0484bb-a715-4082-b6a3-c24200aea47d" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.236:6443: connect: connection refused" Jan 26 16:51:08 crc kubenswrapper[4754]: I0126 16:51:08.718131 4754 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.236:6443: connect: connection refused" Jan 26 16:51:08 crc kubenswrapper[4754]: I0126 16:51:08.730733 4754 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.236:6443: connect: connection refused" Jan 26 16:51:08 crc kubenswrapper[4754]: I0126 16:51:08.731026 4754 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.236:6443: connect: connection refused" Jan 26 16:51:08 crc kubenswrapper[4754]: I0126 16:51:08.731340 4754 status_manager.go:851] "Failed to get status for pod" podUID="ec0484bb-a715-4082-b6a3-c24200aea47d" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.236:6443: connect: connection refused" Jan 26 16:51:08 crc kubenswrapper[4754]: E0126 16:51:08.785637 4754 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.236:6443: connect: connection refused" interval="400ms" Jan 26 16:51:08 crc kubenswrapper[4754]: E0126 16:51:08.832788 4754 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:51:08Z\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:51:08Z\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:51:08Z\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:51:08Z\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"}]}}\" for node \"crc\": Patch \"https://api-int.crc.testing:6443/api/v1/nodes/crc/status?timeout=10s\": dial tcp 38.102.83.236:6443: connect: connection refused" Jan 26 16:51:08 crc kubenswrapper[4754]: E0126 16:51:08.833093 4754 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.236:6443: connect: connection refused" Jan 26 16:51:08 crc kubenswrapper[4754]: E0126 16:51:08.833387 4754 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.236:6443: connect: connection refused" Jan 26 16:51:08 crc kubenswrapper[4754]: E0126 16:51:08.833648 4754 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.236:6443: connect: connection refused" Jan 26 16:51:08 crc kubenswrapper[4754]: E0126 16:51:08.834191 4754 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.236:6443: connect: connection refused" Jan 26 16:51:08 crc kubenswrapper[4754]: E0126 16:51:08.834213 4754 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Jan 26 16:51:09 crc kubenswrapper[4754]: E0126 16:51:09.186833 4754 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.236:6443: connect: connection refused" interval="800ms" Jan 26 16:51:09 crc kubenswrapper[4754]: I0126 16:51:09.775050 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b27818a5e8e43d0dc095d08835c792" path="/var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/volumes" Jan 26 16:51:09 crc kubenswrapper[4754]: E0126 16:51:09.988388 4754 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.236:6443: connect: connection refused" interval="1.6s" Jan 26 16:51:11 crc kubenswrapper[4754]: E0126 16:51:11.589714 4754 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.236:6443: connect: connection refused" interval="3.2s" Jan 26 16:51:13 crc kubenswrapper[4754]: I0126 16:51:13.771024 4754 status_manager.go:851] "Failed to get status for pod" podUID="ec0484bb-a715-4082-b6a3-c24200aea47d" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.236:6443: connect: connection refused" Jan 26 16:51:13 crc kubenswrapper[4754]: I0126 16:51:13.772325 4754 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.236:6443: connect: connection refused" Jan 26 16:51:14 crc kubenswrapper[4754]: E0126 16:51:14.791011 4754 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.236:6443: connect: connection refused" interval="6.4s" Jan 26 16:51:18 crc kubenswrapper[4754]: E0126 16:51:18.023878 4754 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.236:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.188e5600f3d902c6 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-01-26 16:51:05.694257862 +0000 UTC m=+232.218438296,LastTimestamp:2026-01-26 16:51:05.694257862 +0000 UTC m=+232.218438296,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Jan 26 16:51:18 crc kubenswrapper[4754]: I0126 16:51:18.770611 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Jan 26 16:51:18 crc kubenswrapper[4754]: I0126 16:51:18.770700 4754 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="a74507c941b9de4d0401c8f1d53bb79ed11359628d3405cdab7f5474a75c862e" exitCode=1 Jan 26 16:51:18 crc kubenswrapper[4754]: I0126 16:51:18.770734 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"a74507c941b9de4d0401c8f1d53bb79ed11359628d3405cdab7f5474a75c862e"} Jan 26 16:51:18 crc kubenswrapper[4754]: I0126 16:51:18.771224 4754 scope.go:117] "RemoveContainer" containerID="a74507c941b9de4d0401c8f1d53bb79ed11359628d3405cdab7f5474a75c862e" Jan 26 16:51:18 crc kubenswrapper[4754]: I0126 16:51:18.771781 4754 status_manager.go:851] "Failed to get status for pod" podUID="ec0484bb-a715-4082-b6a3-c24200aea47d" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.236:6443: connect: connection refused" Jan 26 16:51:18 crc kubenswrapper[4754]: I0126 16:51:18.772152 4754 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.236:6443: connect: connection refused" Jan 26 16:51:18 crc kubenswrapper[4754]: I0126 16:51:18.772476 4754 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.236:6443: connect: connection refused" Jan 26 16:51:19 crc kubenswrapper[4754]: E0126 16:51:18.999705 4754 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:51:18Z\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:51:18Z\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:51:18Z\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:51:18Z\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"}]}}\" for node \"crc\": Patch \"https://api-int.crc.testing:6443/api/v1/nodes/crc/status?timeout=10s\": dial tcp 38.102.83.236:6443: connect: connection refused" Jan 26 16:51:19 crc kubenswrapper[4754]: E0126 16:51:19.000525 4754 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.236:6443: connect: connection refused" Jan 26 16:51:19 crc kubenswrapper[4754]: E0126 16:51:19.001142 4754 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.236:6443: connect: connection refused" Jan 26 16:51:19 crc kubenswrapper[4754]: E0126 16:51:19.001506 4754 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.236:6443: connect: connection refused" Jan 26 16:51:19 crc kubenswrapper[4754]: E0126 16:51:19.001849 4754 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.236:6443: connect: connection refused" Jan 26 16:51:19 crc kubenswrapper[4754]: E0126 16:51:19.001881 4754 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Jan 26 16:51:19 crc kubenswrapper[4754]: I0126 16:51:19.767166 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 26 16:51:19 crc kubenswrapper[4754]: I0126 16:51:19.768251 4754 status_manager.go:851] "Failed to get status for pod" podUID="ec0484bb-a715-4082-b6a3-c24200aea47d" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.236:6443: connect: connection refused" Jan 26 16:51:19 crc kubenswrapper[4754]: I0126 16:51:19.768799 4754 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.236:6443: connect: connection refused" Jan 26 16:51:19 crc kubenswrapper[4754]: I0126 16:51:19.769197 4754 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.236:6443: connect: connection refused" Jan 26 16:51:19 crc kubenswrapper[4754]: I0126 16:51:19.783989 4754 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="b89aa081-0a2e-4757-9210-e2a6c228bee9" Jan 26 16:51:19 crc kubenswrapper[4754]: I0126 16:51:19.784024 4754 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="b89aa081-0a2e-4757-9210-e2a6c228bee9" Jan 26 16:51:19 crc kubenswrapper[4754]: E0126 16:51:19.784400 4754 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.236:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 26 16:51:19 crc kubenswrapper[4754]: I0126 16:51:19.784948 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 26 16:51:19 crc kubenswrapper[4754]: I0126 16:51:19.784951 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Jan 26 16:51:19 crc kubenswrapper[4754]: I0126 16:51:19.785071 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"c7b7a608891fcadfeb9c04aa2a83d0a200cbfa4612dcd423e4facf14dfe78d2c"} Jan 26 16:51:19 crc kubenswrapper[4754]: I0126 16:51:19.785818 4754 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.236:6443: connect: connection refused" Jan 26 16:51:19 crc kubenswrapper[4754]: I0126 16:51:19.786220 4754 status_manager.go:851] "Failed to get status for pod" podUID="ec0484bb-a715-4082-b6a3-c24200aea47d" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.236:6443: connect: connection refused" Jan 26 16:51:19 crc kubenswrapper[4754]: I0126 16:51:19.786430 4754 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.236:6443: connect: connection refused" Jan 26 16:51:20 crc kubenswrapper[4754]: I0126 16:51:20.798019 4754 generic.go:334] "Generic (PLEG): container finished" podID="71bb4a3aecc4ba5b26c4b7318770ce13" containerID="2011a1cb2ef3e3279ab2945263870077b246c32665776e560b68acf7885b4dc0" exitCode=0 Jan 26 16:51:20 crc kubenswrapper[4754]: I0126 16:51:20.798139 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerDied","Data":"2011a1cb2ef3e3279ab2945263870077b246c32665776e560b68acf7885b4dc0"} Jan 26 16:51:20 crc kubenswrapper[4754]: I0126 16:51:20.798326 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"52655e217c2fa0399366ae24049511f2bcfb5135bd114962e44c27648745cc71"} Jan 26 16:51:20 crc kubenswrapper[4754]: I0126 16:51:20.798605 4754 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="b89aa081-0a2e-4757-9210-e2a6c228bee9" Jan 26 16:51:20 crc kubenswrapper[4754]: I0126 16:51:20.798620 4754 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="b89aa081-0a2e-4757-9210-e2a6c228bee9" Jan 26 16:51:20 crc kubenswrapper[4754]: I0126 16:51:20.799406 4754 status_manager.go:851] "Failed to get status for pod" podUID="ec0484bb-a715-4082-b6a3-c24200aea47d" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.236:6443: connect: connection refused" Jan 26 16:51:20 crc kubenswrapper[4754]: E0126 16:51:20.799414 4754 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.236:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 26 16:51:20 crc kubenswrapper[4754]: I0126 16:51:20.799757 4754 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.236:6443: connect: connection refused" Jan 26 16:51:20 crc kubenswrapper[4754]: I0126 16:51:20.800146 4754 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.236:6443: connect: connection refused" Jan 26 16:51:21 crc kubenswrapper[4754]: I0126 16:51:21.811993 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"3e949ca37ad25c04301370a128bff94cb79834eab64a8865a0a60811cfe17ef6"} Jan 26 16:51:21 crc kubenswrapper[4754]: I0126 16:51:21.812314 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"240865f0ec0ecf260196eba48a8ee0faeb2ac2e727dd321e1471da90b20d1a3e"} Jan 26 16:51:21 crc kubenswrapper[4754]: I0126 16:51:21.812330 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"25e61a5c1623df2b14b04cb5b7c1ff0b960bd531a2a8bd9f41ae58df315dc4f7"} Jan 26 16:51:21 crc kubenswrapper[4754]: I0126 16:51:21.812341 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"91af8e5ab7dd055b6770cd14566e08785524de0ba207891b485a4a39befe34cb"} Jan 26 16:51:21 crc kubenswrapper[4754]: I0126 16:51:21.812352 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"f4edfe6cefe1e3ad92bce46dc0792a593429ee3961b94987bf4b65fbd276ae79"} Jan 26 16:51:21 crc kubenswrapper[4754]: I0126 16:51:21.812607 4754 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="b89aa081-0a2e-4757-9210-e2a6c228bee9" Jan 26 16:51:21 crc kubenswrapper[4754]: I0126 16:51:21.812622 4754 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="b89aa081-0a2e-4757-9210-e2a6c228bee9" Jan 26 16:51:21 crc kubenswrapper[4754]: I0126 16:51:21.812957 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 26 16:51:24 crc kubenswrapper[4754]: I0126 16:51:24.785675 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 26 16:51:24 crc kubenswrapper[4754]: I0126 16:51:24.786012 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 26 16:51:24 crc kubenswrapper[4754]: I0126 16:51:24.790952 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 26 16:51:25 crc kubenswrapper[4754]: I0126 16:51:25.457247 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 26 16:51:25 crc kubenswrapper[4754]: I0126 16:51:25.753968 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 26 16:51:25 crc kubenswrapper[4754]: I0126 16:51:25.754579 4754 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Jan 26 16:51:25 crc kubenswrapper[4754]: I0126 16:51:25.754739 4754 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Jan 26 16:51:27 crc kubenswrapper[4754]: I0126 16:51:27.667334 4754 kubelet.go:1914] "Deleted mirror pod because it is outdated" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 26 16:51:27 crc kubenswrapper[4754]: I0126 16:51:27.739826 4754 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="610e0ebe-3a25-4459-97fc-d8c386644444" Jan 26 16:51:27 crc kubenswrapper[4754]: I0126 16:51:27.840861 4754 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="b89aa081-0a2e-4757-9210-e2a6c228bee9" Jan 26 16:51:27 crc kubenswrapper[4754]: I0126 16:51:27.840890 4754 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="b89aa081-0a2e-4757-9210-e2a6c228bee9" Jan 26 16:51:27 crc kubenswrapper[4754]: I0126 16:51:27.844221 4754 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="610e0ebe-3a25-4459-97fc-d8c386644444" Jan 26 16:51:27 crc kubenswrapper[4754]: I0126 16:51:27.844637 4754 status_manager.go:308] "Container readiness changed before pod has synced" pod="openshift-kube-apiserver/kube-apiserver-crc" containerID="cri-o://f4edfe6cefe1e3ad92bce46dc0792a593429ee3961b94987bf4b65fbd276ae79" Jan 26 16:51:27 crc kubenswrapper[4754]: I0126 16:51:27.844681 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 26 16:51:28 crc kubenswrapper[4754]: I0126 16:51:28.846364 4754 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="b89aa081-0a2e-4757-9210-e2a6c228bee9" Jan 26 16:51:28 crc kubenswrapper[4754]: I0126 16:51:28.846694 4754 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="b89aa081-0a2e-4757-9210-e2a6c228bee9" Jan 26 16:51:28 crc kubenswrapper[4754]: I0126 16:51:28.849013 4754 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="610e0ebe-3a25-4459-97fc-d8c386644444" Jan 26 16:51:35 crc kubenswrapper[4754]: I0126 16:51:35.754133 4754 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Jan 26 16:51:35 crc kubenswrapper[4754]: I0126 16:51:35.754741 4754 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Jan 26 16:51:37 crc kubenswrapper[4754]: I0126 16:51:37.392488 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Jan 26 16:51:37 crc kubenswrapper[4754]: I0126 16:51:37.412416 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Jan 26 16:51:37 crc kubenswrapper[4754]: I0126 16:51:37.772890 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Jan 26 16:51:38 crc kubenswrapper[4754]: I0126 16:51:38.212091 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Jan 26 16:51:38 crc kubenswrapper[4754]: I0126 16:51:38.952612 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Jan 26 16:51:39 crc kubenswrapper[4754]: I0126 16:51:39.141280 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Jan 26 16:51:39 crc kubenswrapper[4754]: I0126 16:51:39.225100 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Jan 26 16:51:39 crc kubenswrapper[4754]: I0126 16:51:39.506596 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Jan 26 16:51:39 crc kubenswrapper[4754]: I0126 16:51:39.548382 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Jan 26 16:51:39 crc kubenswrapper[4754]: I0126 16:51:39.699002 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Jan 26 16:51:39 crc kubenswrapper[4754]: I0126 16:51:39.801955 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Jan 26 16:51:39 crc kubenswrapper[4754]: I0126 16:51:39.913463 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Jan 26 16:51:40 crc kubenswrapper[4754]: I0126 16:51:40.020508 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Jan 26 16:51:40 crc kubenswrapper[4754]: I0126 16:51:40.066474 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Jan 26 16:51:40 crc kubenswrapper[4754]: I0126 16:51:40.228846 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Jan 26 16:51:40 crc kubenswrapper[4754]: I0126 16:51:40.517804 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Jan 26 16:51:40 crc kubenswrapper[4754]: I0126 16:51:40.530288 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Jan 26 16:51:40 crc kubenswrapper[4754]: I0126 16:51:40.573605 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Jan 26 16:51:40 crc kubenswrapper[4754]: I0126 16:51:40.599389 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Jan 26 16:51:40 crc kubenswrapper[4754]: I0126 16:51:40.704716 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Jan 26 16:51:40 crc kubenswrapper[4754]: I0126 16:51:40.748317 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Jan 26 16:51:40 crc kubenswrapper[4754]: I0126 16:51:40.763256 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Jan 26 16:51:40 crc kubenswrapper[4754]: I0126 16:51:40.772805 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Jan 26 16:51:40 crc kubenswrapper[4754]: I0126 16:51:40.829540 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Jan 26 16:51:40 crc kubenswrapper[4754]: I0126 16:51:40.874210 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Jan 26 16:51:40 crc kubenswrapper[4754]: I0126 16:51:40.908861 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Jan 26 16:51:40 crc kubenswrapper[4754]: I0126 16:51:40.953177 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Jan 26 16:51:40 crc kubenswrapper[4754]: I0126 16:51:40.987030 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Jan 26 16:51:41 crc kubenswrapper[4754]: I0126 16:51:41.007414 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Jan 26 16:51:41 crc kubenswrapper[4754]: I0126 16:51:41.040898 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Jan 26 16:51:41 crc kubenswrapper[4754]: I0126 16:51:41.041844 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Jan 26 16:51:41 crc kubenswrapper[4754]: I0126 16:51:41.050862 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Jan 26 16:51:41 crc kubenswrapper[4754]: I0126 16:51:41.110818 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Jan 26 16:51:41 crc kubenswrapper[4754]: I0126 16:51:41.147086 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Jan 26 16:51:41 crc kubenswrapper[4754]: I0126 16:51:41.174307 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Jan 26 16:51:41 crc kubenswrapper[4754]: I0126 16:51:41.233997 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Jan 26 16:51:41 crc kubenswrapper[4754]: I0126 16:51:41.285737 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Jan 26 16:51:41 crc kubenswrapper[4754]: I0126 16:51:41.481100 4754 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Jan 26 16:51:41 crc kubenswrapper[4754]: I0126 16:51:41.492630 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Jan 26 16:51:41 crc kubenswrapper[4754]: I0126 16:51:41.605542 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Jan 26 16:51:41 crc kubenswrapper[4754]: I0126 16:51:41.611412 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Jan 26 16:51:41 crc kubenswrapper[4754]: I0126 16:51:41.704943 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Jan 26 16:51:41 crc kubenswrapper[4754]: I0126 16:51:41.712094 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Jan 26 16:51:41 crc kubenswrapper[4754]: I0126 16:51:41.752163 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Jan 26 16:51:41 crc kubenswrapper[4754]: I0126 16:51:41.758954 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Jan 26 16:51:41 crc kubenswrapper[4754]: I0126 16:51:41.930865 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Jan 26 16:51:42 crc kubenswrapper[4754]: I0126 16:51:42.113814 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Jan 26 16:51:42 crc kubenswrapper[4754]: I0126 16:51:42.137105 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Jan 26 16:51:42 crc kubenswrapper[4754]: I0126 16:51:42.210446 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Jan 26 16:51:42 crc kubenswrapper[4754]: I0126 16:51:42.231101 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Jan 26 16:51:42 crc kubenswrapper[4754]: I0126 16:51:42.235089 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Jan 26 16:51:42 crc kubenswrapper[4754]: I0126 16:51:42.363318 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Jan 26 16:51:42 crc kubenswrapper[4754]: I0126 16:51:42.413575 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Jan 26 16:51:42 crc kubenswrapper[4754]: I0126 16:51:42.488521 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Jan 26 16:51:42 crc kubenswrapper[4754]: I0126 16:51:42.500491 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Jan 26 16:51:42 crc kubenswrapper[4754]: I0126 16:51:42.630027 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Jan 26 16:51:42 crc kubenswrapper[4754]: I0126 16:51:42.636689 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Jan 26 16:51:42 crc kubenswrapper[4754]: I0126 16:51:42.662222 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Jan 26 16:51:42 crc kubenswrapper[4754]: I0126 16:51:42.720264 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Jan 26 16:51:42 crc kubenswrapper[4754]: I0126 16:51:42.753194 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Jan 26 16:51:42 crc kubenswrapper[4754]: I0126 16:51:42.762536 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Jan 26 16:51:42 crc kubenswrapper[4754]: I0126 16:51:42.781002 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Jan 26 16:51:42 crc kubenswrapper[4754]: I0126 16:51:42.783740 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Jan 26 16:51:42 crc kubenswrapper[4754]: I0126 16:51:42.803619 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Jan 26 16:51:42 crc kubenswrapper[4754]: I0126 16:51:42.853459 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Jan 26 16:51:42 crc kubenswrapper[4754]: I0126 16:51:42.883979 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Jan 26 16:51:42 crc kubenswrapper[4754]: I0126 16:51:42.906626 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Jan 26 16:51:42 crc kubenswrapper[4754]: I0126 16:51:42.952053 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Jan 26 16:51:43 crc kubenswrapper[4754]: I0126 16:51:43.031488 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Jan 26 16:51:43 crc kubenswrapper[4754]: I0126 16:51:43.068622 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Jan 26 16:51:43 crc kubenswrapper[4754]: I0126 16:51:43.200960 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Jan 26 16:51:43 crc kubenswrapper[4754]: I0126 16:51:43.204331 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Jan 26 16:51:43 crc kubenswrapper[4754]: I0126 16:51:43.210957 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Jan 26 16:51:43 crc kubenswrapper[4754]: I0126 16:51:43.460522 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Jan 26 16:51:43 crc kubenswrapper[4754]: I0126 16:51:43.460544 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Jan 26 16:51:43 crc kubenswrapper[4754]: I0126 16:51:43.513448 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Jan 26 16:51:43 crc kubenswrapper[4754]: I0126 16:51:43.542041 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Jan 26 16:51:43 crc kubenswrapper[4754]: I0126 16:51:43.567770 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Jan 26 16:51:43 crc kubenswrapper[4754]: I0126 16:51:43.609813 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Jan 26 16:51:43 crc kubenswrapper[4754]: I0126 16:51:43.616310 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Jan 26 16:51:43 crc kubenswrapper[4754]: I0126 16:51:43.644164 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Jan 26 16:51:43 crc kubenswrapper[4754]: I0126 16:51:43.736743 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Jan 26 16:51:43 crc kubenswrapper[4754]: I0126 16:51:43.765698 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Jan 26 16:51:43 crc kubenswrapper[4754]: I0126 16:51:43.830948 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Jan 26 16:51:43 crc kubenswrapper[4754]: I0126 16:51:43.926077 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Jan 26 16:51:43 crc kubenswrapper[4754]: I0126 16:51:43.927880 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Jan 26 16:51:44 crc kubenswrapper[4754]: I0126 16:51:44.057710 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Jan 26 16:51:44 crc kubenswrapper[4754]: I0126 16:51:44.141852 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Jan 26 16:51:44 crc kubenswrapper[4754]: I0126 16:51:44.232106 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Jan 26 16:51:44 crc kubenswrapper[4754]: I0126 16:51:44.266102 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Jan 26 16:51:44 crc kubenswrapper[4754]: I0126 16:51:44.318834 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Jan 26 16:51:44 crc kubenswrapper[4754]: I0126 16:51:44.378474 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Jan 26 16:51:44 crc kubenswrapper[4754]: I0126 16:51:44.481962 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Jan 26 16:51:44 crc kubenswrapper[4754]: I0126 16:51:44.499008 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Jan 26 16:51:44 crc kubenswrapper[4754]: I0126 16:51:44.543265 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Jan 26 16:51:44 crc kubenswrapper[4754]: I0126 16:51:44.549309 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Jan 26 16:51:44 crc kubenswrapper[4754]: I0126 16:51:44.724587 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Jan 26 16:51:44 crc kubenswrapper[4754]: I0126 16:51:44.768387 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Jan 26 16:51:44 crc kubenswrapper[4754]: I0126 16:51:44.842042 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Jan 26 16:51:44 crc kubenswrapper[4754]: I0126 16:51:44.855257 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Jan 26 16:51:45 crc kubenswrapper[4754]: I0126 16:51:45.089746 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Jan 26 16:51:45 crc kubenswrapper[4754]: I0126 16:51:45.231038 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Jan 26 16:51:45 crc kubenswrapper[4754]: I0126 16:51:45.234075 4754 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Jan 26 16:51:45 crc kubenswrapper[4754]: I0126 16:51:45.267221 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Jan 26 16:51:45 crc kubenswrapper[4754]: I0126 16:51:45.297724 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Jan 26 16:51:45 crc kubenswrapper[4754]: I0126 16:51:45.323790 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Jan 26 16:51:45 crc kubenswrapper[4754]: I0126 16:51:45.325151 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Jan 26 16:51:45 crc kubenswrapper[4754]: I0126 16:51:45.443304 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Jan 26 16:51:45 crc kubenswrapper[4754]: I0126 16:51:45.530539 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Jan 26 16:51:45 crc kubenswrapper[4754]: I0126 16:51:45.539734 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Jan 26 16:51:45 crc kubenswrapper[4754]: I0126 16:51:45.558034 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Jan 26 16:51:45 crc kubenswrapper[4754]: I0126 16:51:45.743918 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Jan 26 16:51:45 crc kubenswrapper[4754]: I0126 16:51:45.745337 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Jan 26 16:51:45 crc kubenswrapper[4754]: I0126 16:51:45.753236 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Jan 26 16:51:45 crc kubenswrapper[4754]: I0126 16:51:45.753765 4754 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Jan 26 16:51:45 crc kubenswrapper[4754]: I0126 16:51:45.753827 4754 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Jan 26 16:51:45 crc kubenswrapper[4754]: I0126 16:51:45.753881 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 26 16:51:45 crc kubenswrapper[4754]: I0126 16:51:45.754479 4754 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="kube-controller-manager" containerStatusID={"Type":"cri-o","ID":"c7b7a608891fcadfeb9c04aa2a83d0a200cbfa4612dcd423e4facf14dfe78d2c"} pod="openshift-kube-controller-manager/kube-controller-manager-crc" containerMessage="Container kube-controller-manager failed startup probe, will be restarted" Jan 26 16:51:45 crc kubenswrapper[4754]: I0126 16:51:45.754630 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" containerID="cri-o://c7b7a608891fcadfeb9c04aa2a83d0a200cbfa4612dcd423e4facf14dfe78d2c" gracePeriod=30 Jan 26 16:51:45 crc kubenswrapper[4754]: I0126 16:51:45.764697 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Jan 26 16:51:45 crc kubenswrapper[4754]: I0126 16:51:45.879646 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Jan 26 16:51:45 crc kubenswrapper[4754]: I0126 16:51:45.957607 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Jan 26 16:51:46 crc kubenswrapper[4754]: I0126 16:51:46.040433 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Jan 26 16:51:46 crc kubenswrapper[4754]: I0126 16:51:46.069691 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Jan 26 16:51:46 crc kubenswrapper[4754]: I0126 16:51:46.091543 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Jan 26 16:51:46 crc kubenswrapper[4754]: I0126 16:51:46.093480 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Jan 26 16:51:46 crc kubenswrapper[4754]: I0126 16:51:46.110413 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Jan 26 16:51:46 crc kubenswrapper[4754]: I0126 16:51:46.407238 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Jan 26 16:51:46 crc kubenswrapper[4754]: I0126 16:51:46.449402 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Jan 26 16:51:46 crc kubenswrapper[4754]: I0126 16:51:46.495161 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Jan 26 16:51:46 crc kubenswrapper[4754]: I0126 16:51:46.521338 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Jan 26 16:51:46 crc kubenswrapper[4754]: I0126 16:51:46.531699 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Jan 26 16:51:46 crc kubenswrapper[4754]: I0126 16:51:46.665501 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Jan 26 16:51:46 crc kubenswrapper[4754]: I0126 16:51:46.711859 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Jan 26 16:51:46 crc kubenswrapper[4754]: I0126 16:51:46.810875 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Jan 26 16:51:46 crc kubenswrapper[4754]: I0126 16:51:46.937544 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Jan 26 16:51:47 crc kubenswrapper[4754]: I0126 16:51:47.064274 4754 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Jan 26 16:51:47 crc kubenswrapper[4754]: I0126 16:51:47.068504 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podStartSLOduration=42.06848685 podStartE2EDuration="42.06848685s" podCreationTimestamp="2026-01-26 16:51:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 16:51:27.703685091 +0000 UTC m=+254.227865535" watchObservedRunningTime="2026-01-26 16:51:47.06848685 +0000 UTC m=+273.592667284" Jan 26 16:51:47 crc kubenswrapper[4754]: I0126 16:51:47.068913 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Jan 26 16:51:47 crc kubenswrapper[4754]: I0126 16:51:47.068957 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Jan 26 16:51:47 crc kubenswrapper[4754]: I0126 16:51:47.074048 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 26 16:51:47 crc kubenswrapper[4754]: I0126 16:51:47.102133 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=20.102114098 podStartE2EDuration="20.102114098s" podCreationTimestamp="2026-01-26 16:51:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 16:51:47.100219013 +0000 UTC m=+273.624399467" watchObservedRunningTime="2026-01-26 16:51:47.102114098 +0000 UTC m=+273.626294532" Jan 26 16:51:47 crc kubenswrapper[4754]: I0126 16:51:47.214147 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Jan 26 16:51:47 crc kubenswrapper[4754]: I0126 16:51:47.305085 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Jan 26 16:51:47 crc kubenswrapper[4754]: I0126 16:51:47.335532 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Jan 26 16:51:47 crc kubenswrapper[4754]: I0126 16:51:47.377719 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Jan 26 16:51:47 crc kubenswrapper[4754]: I0126 16:51:47.415300 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Jan 26 16:51:47 crc kubenswrapper[4754]: I0126 16:51:47.443254 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Jan 26 16:51:47 crc kubenswrapper[4754]: I0126 16:51:47.458455 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Jan 26 16:51:47 crc kubenswrapper[4754]: I0126 16:51:47.468660 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Jan 26 16:51:47 crc kubenswrapper[4754]: I0126 16:51:47.490116 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Jan 26 16:51:47 crc kubenswrapper[4754]: I0126 16:51:47.490475 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Jan 26 16:51:47 crc kubenswrapper[4754]: I0126 16:51:47.547638 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Jan 26 16:51:47 crc kubenswrapper[4754]: I0126 16:51:47.556380 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Jan 26 16:51:47 crc kubenswrapper[4754]: I0126 16:51:47.571337 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Jan 26 16:51:47 crc kubenswrapper[4754]: I0126 16:51:47.577108 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Jan 26 16:51:47 crc kubenswrapper[4754]: I0126 16:51:47.595386 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Jan 26 16:51:47 crc kubenswrapper[4754]: I0126 16:51:47.704704 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Jan 26 16:51:47 crc kubenswrapper[4754]: I0126 16:51:47.800271 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Jan 26 16:51:47 crc kubenswrapper[4754]: I0126 16:51:47.840917 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Jan 26 16:51:47 crc kubenswrapper[4754]: I0126 16:51:47.843148 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Jan 26 16:51:47 crc kubenswrapper[4754]: I0126 16:51:47.875915 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Jan 26 16:51:47 crc kubenswrapper[4754]: I0126 16:51:47.877042 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Jan 26 16:51:47 crc kubenswrapper[4754]: I0126 16:51:47.941206 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Jan 26 16:51:47 crc kubenswrapper[4754]: I0126 16:51:47.981985 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Jan 26 16:51:48 crc kubenswrapper[4754]: I0126 16:51:48.002007 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Jan 26 16:51:48 crc kubenswrapper[4754]: I0126 16:51:48.056169 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Jan 26 16:51:48 crc kubenswrapper[4754]: I0126 16:51:48.076894 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Jan 26 16:51:48 crc kubenswrapper[4754]: I0126 16:51:48.097522 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Jan 26 16:51:48 crc kubenswrapper[4754]: I0126 16:51:48.148028 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Jan 26 16:51:48 crc kubenswrapper[4754]: I0126 16:51:48.152434 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Jan 26 16:51:48 crc kubenswrapper[4754]: I0126 16:51:48.262980 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Jan 26 16:51:48 crc kubenswrapper[4754]: I0126 16:51:48.272254 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Jan 26 16:51:48 crc kubenswrapper[4754]: I0126 16:51:48.313591 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Jan 26 16:51:48 crc kubenswrapper[4754]: I0126 16:51:48.348700 4754 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Jan 26 16:51:48 crc kubenswrapper[4754]: I0126 16:51:48.426559 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Jan 26 16:51:48 crc kubenswrapper[4754]: I0126 16:51:48.444852 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Jan 26 16:51:48 crc kubenswrapper[4754]: I0126 16:51:48.455737 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Jan 26 16:51:48 crc kubenswrapper[4754]: I0126 16:51:48.480173 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Jan 26 16:51:48 crc kubenswrapper[4754]: I0126 16:51:48.487874 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Jan 26 16:51:48 crc kubenswrapper[4754]: I0126 16:51:48.492427 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Jan 26 16:51:48 crc kubenswrapper[4754]: I0126 16:51:48.556611 4754 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Jan 26 16:51:48 crc kubenswrapper[4754]: I0126 16:51:48.737371 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Jan 26 16:51:48 crc kubenswrapper[4754]: I0126 16:51:48.836398 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Jan 26 16:51:48 crc kubenswrapper[4754]: I0126 16:51:48.861412 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Jan 26 16:51:48 crc kubenswrapper[4754]: I0126 16:51:48.942643 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Jan 26 16:51:48 crc kubenswrapper[4754]: I0126 16:51:48.966229 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Jan 26 16:51:48 crc kubenswrapper[4754]: I0126 16:51:48.991299 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Jan 26 16:51:49 crc kubenswrapper[4754]: I0126 16:51:49.000641 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Jan 26 16:51:49 crc kubenswrapper[4754]: I0126 16:51:49.007958 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Jan 26 16:51:49 crc kubenswrapper[4754]: I0126 16:51:49.031993 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Jan 26 16:51:49 crc kubenswrapper[4754]: I0126 16:51:49.048754 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Jan 26 16:51:49 crc kubenswrapper[4754]: I0126 16:51:49.090327 4754 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Jan 26 16:51:49 crc kubenswrapper[4754]: I0126 16:51:49.090554 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" containerID="cri-o://5d9edf9c246049d79a08dffa69380855cc2d0279daea7432bf86307a01b371c3" gracePeriod=5 Jan 26 16:51:49 crc kubenswrapper[4754]: I0126 16:51:49.155930 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Jan 26 16:51:49 crc kubenswrapper[4754]: I0126 16:51:49.200783 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Jan 26 16:51:49 crc kubenswrapper[4754]: I0126 16:51:49.250785 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Jan 26 16:51:49 crc kubenswrapper[4754]: I0126 16:51:49.340248 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Jan 26 16:51:49 crc kubenswrapper[4754]: I0126 16:51:49.356705 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Jan 26 16:51:49 crc kubenswrapper[4754]: I0126 16:51:49.449507 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Jan 26 16:51:49 crc kubenswrapper[4754]: I0126 16:51:49.499629 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Jan 26 16:51:49 crc kubenswrapper[4754]: I0126 16:51:49.500592 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Jan 26 16:51:49 crc kubenswrapper[4754]: I0126 16:51:49.519757 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Jan 26 16:51:49 crc kubenswrapper[4754]: I0126 16:51:49.605932 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Jan 26 16:51:49 crc kubenswrapper[4754]: I0126 16:51:49.691293 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Jan 26 16:51:49 crc kubenswrapper[4754]: I0126 16:51:49.767069 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Jan 26 16:51:49 crc kubenswrapper[4754]: I0126 16:51:49.915635 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Jan 26 16:51:50 crc kubenswrapper[4754]: I0126 16:51:50.032923 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Jan 26 16:51:50 crc kubenswrapper[4754]: I0126 16:51:50.082448 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Jan 26 16:51:50 crc kubenswrapper[4754]: I0126 16:51:50.183209 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Jan 26 16:51:50 crc kubenswrapper[4754]: I0126 16:51:50.254374 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Jan 26 16:51:50 crc kubenswrapper[4754]: I0126 16:51:50.264053 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Jan 26 16:51:50 crc kubenswrapper[4754]: I0126 16:51:50.265369 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Jan 26 16:51:50 crc kubenswrapper[4754]: I0126 16:51:50.311717 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Jan 26 16:51:50 crc kubenswrapper[4754]: I0126 16:51:50.378963 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Jan 26 16:51:50 crc kubenswrapper[4754]: I0126 16:51:50.416771 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Jan 26 16:51:50 crc kubenswrapper[4754]: I0126 16:51:50.475191 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Jan 26 16:51:50 crc kubenswrapper[4754]: I0126 16:51:50.478231 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Jan 26 16:51:50 crc kubenswrapper[4754]: I0126 16:51:50.499958 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Jan 26 16:51:50 crc kubenswrapper[4754]: I0126 16:51:50.548102 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Jan 26 16:51:50 crc kubenswrapper[4754]: I0126 16:51:50.639758 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Jan 26 16:51:50 crc kubenswrapper[4754]: I0126 16:51:50.961191 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Jan 26 16:51:51 crc kubenswrapper[4754]: I0126 16:51:51.130998 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Jan 26 16:51:51 crc kubenswrapper[4754]: I0126 16:51:51.150711 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Jan 26 16:51:51 crc kubenswrapper[4754]: I0126 16:51:51.175849 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Jan 26 16:51:51 crc kubenswrapper[4754]: I0126 16:51:51.180257 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Jan 26 16:51:51 crc kubenswrapper[4754]: I0126 16:51:51.293861 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Jan 26 16:51:51 crc kubenswrapper[4754]: I0126 16:51:51.317487 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Jan 26 16:51:51 crc kubenswrapper[4754]: I0126 16:51:51.321740 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Jan 26 16:51:51 crc kubenswrapper[4754]: I0126 16:51:51.344572 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Jan 26 16:51:51 crc kubenswrapper[4754]: I0126 16:51:51.431544 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Jan 26 16:51:51 crc kubenswrapper[4754]: I0126 16:51:51.709999 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Jan 26 16:51:51 crc kubenswrapper[4754]: I0126 16:51:51.896010 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Jan 26 16:51:51 crc kubenswrapper[4754]: I0126 16:51:51.963298 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Jan 26 16:51:52 crc kubenswrapper[4754]: I0126 16:51:52.008909 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Jan 26 16:51:52 crc kubenswrapper[4754]: I0126 16:51:52.236714 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Jan 26 16:51:52 crc kubenswrapper[4754]: I0126 16:51:52.258517 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Jan 26 16:51:52 crc kubenswrapper[4754]: I0126 16:51:52.326092 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Jan 26 16:51:52 crc kubenswrapper[4754]: I0126 16:51:52.620313 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Jan 26 16:51:52 crc kubenswrapper[4754]: I0126 16:51:52.670768 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Jan 26 16:51:52 crc kubenswrapper[4754]: I0126 16:51:52.743109 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Jan 26 16:51:52 crc kubenswrapper[4754]: I0126 16:51:52.743527 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Jan 26 16:51:52 crc kubenswrapper[4754]: I0126 16:51:52.927633 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Jan 26 16:51:53 crc kubenswrapper[4754]: I0126 16:51:53.065328 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Jan 26 16:51:53 crc kubenswrapper[4754]: I0126 16:51:53.106209 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Jan 26 16:51:53 crc kubenswrapper[4754]: I0126 16:51:53.239431 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Jan 26 16:51:53 crc kubenswrapper[4754]: I0126 16:51:53.298868 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Jan 26 16:51:53 crc kubenswrapper[4754]: I0126 16:51:53.320253 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Jan 26 16:51:53 crc kubenswrapper[4754]: I0126 16:51:53.364911 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Jan 26 16:51:53 crc kubenswrapper[4754]: I0126 16:51:53.383143 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Jan 26 16:51:53 crc kubenswrapper[4754]: I0126 16:51:53.403127 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Jan 26 16:51:53 crc kubenswrapper[4754]: I0126 16:51:53.412126 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Jan 26 16:51:53 crc kubenswrapper[4754]: I0126 16:51:53.451237 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Jan 26 16:51:53 crc kubenswrapper[4754]: I0126 16:51:53.629980 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Jan 26 16:51:53 crc kubenswrapper[4754]: I0126 16:51:53.858689 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Jan 26 16:51:54 crc kubenswrapper[4754]: I0126 16:51:54.184464 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Jan 26 16:51:54 crc kubenswrapper[4754]: I0126 16:51:54.250058 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Jan 26 16:51:54 crc kubenswrapper[4754]: I0126 16:51:54.625971 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Jan 26 16:51:54 crc kubenswrapper[4754]: I0126 16:51:54.656152 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Jan 26 16:51:54 crc kubenswrapper[4754]: I0126 16:51:54.656240 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 26 16:51:54 crc kubenswrapper[4754]: I0126 16:51:54.790425 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Jan 26 16:51:54 crc kubenswrapper[4754]: I0126 16:51:54.790959 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Jan 26 16:51:54 crc kubenswrapper[4754]: I0126 16:51:54.791199 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Jan 26 16:51:54 crc kubenswrapper[4754]: I0126 16:51:54.790603 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests" (OuterVolumeSpecName: "manifests") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 16:51:54 crc kubenswrapper[4754]: I0126 16:51:54.791387 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock" (OuterVolumeSpecName: "var-lock") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 16:51:54 crc kubenswrapper[4754]: I0126 16:51:54.791437 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Jan 26 16:51:54 crc kubenswrapper[4754]: I0126 16:51:54.791690 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Jan 26 16:51:54 crc kubenswrapper[4754]: I0126 16:51:54.791768 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 16:51:54 crc kubenswrapper[4754]: I0126 16:51:54.792082 4754 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") on node \"crc\" DevicePath \"\"" Jan 26 16:51:54 crc kubenswrapper[4754]: I0126 16:51:54.792269 4754 reconciler_common.go:293] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") on node \"crc\" DevicePath \"\"" Jan 26 16:51:54 crc kubenswrapper[4754]: I0126 16:51:54.792297 4754 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") on node \"crc\" DevicePath \"\"" Jan 26 16:51:54 crc kubenswrapper[4754]: I0126 16:51:54.792083 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log" (OuterVolumeSpecName: "var-log") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 16:51:54 crc kubenswrapper[4754]: I0126 16:51:54.800817 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 16:51:54 crc kubenswrapper[4754]: I0126 16:51:54.893026 4754 reconciler_common.go:293] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") on node \"crc\" DevicePath \"\"" Jan 26 16:51:54 crc kubenswrapper[4754]: I0126 16:51:54.893069 4754 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") on node \"crc\" DevicePath \"\"" Jan 26 16:51:54 crc kubenswrapper[4754]: I0126 16:51:54.982511 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Jan 26 16:51:54 crc kubenswrapper[4754]: I0126 16:51:54.983231 4754 generic.go:334] "Generic (PLEG): container finished" podID="f85e55b1a89d02b0cb034b1ea31ed45a" containerID="5d9edf9c246049d79a08dffa69380855cc2d0279daea7432bf86307a01b371c3" exitCode=137 Jan 26 16:51:54 crc kubenswrapper[4754]: I0126 16:51:54.983325 4754 scope.go:117] "RemoveContainer" containerID="5d9edf9c246049d79a08dffa69380855cc2d0279daea7432bf86307a01b371c3" Jan 26 16:51:54 crc kubenswrapper[4754]: I0126 16:51:54.983365 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 26 16:51:55 crc kubenswrapper[4754]: I0126 16:51:55.004088 4754 scope.go:117] "RemoveContainer" containerID="5d9edf9c246049d79a08dffa69380855cc2d0279daea7432bf86307a01b371c3" Jan 26 16:51:55 crc kubenswrapper[4754]: E0126 16:51:55.004655 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5d9edf9c246049d79a08dffa69380855cc2d0279daea7432bf86307a01b371c3\": container with ID starting with 5d9edf9c246049d79a08dffa69380855cc2d0279daea7432bf86307a01b371c3 not found: ID does not exist" containerID="5d9edf9c246049d79a08dffa69380855cc2d0279daea7432bf86307a01b371c3" Jan 26 16:51:55 crc kubenswrapper[4754]: I0126 16:51:55.004699 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5d9edf9c246049d79a08dffa69380855cc2d0279daea7432bf86307a01b371c3"} err="failed to get container status \"5d9edf9c246049d79a08dffa69380855cc2d0279daea7432bf86307a01b371c3\": rpc error: code = NotFound desc = could not find container \"5d9edf9c246049d79a08dffa69380855cc2d0279daea7432bf86307a01b371c3\": container with ID starting with 5d9edf9c246049d79a08dffa69380855cc2d0279daea7432bf86307a01b371c3 not found: ID does not exist" Jan 26 16:51:55 crc kubenswrapper[4754]: I0126 16:51:55.775089 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" path="/var/lib/kubelet/pods/f85e55b1a89d02b0cb034b1ea31ed45a/volumes" Jan 26 16:51:55 crc kubenswrapper[4754]: I0126 16:51:55.775527 4754 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="" Jan 26 16:51:55 crc kubenswrapper[4754]: I0126 16:51:55.784767 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Jan 26 16:51:55 crc kubenswrapper[4754]: I0126 16:51:55.784803 4754 kubelet.go:2649] "Unable to find pod for mirror pod, skipping" mirrorPod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" mirrorPodUID="12eb1d6a-2394-4275-9280-2442c39caf82" Jan 26 16:51:55 crc kubenswrapper[4754]: I0126 16:51:55.788187 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Jan 26 16:51:55 crc kubenswrapper[4754]: I0126 16:51:55.788233 4754 kubelet.go:2673] "Unable to find pod for mirror pod, skipping" mirrorPod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" mirrorPodUID="12eb1d6a-2394-4275-9280-2442c39caf82" Jan 26 16:52:02 crc kubenswrapper[4754]: I0126 16:52:02.503306 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Jan 26 16:52:06 crc kubenswrapper[4754]: I0126 16:52:06.625315 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Jan 26 16:52:08 crc kubenswrapper[4754]: I0126 16:52:08.057888 4754 generic.go:334] "Generic (PLEG): container finished" podID="f06be1d1-4134-4ad7-9233-b4099687bfe2" containerID="089c7bd3b8cd79c825b9c58787fb66cd12628763659360ce5e1b716c6f8fa7d0" exitCode=0 Jan 26 16:52:08 crc kubenswrapper[4754]: I0126 16:52:08.058005 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-7xsg2" event={"ID":"f06be1d1-4134-4ad7-9233-b4099687bfe2","Type":"ContainerDied","Data":"089c7bd3b8cd79c825b9c58787fb66cd12628763659360ce5e1b716c6f8fa7d0"} Jan 26 16:52:08 crc kubenswrapper[4754]: I0126 16:52:08.058566 4754 scope.go:117] "RemoveContainer" containerID="089c7bd3b8cd79c825b9c58787fb66cd12628763659360ce5e1b716c6f8fa7d0" Jan 26 16:52:09 crc kubenswrapper[4754]: I0126 16:52:09.064122 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-7xsg2" event={"ID":"f06be1d1-4134-4ad7-9233-b4099687bfe2","Type":"ContainerStarted","Data":"57b65fbcda96a2d7c3ad5056427ff581fc02c615544c7b716f61866017c20b02"} Jan 26 16:52:09 crc kubenswrapper[4754]: I0126 16:52:09.065481 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-7xsg2" Jan 26 16:52:09 crc kubenswrapper[4754]: I0126 16:52:09.067057 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-7xsg2" Jan 26 16:52:10 crc kubenswrapper[4754]: I0126 16:52:10.881911 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Jan 26 16:52:13 crc kubenswrapper[4754]: I0126 16:52:13.629115 4754 cert_rotation.go:91] certificate rotation detected, shutting down client connections to start using new credentials Jan 26 16:52:16 crc kubenswrapper[4754]: I0126 16:52:16.029170 4754 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Jan 26 16:52:16 crc kubenswrapper[4754]: I0126 16:52:16.106007 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/1.log" Jan 26 16:52:16 crc kubenswrapper[4754]: I0126 16:52:16.108019 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Jan 26 16:52:16 crc kubenswrapper[4754]: I0126 16:52:16.108069 4754 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="c7b7a608891fcadfeb9c04aa2a83d0a200cbfa4612dcd423e4facf14dfe78d2c" exitCode=137 Jan 26 16:52:16 crc kubenswrapper[4754]: I0126 16:52:16.108101 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"c7b7a608891fcadfeb9c04aa2a83d0a200cbfa4612dcd423e4facf14dfe78d2c"} Jan 26 16:52:16 crc kubenswrapper[4754]: I0126 16:52:16.108147 4754 scope.go:117] "RemoveContainer" containerID="a74507c941b9de4d0401c8f1d53bb79ed11359628d3405cdab7f5474a75c862e" Jan 26 16:52:17 crc kubenswrapper[4754]: I0126 16:52:17.116796 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/1.log" Jan 26 16:52:17 crc kubenswrapper[4754]: I0126 16:52:17.117909 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"52798d65672e2f789bcf8d80888e2dd617ea0132e2201b78bd133254a424659c"} Jan 26 16:52:20 crc kubenswrapper[4754]: I0126 16:52:20.371810 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Jan 26 16:52:22 crc kubenswrapper[4754]: I0126 16:52:22.361790 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Jan 26 16:52:25 crc kubenswrapper[4754]: I0126 16:52:25.457392 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 26 16:52:25 crc kubenswrapper[4754]: I0126 16:52:25.753610 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 26 16:52:25 crc kubenswrapper[4754]: I0126 16:52:25.759640 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 26 16:52:26 crc kubenswrapper[4754]: I0126 16:52:26.191384 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 26 16:52:32 crc kubenswrapper[4754]: I0126 16:52:32.884905 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-677j9"] Jan 26 16:52:32 crc kubenswrapper[4754]: I0126 16:52:32.885787 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-677j9" podUID="4d81d21a-83f5-402e-91fd-d8268a6b49be" containerName="controller-manager" containerID="cri-o://9ba363f3729d6d97c45b7a749c56183deb835a07527b3f7da097b0eabf031c27" gracePeriod=30 Jan 26 16:52:32 crc kubenswrapper[4754]: I0126 16:52:32.889381 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-m2525"] Jan 26 16:52:32 crc kubenswrapper[4754]: I0126 16:52:32.892762 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m2525" podUID="587cde2e-d0a4-406b-ab72-26c7fe4b6707" containerName="route-controller-manager" containerID="cri-o://10faa4395f7c506f3b98bb9824f69a2aedc5484d3299d85f666cb4dbbab259c8" gracePeriod=30 Jan 26 16:52:33 crc kubenswrapper[4754]: I0126 16:52:33.242394 4754 generic.go:334] "Generic (PLEG): container finished" podID="587cde2e-d0a4-406b-ab72-26c7fe4b6707" containerID="10faa4395f7c506f3b98bb9824f69a2aedc5484d3299d85f666cb4dbbab259c8" exitCode=0 Jan 26 16:52:33 crc kubenswrapper[4754]: I0126 16:52:33.242576 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m2525" event={"ID":"587cde2e-d0a4-406b-ab72-26c7fe4b6707","Type":"ContainerDied","Data":"10faa4395f7c506f3b98bb9824f69a2aedc5484d3299d85f666cb4dbbab259c8"} Jan 26 16:52:33 crc kubenswrapper[4754]: I0126 16:52:33.245916 4754 generic.go:334] "Generic (PLEG): container finished" podID="4d81d21a-83f5-402e-91fd-d8268a6b49be" containerID="9ba363f3729d6d97c45b7a749c56183deb835a07527b3f7da097b0eabf031c27" exitCode=0 Jan 26 16:52:33 crc kubenswrapper[4754]: I0126 16:52:33.245954 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-677j9" event={"ID":"4d81d21a-83f5-402e-91fd-d8268a6b49be","Type":"ContainerDied","Data":"9ba363f3729d6d97c45b7a749c56183deb835a07527b3f7da097b0eabf031c27"} Jan 26 16:52:33 crc kubenswrapper[4754]: I0126 16:52:33.377224 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-677j9" Jan 26 16:52:33 crc kubenswrapper[4754]: I0126 16:52:33.385803 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m2525" Jan 26 16:52:33 crc kubenswrapper[4754]: I0126 16:52:33.474417 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-llgrp\" (UniqueName: \"kubernetes.io/projected/587cde2e-d0a4-406b-ab72-26c7fe4b6707-kube-api-access-llgrp\") pod \"587cde2e-d0a4-406b-ab72-26c7fe4b6707\" (UID: \"587cde2e-d0a4-406b-ab72-26c7fe4b6707\") " Jan 26 16:52:33 crc kubenswrapper[4754]: I0126 16:52:33.474485 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4d81d21a-83f5-402e-91fd-d8268a6b49be-config\") pod \"4d81d21a-83f5-402e-91fd-d8268a6b49be\" (UID: \"4d81d21a-83f5-402e-91fd-d8268a6b49be\") " Jan 26 16:52:33 crc kubenswrapper[4754]: I0126 16:52:33.474507 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4d81d21a-83f5-402e-91fd-d8268a6b49be-serving-cert\") pod \"4d81d21a-83f5-402e-91fd-d8268a6b49be\" (UID: \"4d81d21a-83f5-402e-91fd-d8268a6b49be\") " Jan 26 16:52:33 crc kubenswrapper[4754]: I0126 16:52:33.474537 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4d81d21a-83f5-402e-91fd-d8268a6b49be-client-ca\") pod \"4d81d21a-83f5-402e-91fd-d8268a6b49be\" (UID: \"4d81d21a-83f5-402e-91fd-d8268a6b49be\") " Jan 26 16:52:33 crc kubenswrapper[4754]: I0126 16:52:33.474561 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/4d81d21a-83f5-402e-91fd-d8268a6b49be-proxy-ca-bundles\") pod \"4d81d21a-83f5-402e-91fd-d8268a6b49be\" (UID: \"4d81d21a-83f5-402e-91fd-d8268a6b49be\") " Jan 26 16:52:33 crc kubenswrapper[4754]: I0126 16:52:33.474590 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/587cde2e-d0a4-406b-ab72-26c7fe4b6707-client-ca\") pod \"587cde2e-d0a4-406b-ab72-26c7fe4b6707\" (UID: \"587cde2e-d0a4-406b-ab72-26c7fe4b6707\") " Jan 26 16:52:33 crc kubenswrapper[4754]: I0126 16:52:33.474617 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vdvx6\" (UniqueName: \"kubernetes.io/projected/4d81d21a-83f5-402e-91fd-d8268a6b49be-kube-api-access-vdvx6\") pod \"4d81d21a-83f5-402e-91fd-d8268a6b49be\" (UID: \"4d81d21a-83f5-402e-91fd-d8268a6b49be\") " Jan 26 16:52:33 crc kubenswrapper[4754]: I0126 16:52:33.474634 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/587cde2e-d0a4-406b-ab72-26c7fe4b6707-config\") pod \"587cde2e-d0a4-406b-ab72-26c7fe4b6707\" (UID: \"587cde2e-d0a4-406b-ab72-26c7fe4b6707\") " Jan 26 16:52:33 crc kubenswrapper[4754]: I0126 16:52:33.474650 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/587cde2e-d0a4-406b-ab72-26c7fe4b6707-serving-cert\") pod \"587cde2e-d0a4-406b-ab72-26c7fe4b6707\" (UID: \"587cde2e-d0a4-406b-ab72-26c7fe4b6707\") " Jan 26 16:52:33 crc kubenswrapper[4754]: I0126 16:52:33.476471 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4d81d21a-83f5-402e-91fd-d8268a6b49be-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "4d81d21a-83f5-402e-91fd-d8268a6b49be" (UID: "4d81d21a-83f5-402e-91fd-d8268a6b49be"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 16:52:33 crc kubenswrapper[4754]: I0126 16:52:33.476476 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/587cde2e-d0a4-406b-ab72-26c7fe4b6707-client-ca" (OuterVolumeSpecName: "client-ca") pod "587cde2e-d0a4-406b-ab72-26c7fe4b6707" (UID: "587cde2e-d0a4-406b-ab72-26c7fe4b6707"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 16:52:33 crc kubenswrapper[4754]: I0126 16:52:33.477023 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/587cde2e-d0a4-406b-ab72-26c7fe4b6707-config" (OuterVolumeSpecName: "config") pod "587cde2e-d0a4-406b-ab72-26c7fe4b6707" (UID: "587cde2e-d0a4-406b-ab72-26c7fe4b6707"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 16:52:33 crc kubenswrapper[4754]: I0126 16:52:33.478268 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4d81d21a-83f5-402e-91fd-d8268a6b49be-client-ca" (OuterVolumeSpecName: "client-ca") pod "4d81d21a-83f5-402e-91fd-d8268a6b49be" (UID: "4d81d21a-83f5-402e-91fd-d8268a6b49be"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 16:52:33 crc kubenswrapper[4754]: I0126 16:52:33.478799 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4d81d21a-83f5-402e-91fd-d8268a6b49be-config" (OuterVolumeSpecName: "config") pod "4d81d21a-83f5-402e-91fd-d8268a6b49be" (UID: "4d81d21a-83f5-402e-91fd-d8268a6b49be"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 16:52:33 crc kubenswrapper[4754]: I0126 16:52:33.481295 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4d81d21a-83f5-402e-91fd-d8268a6b49be-kube-api-access-vdvx6" (OuterVolumeSpecName: "kube-api-access-vdvx6") pod "4d81d21a-83f5-402e-91fd-d8268a6b49be" (UID: "4d81d21a-83f5-402e-91fd-d8268a6b49be"). InnerVolumeSpecName "kube-api-access-vdvx6". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:52:33 crc kubenswrapper[4754]: I0126 16:52:33.481406 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d81d21a-83f5-402e-91fd-d8268a6b49be-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "4d81d21a-83f5-402e-91fd-d8268a6b49be" (UID: "4d81d21a-83f5-402e-91fd-d8268a6b49be"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 16:52:33 crc kubenswrapper[4754]: I0126 16:52:33.482390 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/587cde2e-d0a4-406b-ab72-26c7fe4b6707-kube-api-access-llgrp" (OuterVolumeSpecName: "kube-api-access-llgrp") pod "587cde2e-d0a4-406b-ab72-26c7fe4b6707" (UID: "587cde2e-d0a4-406b-ab72-26c7fe4b6707"). InnerVolumeSpecName "kube-api-access-llgrp". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:52:33 crc kubenswrapper[4754]: I0126 16:52:33.487260 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/587cde2e-d0a4-406b-ab72-26c7fe4b6707-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "587cde2e-d0a4-406b-ab72-26c7fe4b6707" (UID: "587cde2e-d0a4-406b-ab72-26c7fe4b6707"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 16:52:33 crc kubenswrapper[4754]: I0126 16:52:33.576290 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-llgrp\" (UniqueName: \"kubernetes.io/projected/587cde2e-d0a4-406b-ab72-26c7fe4b6707-kube-api-access-llgrp\") on node \"crc\" DevicePath \"\"" Jan 26 16:52:33 crc kubenswrapper[4754]: I0126 16:52:33.576608 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-59b6cc659-w9w2t"] Jan 26 16:52:33 crc kubenswrapper[4754]: I0126 16:52:33.576631 4754 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4d81d21a-83f5-402e-91fd-d8268a6b49be-config\") on node \"crc\" DevicePath \"\"" Jan 26 16:52:33 crc kubenswrapper[4754]: I0126 16:52:33.576704 4754 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4d81d21a-83f5-402e-91fd-d8268a6b49be-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 26 16:52:33 crc kubenswrapper[4754]: I0126 16:52:33.576716 4754 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4d81d21a-83f5-402e-91fd-d8268a6b49be-client-ca\") on node \"crc\" DevicePath \"\"" Jan 26 16:52:33 crc kubenswrapper[4754]: I0126 16:52:33.576726 4754 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/4d81d21a-83f5-402e-91fd-d8268a6b49be-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Jan 26 16:52:33 crc kubenswrapper[4754]: I0126 16:52:33.576736 4754 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/587cde2e-d0a4-406b-ab72-26c7fe4b6707-client-ca\") on node \"crc\" DevicePath \"\"" Jan 26 16:52:33 crc kubenswrapper[4754]: I0126 16:52:33.576745 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vdvx6\" (UniqueName: \"kubernetes.io/projected/4d81d21a-83f5-402e-91fd-d8268a6b49be-kube-api-access-vdvx6\") on node \"crc\" DevicePath \"\"" Jan 26 16:52:33 crc kubenswrapper[4754]: I0126 16:52:33.576759 4754 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/587cde2e-d0a4-406b-ab72-26c7fe4b6707-config\") on node \"crc\" DevicePath \"\"" Jan 26 16:52:33 crc kubenswrapper[4754]: I0126 16:52:33.576768 4754 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/587cde2e-d0a4-406b-ab72-26c7fe4b6707-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 26 16:52:33 crc kubenswrapper[4754]: E0126 16:52:33.576873 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ec0484bb-a715-4082-b6a3-c24200aea47d" containerName="installer" Jan 26 16:52:33 crc kubenswrapper[4754]: I0126 16:52:33.576892 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="ec0484bb-a715-4082-b6a3-c24200aea47d" containerName="installer" Jan 26 16:52:33 crc kubenswrapper[4754]: E0126 16:52:33.576905 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d81d21a-83f5-402e-91fd-d8268a6b49be" containerName="controller-manager" Jan 26 16:52:33 crc kubenswrapper[4754]: I0126 16:52:33.576912 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d81d21a-83f5-402e-91fd-d8268a6b49be" containerName="controller-manager" Jan 26 16:52:33 crc kubenswrapper[4754]: E0126 16:52:33.576927 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Jan 26 16:52:33 crc kubenswrapper[4754]: I0126 16:52:33.576936 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Jan 26 16:52:33 crc kubenswrapper[4754]: E0126 16:52:33.576955 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="587cde2e-d0a4-406b-ab72-26c7fe4b6707" containerName="route-controller-manager" Jan 26 16:52:33 crc kubenswrapper[4754]: I0126 16:52:33.576963 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="587cde2e-d0a4-406b-ab72-26c7fe4b6707" containerName="route-controller-manager" Jan 26 16:52:33 crc kubenswrapper[4754]: I0126 16:52:33.577057 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Jan 26 16:52:33 crc kubenswrapper[4754]: I0126 16:52:33.577065 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="4d81d21a-83f5-402e-91fd-d8268a6b49be" containerName="controller-manager" Jan 26 16:52:33 crc kubenswrapper[4754]: I0126 16:52:33.577076 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="ec0484bb-a715-4082-b6a3-c24200aea47d" containerName="installer" Jan 26 16:52:33 crc kubenswrapper[4754]: I0126 16:52:33.577087 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="587cde2e-d0a4-406b-ab72-26c7fe4b6707" containerName="route-controller-manager" Jan 26 16:52:33 crc kubenswrapper[4754]: I0126 16:52:33.577495 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-59b6cc659-w9w2t" Jan 26 16:52:33 crc kubenswrapper[4754]: I0126 16:52:33.593658 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-59b6cc659-w9w2t"] Jan 26 16:52:33 crc kubenswrapper[4754]: I0126 16:52:33.677886 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/308d80ff-1c84-4fe7-93d4-c1f66a82fc91-config\") pod \"route-controller-manager-59b6cc659-w9w2t\" (UID: \"308d80ff-1c84-4fe7-93d4-c1f66a82fc91\") " pod="openshift-route-controller-manager/route-controller-manager-59b6cc659-w9w2t" Jan 26 16:52:33 crc kubenswrapper[4754]: I0126 16:52:33.677934 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/308d80ff-1c84-4fe7-93d4-c1f66a82fc91-serving-cert\") pod \"route-controller-manager-59b6cc659-w9w2t\" (UID: \"308d80ff-1c84-4fe7-93d4-c1f66a82fc91\") " pod="openshift-route-controller-manager/route-controller-manager-59b6cc659-w9w2t" Jan 26 16:52:33 crc kubenswrapper[4754]: I0126 16:52:33.677998 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6mw6c\" (UniqueName: \"kubernetes.io/projected/308d80ff-1c84-4fe7-93d4-c1f66a82fc91-kube-api-access-6mw6c\") pod \"route-controller-manager-59b6cc659-w9w2t\" (UID: \"308d80ff-1c84-4fe7-93d4-c1f66a82fc91\") " pod="openshift-route-controller-manager/route-controller-manager-59b6cc659-w9w2t" Jan 26 16:52:33 crc kubenswrapper[4754]: I0126 16:52:33.678234 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/308d80ff-1c84-4fe7-93d4-c1f66a82fc91-client-ca\") pod \"route-controller-manager-59b6cc659-w9w2t\" (UID: \"308d80ff-1c84-4fe7-93d4-c1f66a82fc91\") " pod="openshift-route-controller-manager/route-controller-manager-59b6cc659-w9w2t" Jan 26 16:52:33 crc kubenswrapper[4754]: I0126 16:52:33.779162 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/308d80ff-1c84-4fe7-93d4-c1f66a82fc91-config\") pod \"route-controller-manager-59b6cc659-w9w2t\" (UID: \"308d80ff-1c84-4fe7-93d4-c1f66a82fc91\") " pod="openshift-route-controller-manager/route-controller-manager-59b6cc659-w9w2t" Jan 26 16:52:33 crc kubenswrapper[4754]: I0126 16:52:33.779217 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/308d80ff-1c84-4fe7-93d4-c1f66a82fc91-serving-cert\") pod \"route-controller-manager-59b6cc659-w9w2t\" (UID: \"308d80ff-1c84-4fe7-93d4-c1f66a82fc91\") " pod="openshift-route-controller-manager/route-controller-manager-59b6cc659-w9w2t" Jan 26 16:52:33 crc kubenswrapper[4754]: I0126 16:52:33.779242 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6mw6c\" (UniqueName: \"kubernetes.io/projected/308d80ff-1c84-4fe7-93d4-c1f66a82fc91-kube-api-access-6mw6c\") pod \"route-controller-manager-59b6cc659-w9w2t\" (UID: \"308d80ff-1c84-4fe7-93d4-c1f66a82fc91\") " pod="openshift-route-controller-manager/route-controller-manager-59b6cc659-w9w2t" Jan 26 16:52:33 crc kubenswrapper[4754]: I0126 16:52:33.779308 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/308d80ff-1c84-4fe7-93d4-c1f66a82fc91-client-ca\") pod \"route-controller-manager-59b6cc659-w9w2t\" (UID: \"308d80ff-1c84-4fe7-93d4-c1f66a82fc91\") " pod="openshift-route-controller-manager/route-controller-manager-59b6cc659-w9w2t" Jan 26 16:52:33 crc kubenswrapper[4754]: I0126 16:52:33.780379 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/308d80ff-1c84-4fe7-93d4-c1f66a82fc91-client-ca\") pod \"route-controller-manager-59b6cc659-w9w2t\" (UID: \"308d80ff-1c84-4fe7-93d4-c1f66a82fc91\") " pod="openshift-route-controller-manager/route-controller-manager-59b6cc659-w9w2t" Jan 26 16:52:33 crc kubenswrapper[4754]: I0126 16:52:33.780530 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/308d80ff-1c84-4fe7-93d4-c1f66a82fc91-config\") pod \"route-controller-manager-59b6cc659-w9w2t\" (UID: \"308d80ff-1c84-4fe7-93d4-c1f66a82fc91\") " pod="openshift-route-controller-manager/route-controller-manager-59b6cc659-w9w2t" Jan 26 16:52:33 crc kubenswrapper[4754]: I0126 16:52:33.783253 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/308d80ff-1c84-4fe7-93d4-c1f66a82fc91-serving-cert\") pod \"route-controller-manager-59b6cc659-w9w2t\" (UID: \"308d80ff-1c84-4fe7-93d4-c1f66a82fc91\") " pod="openshift-route-controller-manager/route-controller-manager-59b6cc659-w9w2t" Jan 26 16:52:33 crc kubenswrapper[4754]: I0126 16:52:33.800129 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6mw6c\" (UniqueName: \"kubernetes.io/projected/308d80ff-1c84-4fe7-93d4-c1f66a82fc91-kube-api-access-6mw6c\") pod \"route-controller-manager-59b6cc659-w9w2t\" (UID: \"308d80ff-1c84-4fe7-93d4-c1f66a82fc91\") " pod="openshift-route-controller-manager/route-controller-manager-59b6cc659-w9w2t" Jan 26 16:52:33 crc kubenswrapper[4754]: I0126 16:52:33.893080 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-59b6cc659-w9w2t" Jan 26 16:52:34 crc kubenswrapper[4754]: I0126 16:52:34.076227 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-d5469898f-fs6b7"] Jan 26 16:52:34 crc kubenswrapper[4754]: I0126 16:52:34.077127 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-d5469898f-fs6b7" Jan 26 16:52:34 crc kubenswrapper[4754]: I0126 16:52:34.087003 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-d5469898f-fs6b7"] Jan 26 16:52:34 crc kubenswrapper[4754]: I0126 16:52:34.094613 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-59b6cc659-w9w2t"] Jan 26 16:52:34 crc kubenswrapper[4754]: I0126 16:52:34.187492 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7b233668-d1fb-40c4-bfd4-96374a7fbc81-client-ca\") pod \"controller-manager-d5469898f-fs6b7\" (UID: \"7b233668-d1fb-40c4-bfd4-96374a7fbc81\") " pod="openshift-controller-manager/controller-manager-d5469898f-fs6b7" Jan 26 16:52:34 crc kubenswrapper[4754]: I0126 16:52:34.187866 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d67wv\" (UniqueName: \"kubernetes.io/projected/7b233668-d1fb-40c4-bfd4-96374a7fbc81-kube-api-access-d67wv\") pod \"controller-manager-d5469898f-fs6b7\" (UID: \"7b233668-d1fb-40c4-bfd4-96374a7fbc81\") " pod="openshift-controller-manager/controller-manager-d5469898f-fs6b7" Jan 26 16:52:34 crc kubenswrapper[4754]: I0126 16:52:34.187899 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7b233668-d1fb-40c4-bfd4-96374a7fbc81-config\") pod \"controller-manager-d5469898f-fs6b7\" (UID: \"7b233668-d1fb-40c4-bfd4-96374a7fbc81\") " pod="openshift-controller-manager/controller-manager-d5469898f-fs6b7" Jan 26 16:52:34 crc kubenswrapper[4754]: I0126 16:52:34.187919 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7b233668-d1fb-40c4-bfd4-96374a7fbc81-proxy-ca-bundles\") pod \"controller-manager-d5469898f-fs6b7\" (UID: \"7b233668-d1fb-40c4-bfd4-96374a7fbc81\") " pod="openshift-controller-manager/controller-manager-d5469898f-fs6b7" Jan 26 16:52:34 crc kubenswrapper[4754]: I0126 16:52:34.187940 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7b233668-d1fb-40c4-bfd4-96374a7fbc81-serving-cert\") pod \"controller-manager-d5469898f-fs6b7\" (UID: \"7b233668-d1fb-40c4-bfd4-96374a7fbc81\") " pod="openshift-controller-manager/controller-manager-d5469898f-fs6b7" Jan 26 16:52:34 crc kubenswrapper[4754]: I0126 16:52:34.254698 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-59b6cc659-w9w2t" event={"ID":"308d80ff-1c84-4fe7-93d4-c1f66a82fc91","Type":"ContainerStarted","Data":"8c610acc374485ddbae13f42665fd3ddb01f2447b9af28ab1583707240479578"} Jan 26 16:52:34 crc kubenswrapper[4754]: I0126 16:52:34.254740 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-59b6cc659-w9w2t" event={"ID":"308d80ff-1c84-4fe7-93d4-c1f66a82fc91","Type":"ContainerStarted","Data":"17c889ab87bc42fa58318bc251afe93c9112510367a72c085614d8e7232a4297"} Jan 26 16:52:34 crc kubenswrapper[4754]: I0126 16:52:34.255195 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-59b6cc659-w9w2t" Jan 26 16:52:34 crc kubenswrapper[4754]: I0126 16:52:34.256373 4754 patch_prober.go:28] interesting pod/route-controller-manager-59b6cc659-w9w2t container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.57:8443/healthz\": dial tcp 10.217.0.57:8443: connect: connection refused" start-of-body= Jan 26 16:52:34 crc kubenswrapper[4754]: I0126 16:52:34.256408 4754 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-59b6cc659-w9w2t" podUID="308d80ff-1c84-4fe7-93d4-c1f66a82fc91" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.57:8443/healthz\": dial tcp 10.217.0.57:8443: connect: connection refused" Jan 26 16:52:34 crc kubenswrapper[4754]: I0126 16:52:34.257022 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-677j9" event={"ID":"4d81d21a-83f5-402e-91fd-d8268a6b49be","Type":"ContainerDied","Data":"df3f65a2226610c65f3906b6fb76f4060198225d530bb9b11c214dfb40d1ed9f"} Jan 26 16:52:34 crc kubenswrapper[4754]: I0126 16:52:34.257054 4754 scope.go:117] "RemoveContainer" containerID="9ba363f3729d6d97c45b7a749c56183deb835a07527b3f7da097b0eabf031c27" Jan 26 16:52:34 crc kubenswrapper[4754]: I0126 16:52:34.257202 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-677j9" Jan 26 16:52:34 crc kubenswrapper[4754]: I0126 16:52:34.262972 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m2525" event={"ID":"587cde2e-d0a4-406b-ab72-26c7fe4b6707","Type":"ContainerDied","Data":"2f4c3587fd3d9e40fdd038a747a6ef5602c25a1ee3b35c91c551084566eed866"} Jan 26 16:52:34 crc kubenswrapper[4754]: I0126 16:52:34.263156 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m2525" Jan 26 16:52:34 crc kubenswrapper[4754]: I0126 16:52:34.268648 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-59b6cc659-w9w2t" podStartSLOduration=1.268626436 podStartE2EDuration="1.268626436s" podCreationTimestamp="2026-01-26 16:52:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 16:52:34.267213415 +0000 UTC m=+320.791393849" watchObservedRunningTime="2026-01-26 16:52:34.268626436 +0000 UTC m=+320.792806870" Jan 26 16:52:34 crc kubenswrapper[4754]: I0126 16:52:34.283399 4754 scope.go:117] "RemoveContainer" containerID="10faa4395f7c506f3b98bb9824f69a2aedc5484d3299d85f666cb4dbbab259c8" Jan 26 16:52:34 crc kubenswrapper[4754]: I0126 16:52:34.288145 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-m2525"] Jan 26 16:52:34 crc kubenswrapper[4754]: I0126 16:52:34.289483 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7b233668-d1fb-40c4-bfd4-96374a7fbc81-client-ca\") pod \"controller-manager-d5469898f-fs6b7\" (UID: \"7b233668-d1fb-40c4-bfd4-96374a7fbc81\") " pod="openshift-controller-manager/controller-manager-d5469898f-fs6b7" Jan 26 16:52:34 crc kubenswrapper[4754]: I0126 16:52:34.289536 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d67wv\" (UniqueName: \"kubernetes.io/projected/7b233668-d1fb-40c4-bfd4-96374a7fbc81-kube-api-access-d67wv\") pod \"controller-manager-d5469898f-fs6b7\" (UID: \"7b233668-d1fb-40c4-bfd4-96374a7fbc81\") " pod="openshift-controller-manager/controller-manager-d5469898f-fs6b7" Jan 26 16:52:34 crc kubenswrapper[4754]: I0126 16:52:34.289563 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7b233668-d1fb-40c4-bfd4-96374a7fbc81-config\") pod \"controller-manager-d5469898f-fs6b7\" (UID: \"7b233668-d1fb-40c4-bfd4-96374a7fbc81\") " pod="openshift-controller-manager/controller-manager-d5469898f-fs6b7" Jan 26 16:52:34 crc kubenswrapper[4754]: I0126 16:52:34.289584 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7b233668-d1fb-40c4-bfd4-96374a7fbc81-proxy-ca-bundles\") pod \"controller-manager-d5469898f-fs6b7\" (UID: \"7b233668-d1fb-40c4-bfd4-96374a7fbc81\") " pod="openshift-controller-manager/controller-manager-d5469898f-fs6b7" Jan 26 16:52:34 crc kubenswrapper[4754]: I0126 16:52:34.289604 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7b233668-d1fb-40c4-bfd4-96374a7fbc81-serving-cert\") pod \"controller-manager-d5469898f-fs6b7\" (UID: \"7b233668-d1fb-40c4-bfd4-96374a7fbc81\") " pod="openshift-controller-manager/controller-manager-d5469898f-fs6b7" Jan 26 16:52:34 crc kubenswrapper[4754]: I0126 16:52:34.290502 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7b233668-d1fb-40c4-bfd4-96374a7fbc81-client-ca\") pod \"controller-manager-d5469898f-fs6b7\" (UID: \"7b233668-d1fb-40c4-bfd4-96374a7fbc81\") " pod="openshift-controller-manager/controller-manager-d5469898f-fs6b7" Jan 26 16:52:34 crc kubenswrapper[4754]: I0126 16:52:34.291084 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7b233668-d1fb-40c4-bfd4-96374a7fbc81-config\") pod \"controller-manager-d5469898f-fs6b7\" (UID: \"7b233668-d1fb-40c4-bfd4-96374a7fbc81\") " pod="openshift-controller-manager/controller-manager-d5469898f-fs6b7" Jan 26 16:52:34 crc kubenswrapper[4754]: I0126 16:52:34.291212 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7b233668-d1fb-40c4-bfd4-96374a7fbc81-proxy-ca-bundles\") pod \"controller-manager-d5469898f-fs6b7\" (UID: \"7b233668-d1fb-40c4-bfd4-96374a7fbc81\") " pod="openshift-controller-manager/controller-manager-d5469898f-fs6b7" Jan 26 16:52:34 crc kubenswrapper[4754]: I0126 16:52:34.297997 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7b233668-d1fb-40c4-bfd4-96374a7fbc81-serving-cert\") pod \"controller-manager-d5469898f-fs6b7\" (UID: \"7b233668-d1fb-40c4-bfd4-96374a7fbc81\") " pod="openshift-controller-manager/controller-manager-d5469898f-fs6b7" Jan 26 16:52:34 crc kubenswrapper[4754]: I0126 16:52:34.300477 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-m2525"] Jan 26 16:52:34 crc kubenswrapper[4754]: I0126 16:52:34.304764 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d67wv\" (UniqueName: \"kubernetes.io/projected/7b233668-d1fb-40c4-bfd4-96374a7fbc81-kube-api-access-d67wv\") pod \"controller-manager-d5469898f-fs6b7\" (UID: \"7b233668-d1fb-40c4-bfd4-96374a7fbc81\") " pod="openshift-controller-manager/controller-manager-d5469898f-fs6b7" Jan 26 16:52:34 crc kubenswrapper[4754]: I0126 16:52:34.305465 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-677j9"] Jan 26 16:52:34 crc kubenswrapper[4754]: I0126 16:52:34.310861 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-677j9"] Jan 26 16:52:34 crc kubenswrapper[4754]: I0126 16:52:34.427817 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-d5469898f-fs6b7" Jan 26 16:52:34 crc kubenswrapper[4754]: I0126 16:52:34.663165 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-d5469898f-fs6b7"] Jan 26 16:52:34 crc kubenswrapper[4754]: W0126 16:52:34.673511 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7b233668_d1fb_40c4_bfd4_96374a7fbc81.slice/crio-e6f97fc7dfef63a671ab3aa6635822beba68d78daa551df3fc7bfad79342985e WatchSource:0}: Error finding container e6f97fc7dfef63a671ab3aa6635822beba68d78daa551df3fc7bfad79342985e: Status 404 returned error can't find the container with id e6f97fc7dfef63a671ab3aa6635822beba68d78daa551df3fc7bfad79342985e Jan 26 16:52:35 crc kubenswrapper[4754]: I0126 16:52:35.272486 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-d5469898f-fs6b7" event={"ID":"7b233668-d1fb-40c4-bfd4-96374a7fbc81","Type":"ContainerStarted","Data":"002543359d3f5099129099dc09e27d79540d55011c076232de8ebe011a200767"} Jan 26 16:52:35 crc kubenswrapper[4754]: I0126 16:52:35.272719 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-d5469898f-fs6b7" event={"ID":"7b233668-d1fb-40c4-bfd4-96374a7fbc81","Type":"ContainerStarted","Data":"e6f97fc7dfef63a671ab3aa6635822beba68d78daa551df3fc7bfad79342985e"} Jan 26 16:52:35 crc kubenswrapper[4754]: I0126 16:52:35.278893 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-59b6cc659-w9w2t" Jan 26 16:52:35 crc kubenswrapper[4754]: I0126 16:52:35.334918 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-d5469898f-fs6b7" podStartSLOduration=3.334889299 podStartE2EDuration="3.334889299s" podCreationTimestamp="2026-01-26 16:52:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 16:52:35.305505355 +0000 UTC m=+321.829685789" watchObservedRunningTime="2026-01-26 16:52:35.334889299 +0000 UTC m=+321.859069733" Jan 26 16:52:35 crc kubenswrapper[4754]: I0126 16:52:35.774348 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4d81d21a-83f5-402e-91fd-d8268a6b49be" path="/var/lib/kubelet/pods/4d81d21a-83f5-402e-91fd-d8268a6b49be/volumes" Jan 26 16:52:35 crc kubenswrapper[4754]: I0126 16:52:35.774857 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="587cde2e-d0a4-406b-ab72-26c7fe4b6707" path="/var/lib/kubelet/pods/587cde2e-d0a4-406b-ab72-26c7fe4b6707/volumes" Jan 26 16:52:36 crc kubenswrapper[4754]: I0126 16:52:36.277614 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-d5469898f-fs6b7" Jan 26 16:52:36 crc kubenswrapper[4754]: I0126 16:52:36.282326 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-d5469898f-fs6b7" Jan 26 16:52:53 crc kubenswrapper[4754]: I0126 16:52:53.582549 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-d5469898f-fs6b7"] Jan 26 16:52:53 crc kubenswrapper[4754]: I0126 16:52:53.583303 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-d5469898f-fs6b7" podUID="7b233668-d1fb-40c4-bfd4-96374a7fbc81" containerName="controller-manager" containerID="cri-o://002543359d3f5099129099dc09e27d79540d55011c076232de8ebe011a200767" gracePeriod=30 Jan 26 16:52:54 crc kubenswrapper[4754]: I0126 16:52:54.428583 4754 patch_prober.go:28] interesting pod/controller-manager-d5469898f-fs6b7 container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.58:8443/healthz\": dial tcp 10.217.0.58:8443: connect: connection refused" start-of-body= Jan 26 16:52:54 crc kubenswrapper[4754]: I0126 16:52:54.429129 4754 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-d5469898f-fs6b7" podUID="7b233668-d1fb-40c4-bfd4-96374a7fbc81" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.58:8443/healthz\": dial tcp 10.217.0.58:8443: connect: connection refused" Jan 26 16:53:02 crc kubenswrapper[4754]: I0126 16:53:02.144740 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-controller-manager_controller-manager-d5469898f-fs6b7_7b233668-d1fb-40c4-bfd4-96374a7fbc81/controller-manager/0.log" Jan 26 16:53:02 crc kubenswrapper[4754]: I0126 16:53:02.145260 4754 generic.go:334] "Generic (PLEG): container finished" podID="7b233668-d1fb-40c4-bfd4-96374a7fbc81" containerID="002543359d3f5099129099dc09e27d79540d55011c076232de8ebe011a200767" exitCode=-1 Jan 26 16:53:02 crc kubenswrapper[4754]: I0126 16:53:02.145323 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-d5469898f-fs6b7" event={"ID":"7b233668-d1fb-40c4-bfd4-96374a7fbc81","Type":"ContainerDied","Data":"002543359d3f5099129099dc09e27d79540d55011c076232de8ebe011a200767"} Jan 26 16:53:04 crc kubenswrapper[4754]: I0126 16:53:04.428903 4754 patch_prober.go:28] interesting pod/controller-manager-d5469898f-fs6b7 container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.58:8443/healthz\": dial tcp 10.217.0.58:8443: connect: connection refused" start-of-body= Jan 26 16:53:04 crc kubenswrapper[4754]: I0126 16:53:04.429014 4754 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-d5469898f-fs6b7" podUID="7b233668-d1fb-40c4-bfd4-96374a7fbc81" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.58:8443/healthz\": dial tcp 10.217.0.58:8443: connect: connection refused" Jan 26 16:53:07 crc kubenswrapper[4754]: I0126 16:53:07.129080 4754 patch_prober.go:28] interesting pod/machine-config-daemon-x65wv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 26 16:53:07 crc kubenswrapper[4754]: I0126 16:53:07.129489 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 26 16:53:14 crc kubenswrapper[4754]: I0126 16:53:14.429002 4754 patch_prober.go:28] interesting pod/controller-manager-d5469898f-fs6b7 container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.58:8443/healthz\": dial tcp 10.217.0.58:8443: connect: connection refused" start-of-body= Jan 26 16:53:14 crc kubenswrapper[4754]: I0126 16:53:14.429791 4754 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-d5469898f-fs6b7" podUID="7b233668-d1fb-40c4-bfd4-96374a7fbc81" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.58:8443/healthz\": dial tcp 10.217.0.58:8443: connect: connection refused" Jan 26 16:53:16 crc kubenswrapper[4754]: I0126 16:53:16.472051 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-8x46q"] Jan 26 16:53:16 crc kubenswrapper[4754]: I0126 16:53:16.472789 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-8x46q" Jan 26 16:53:16 crc kubenswrapper[4754]: I0126 16:53:16.493793 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-8x46q"] Jan 26 16:53:16 crc kubenswrapper[4754]: I0126 16:53:16.631325 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e50abbad-382f-4b2f-bda0-4fd0e6a1c4e4-bound-sa-token\") pod \"image-registry-66df7c8f76-8x46q\" (UID: \"e50abbad-382f-4b2f-bda0-4fd0e6a1c4e4\") " pod="openshift-image-registry/image-registry-66df7c8f76-8x46q" Jan 26 16:53:16 crc kubenswrapper[4754]: I0126 16:53:16.631373 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e50abbad-382f-4b2f-bda0-4fd0e6a1c4e4-trusted-ca\") pod \"image-registry-66df7c8f76-8x46q\" (UID: \"e50abbad-382f-4b2f-bda0-4fd0e6a1c4e4\") " pod="openshift-image-registry/image-registry-66df7c8f76-8x46q" Jan 26 16:53:16 crc kubenswrapper[4754]: I0126 16:53:16.631395 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/e50abbad-382f-4b2f-bda0-4fd0e6a1c4e4-registry-certificates\") pod \"image-registry-66df7c8f76-8x46q\" (UID: \"e50abbad-382f-4b2f-bda0-4fd0e6a1c4e4\") " pod="openshift-image-registry/image-registry-66df7c8f76-8x46q" Jan 26 16:53:16 crc kubenswrapper[4754]: I0126 16:53:16.631413 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/e50abbad-382f-4b2f-bda0-4fd0e6a1c4e4-registry-tls\") pod \"image-registry-66df7c8f76-8x46q\" (UID: \"e50abbad-382f-4b2f-bda0-4fd0e6a1c4e4\") " pod="openshift-image-registry/image-registry-66df7c8f76-8x46q" Jan 26 16:53:16 crc kubenswrapper[4754]: I0126 16:53:16.631440 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/e50abbad-382f-4b2f-bda0-4fd0e6a1c4e4-installation-pull-secrets\") pod \"image-registry-66df7c8f76-8x46q\" (UID: \"e50abbad-382f-4b2f-bda0-4fd0e6a1c4e4\") " pod="openshift-image-registry/image-registry-66df7c8f76-8x46q" Jan 26 16:53:16 crc kubenswrapper[4754]: I0126 16:53:16.631531 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-8x46q\" (UID: \"e50abbad-382f-4b2f-bda0-4fd0e6a1c4e4\") " pod="openshift-image-registry/image-registry-66df7c8f76-8x46q" Jan 26 16:53:16 crc kubenswrapper[4754]: I0126 16:53:16.631574 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-smnkb\" (UniqueName: \"kubernetes.io/projected/e50abbad-382f-4b2f-bda0-4fd0e6a1c4e4-kube-api-access-smnkb\") pod \"image-registry-66df7c8f76-8x46q\" (UID: \"e50abbad-382f-4b2f-bda0-4fd0e6a1c4e4\") " pod="openshift-image-registry/image-registry-66df7c8f76-8x46q" Jan 26 16:53:16 crc kubenswrapper[4754]: I0126 16:53:16.631594 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/e50abbad-382f-4b2f-bda0-4fd0e6a1c4e4-ca-trust-extracted\") pod \"image-registry-66df7c8f76-8x46q\" (UID: \"e50abbad-382f-4b2f-bda0-4fd0e6a1c4e4\") " pod="openshift-image-registry/image-registry-66df7c8f76-8x46q" Jan 26 16:53:16 crc kubenswrapper[4754]: I0126 16:53:16.657076 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-8x46q\" (UID: \"e50abbad-382f-4b2f-bda0-4fd0e6a1c4e4\") " pod="openshift-image-registry/image-registry-66df7c8f76-8x46q" Jan 26 16:53:16 crc kubenswrapper[4754]: I0126 16:53:16.732174 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e50abbad-382f-4b2f-bda0-4fd0e6a1c4e4-bound-sa-token\") pod \"image-registry-66df7c8f76-8x46q\" (UID: \"e50abbad-382f-4b2f-bda0-4fd0e6a1c4e4\") " pod="openshift-image-registry/image-registry-66df7c8f76-8x46q" Jan 26 16:53:16 crc kubenswrapper[4754]: I0126 16:53:16.732219 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e50abbad-382f-4b2f-bda0-4fd0e6a1c4e4-trusted-ca\") pod \"image-registry-66df7c8f76-8x46q\" (UID: \"e50abbad-382f-4b2f-bda0-4fd0e6a1c4e4\") " pod="openshift-image-registry/image-registry-66df7c8f76-8x46q" Jan 26 16:53:16 crc kubenswrapper[4754]: I0126 16:53:16.732238 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/e50abbad-382f-4b2f-bda0-4fd0e6a1c4e4-registry-certificates\") pod \"image-registry-66df7c8f76-8x46q\" (UID: \"e50abbad-382f-4b2f-bda0-4fd0e6a1c4e4\") " pod="openshift-image-registry/image-registry-66df7c8f76-8x46q" Jan 26 16:53:16 crc kubenswrapper[4754]: I0126 16:53:16.732260 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/e50abbad-382f-4b2f-bda0-4fd0e6a1c4e4-registry-tls\") pod \"image-registry-66df7c8f76-8x46q\" (UID: \"e50abbad-382f-4b2f-bda0-4fd0e6a1c4e4\") " pod="openshift-image-registry/image-registry-66df7c8f76-8x46q" Jan 26 16:53:16 crc kubenswrapper[4754]: I0126 16:53:16.732276 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/e50abbad-382f-4b2f-bda0-4fd0e6a1c4e4-installation-pull-secrets\") pod \"image-registry-66df7c8f76-8x46q\" (UID: \"e50abbad-382f-4b2f-bda0-4fd0e6a1c4e4\") " pod="openshift-image-registry/image-registry-66df7c8f76-8x46q" Jan 26 16:53:16 crc kubenswrapper[4754]: I0126 16:53:16.732336 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-smnkb\" (UniqueName: \"kubernetes.io/projected/e50abbad-382f-4b2f-bda0-4fd0e6a1c4e4-kube-api-access-smnkb\") pod \"image-registry-66df7c8f76-8x46q\" (UID: \"e50abbad-382f-4b2f-bda0-4fd0e6a1c4e4\") " pod="openshift-image-registry/image-registry-66df7c8f76-8x46q" Jan 26 16:53:16 crc kubenswrapper[4754]: I0126 16:53:16.732368 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/e50abbad-382f-4b2f-bda0-4fd0e6a1c4e4-ca-trust-extracted\") pod \"image-registry-66df7c8f76-8x46q\" (UID: \"e50abbad-382f-4b2f-bda0-4fd0e6a1c4e4\") " pod="openshift-image-registry/image-registry-66df7c8f76-8x46q" Jan 26 16:53:16 crc kubenswrapper[4754]: I0126 16:53:16.732811 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/e50abbad-382f-4b2f-bda0-4fd0e6a1c4e4-ca-trust-extracted\") pod \"image-registry-66df7c8f76-8x46q\" (UID: \"e50abbad-382f-4b2f-bda0-4fd0e6a1c4e4\") " pod="openshift-image-registry/image-registry-66df7c8f76-8x46q" Jan 26 16:53:16 crc kubenswrapper[4754]: I0126 16:53:16.733892 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/e50abbad-382f-4b2f-bda0-4fd0e6a1c4e4-registry-certificates\") pod \"image-registry-66df7c8f76-8x46q\" (UID: \"e50abbad-382f-4b2f-bda0-4fd0e6a1c4e4\") " pod="openshift-image-registry/image-registry-66df7c8f76-8x46q" Jan 26 16:53:16 crc kubenswrapper[4754]: I0126 16:53:16.734232 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e50abbad-382f-4b2f-bda0-4fd0e6a1c4e4-trusted-ca\") pod \"image-registry-66df7c8f76-8x46q\" (UID: \"e50abbad-382f-4b2f-bda0-4fd0e6a1c4e4\") " pod="openshift-image-registry/image-registry-66df7c8f76-8x46q" Jan 26 16:53:16 crc kubenswrapper[4754]: I0126 16:53:16.737991 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/e50abbad-382f-4b2f-bda0-4fd0e6a1c4e4-registry-tls\") pod \"image-registry-66df7c8f76-8x46q\" (UID: \"e50abbad-382f-4b2f-bda0-4fd0e6a1c4e4\") " pod="openshift-image-registry/image-registry-66df7c8f76-8x46q" Jan 26 16:53:16 crc kubenswrapper[4754]: I0126 16:53:16.738257 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/e50abbad-382f-4b2f-bda0-4fd0e6a1c4e4-installation-pull-secrets\") pod \"image-registry-66df7c8f76-8x46q\" (UID: \"e50abbad-382f-4b2f-bda0-4fd0e6a1c4e4\") " pod="openshift-image-registry/image-registry-66df7c8f76-8x46q" Jan 26 16:53:16 crc kubenswrapper[4754]: I0126 16:53:16.754348 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e50abbad-382f-4b2f-bda0-4fd0e6a1c4e4-bound-sa-token\") pod \"image-registry-66df7c8f76-8x46q\" (UID: \"e50abbad-382f-4b2f-bda0-4fd0e6a1c4e4\") " pod="openshift-image-registry/image-registry-66df7c8f76-8x46q" Jan 26 16:53:16 crc kubenswrapper[4754]: I0126 16:53:16.756965 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-smnkb\" (UniqueName: \"kubernetes.io/projected/e50abbad-382f-4b2f-bda0-4fd0e6a1c4e4-kube-api-access-smnkb\") pod \"image-registry-66df7c8f76-8x46q\" (UID: \"e50abbad-382f-4b2f-bda0-4fd0e6a1c4e4\") " pod="openshift-image-registry/image-registry-66df7c8f76-8x46q" Jan 26 16:53:16 crc kubenswrapper[4754]: I0126 16:53:16.787406 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-8x46q" Jan 26 16:53:17 crc kubenswrapper[4754]: I0126 16:53:17.201650 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-8x46q"] Jan 26 16:53:17 crc kubenswrapper[4754]: I0126 16:53:17.245737 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-8x46q" event={"ID":"e50abbad-382f-4b2f-bda0-4fd0e6a1c4e4","Type":"ContainerStarted","Data":"520a9b0de2a34903888a3ebacd80127b9452e076ca4c2a75ef913a17b918026b"} Jan 26 16:53:19 crc kubenswrapper[4754]: I0126 16:53:19.678711 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-qzw56"] Jan 26 16:53:19 crc kubenswrapper[4754]: I0126 16:53:19.679541 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-qzw56" podUID="0ca0de6e-d004-4b46-aee4-66eae716184f" containerName="registry-server" containerID="cri-o://c87ab9b579140b0a54235a7b7812977815f23e9829ef9fadcf2f4775443d3c8f" gracePeriod=30 Jan 26 16:53:19 crc kubenswrapper[4754]: I0126 16:53:19.694738 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-cgm62"] Jan 26 16:53:19 crc kubenswrapper[4754]: I0126 16:53:19.695110 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-cgm62" podUID="c2962a85-523e-4893-8a19-b7baa1360c01" containerName="registry-server" containerID="cri-o://3533e22f9c0731ad76ebd37e68266f9c0db47db9605ad038aa9b1e1df2d19b8f" gracePeriod=30 Jan 26 16:53:19 crc kubenswrapper[4754]: I0126 16:53:19.699348 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-7xsg2"] Jan 26 16:53:19 crc kubenswrapper[4754]: I0126 16:53:19.699639 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-7xsg2" podUID="f06be1d1-4134-4ad7-9233-b4099687bfe2" containerName="marketplace-operator" containerID="cri-o://57b65fbcda96a2d7c3ad5056427ff581fc02c615544c7b716f61866017c20b02" gracePeriod=30 Jan 26 16:53:19 crc kubenswrapper[4754]: I0126 16:53:19.718126 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-hhhn5"] Jan 26 16:53:19 crc kubenswrapper[4754]: I0126 16:53:19.718548 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-hhhn5" podUID="e36dc2d0-47de-46a6-aeba-52475754867e" containerName="registry-server" containerID="cri-o://2e25b7bf0c7d3598416895619dfd7bf1e691d76eb5e80aee438983164baf5bf0" gracePeriod=30 Jan 26 16:53:19 crc kubenswrapper[4754]: I0126 16:53:19.724779 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-mp4qz"] Jan 26 16:53:19 crc kubenswrapper[4754]: I0126 16:53:19.725694 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-mp4qz" Jan 26 16:53:19 crc kubenswrapper[4754]: I0126 16:53:19.729036 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-5rvw2"] Jan 26 16:53:19 crc kubenswrapper[4754]: I0126 16:53:19.729347 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-5rvw2" podUID="b1b51a00-e6fe-428a-85ef-9d31badc8457" containerName="registry-server" containerID="cri-o://bcf0497d70be7f24baa4f82c8fe6e79184c60b7185ce37f72e6a06cb1f133a33" gracePeriod=30 Jan 26 16:53:19 crc kubenswrapper[4754]: I0126 16:53:19.754045 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-mp4qz"] Jan 26 16:53:19 crc kubenswrapper[4754]: I0126 16:53:19.895473 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/eedf4e24-8d2e-4ec6-9caa-f5af47592b89-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-mp4qz\" (UID: \"eedf4e24-8d2e-4ec6-9caa-f5af47592b89\") " pod="openshift-marketplace/marketplace-operator-79b997595-mp4qz" Jan 26 16:53:19 crc kubenswrapper[4754]: I0126 16:53:19.895526 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/eedf4e24-8d2e-4ec6-9caa-f5af47592b89-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-mp4qz\" (UID: \"eedf4e24-8d2e-4ec6-9caa-f5af47592b89\") " pod="openshift-marketplace/marketplace-operator-79b997595-mp4qz" Jan 26 16:53:19 crc kubenswrapper[4754]: I0126 16:53:19.895792 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vkg6j\" (UniqueName: \"kubernetes.io/projected/eedf4e24-8d2e-4ec6-9caa-f5af47592b89-kube-api-access-vkg6j\") pod \"marketplace-operator-79b997595-mp4qz\" (UID: \"eedf4e24-8d2e-4ec6-9caa-f5af47592b89\") " pod="openshift-marketplace/marketplace-operator-79b997595-mp4qz" Jan 26 16:53:19 crc kubenswrapper[4754]: I0126 16:53:19.997004 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/eedf4e24-8d2e-4ec6-9caa-f5af47592b89-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-mp4qz\" (UID: \"eedf4e24-8d2e-4ec6-9caa-f5af47592b89\") " pod="openshift-marketplace/marketplace-operator-79b997595-mp4qz" Jan 26 16:53:19 crc kubenswrapper[4754]: I0126 16:53:19.997105 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/eedf4e24-8d2e-4ec6-9caa-f5af47592b89-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-mp4qz\" (UID: \"eedf4e24-8d2e-4ec6-9caa-f5af47592b89\") " pod="openshift-marketplace/marketplace-operator-79b997595-mp4qz" Jan 26 16:53:19 crc kubenswrapper[4754]: I0126 16:53:19.997170 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vkg6j\" (UniqueName: \"kubernetes.io/projected/eedf4e24-8d2e-4ec6-9caa-f5af47592b89-kube-api-access-vkg6j\") pod \"marketplace-operator-79b997595-mp4qz\" (UID: \"eedf4e24-8d2e-4ec6-9caa-f5af47592b89\") " pod="openshift-marketplace/marketplace-operator-79b997595-mp4qz" Jan 26 16:53:19 crc kubenswrapper[4754]: I0126 16:53:19.999116 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/eedf4e24-8d2e-4ec6-9caa-f5af47592b89-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-mp4qz\" (UID: \"eedf4e24-8d2e-4ec6-9caa-f5af47592b89\") " pod="openshift-marketplace/marketplace-operator-79b997595-mp4qz" Jan 26 16:53:20 crc kubenswrapper[4754]: I0126 16:53:20.016413 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/eedf4e24-8d2e-4ec6-9caa-f5af47592b89-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-mp4qz\" (UID: \"eedf4e24-8d2e-4ec6-9caa-f5af47592b89\") " pod="openshift-marketplace/marketplace-operator-79b997595-mp4qz" Jan 26 16:53:20 crc kubenswrapper[4754]: I0126 16:53:20.021338 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vkg6j\" (UniqueName: \"kubernetes.io/projected/eedf4e24-8d2e-4ec6-9caa-f5af47592b89-kube-api-access-vkg6j\") pod \"marketplace-operator-79b997595-mp4qz\" (UID: \"eedf4e24-8d2e-4ec6-9caa-f5af47592b89\") " pod="openshift-marketplace/marketplace-operator-79b997595-mp4qz" Jan 26 16:53:20 crc kubenswrapper[4754]: I0126 16:53:20.047166 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-mp4qz" Jan 26 16:53:20 crc kubenswrapper[4754]: I0126 16:53:20.339692 4754 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-7xsg2 container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.36:8080/healthz\": dial tcp 10.217.0.36:8080: connect: connection refused" start-of-body= Jan 26 16:53:20 crc kubenswrapper[4754]: I0126 16:53:20.339782 4754 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-7xsg2" podUID="f06be1d1-4134-4ad7-9233-b4099687bfe2" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.36:8080/healthz\": dial tcp 10.217.0.36:8080: connect: connection refused" Jan 26 16:53:24 crc kubenswrapper[4754]: I0126 16:53:24.428656 4754 patch_prober.go:28] interesting pod/controller-manager-d5469898f-fs6b7 container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.58:8443/healthz\": dial tcp 10.217.0.58:8443: connect: connection refused" start-of-body= Jan 26 16:53:24 crc kubenswrapper[4754]: I0126 16:53:24.429177 4754 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-d5469898f-fs6b7" podUID="7b233668-d1fb-40c4-bfd4-96374a7fbc81" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.58:8443/healthz\": dial tcp 10.217.0.58:8443: connect: connection refused" Jan 26 16:53:26 crc kubenswrapper[4754]: E0126 16:53:26.392778 4754 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 3533e22f9c0731ad76ebd37e68266f9c0db47db9605ad038aa9b1e1df2d19b8f is running failed: container process not found" containerID="3533e22f9c0731ad76ebd37e68266f9c0db47db9605ad038aa9b1e1df2d19b8f" cmd=["grpc_health_probe","-addr=:50051"] Jan 26 16:53:26 crc kubenswrapper[4754]: E0126 16:53:26.393505 4754 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 3533e22f9c0731ad76ebd37e68266f9c0db47db9605ad038aa9b1e1df2d19b8f is running failed: container process not found" containerID="3533e22f9c0731ad76ebd37e68266f9c0db47db9605ad038aa9b1e1df2d19b8f" cmd=["grpc_health_probe","-addr=:50051"] Jan 26 16:53:26 crc kubenswrapper[4754]: E0126 16:53:26.393932 4754 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 3533e22f9c0731ad76ebd37e68266f9c0db47db9605ad038aa9b1e1df2d19b8f is running failed: container process not found" containerID="3533e22f9c0731ad76ebd37e68266f9c0db47db9605ad038aa9b1e1df2d19b8f" cmd=["grpc_health_probe","-addr=:50051"] Jan 26 16:53:26 crc kubenswrapper[4754]: E0126 16:53:26.394019 4754 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 3533e22f9c0731ad76ebd37e68266f9c0db47db9605ad038aa9b1e1df2d19b8f is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/community-operators-cgm62" podUID="c2962a85-523e-4893-8a19-b7baa1360c01" containerName="registry-server" Jan 26 16:53:26 crc kubenswrapper[4754]: E0126 16:53:26.884583 4754 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of c87ab9b579140b0a54235a7b7812977815f23e9829ef9fadcf2f4775443d3c8f is running failed: container process not found" containerID="c87ab9b579140b0a54235a7b7812977815f23e9829ef9fadcf2f4775443d3c8f" cmd=["grpc_health_probe","-addr=:50051"] Jan 26 16:53:26 crc kubenswrapper[4754]: E0126 16:53:26.885762 4754 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of c87ab9b579140b0a54235a7b7812977815f23e9829ef9fadcf2f4775443d3c8f is running failed: container process not found" containerID="c87ab9b579140b0a54235a7b7812977815f23e9829ef9fadcf2f4775443d3c8f" cmd=["grpc_health_probe","-addr=:50051"] Jan 26 16:53:26 crc kubenswrapper[4754]: E0126 16:53:26.886142 4754 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of c87ab9b579140b0a54235a7b7812977815f23e9829ef9fadcf2f4775443d3c8f is running failed: container process not found" containerID="c87ab9b579140b0a54235a7b7812977815f23e9829ef9fadcf2f4775443d3c8f" cmd=["grpc_health_probe","-addr=:50051"] Jan 26 16:53:26 crc kubenswrapper[4754]: E0126 16:53:26.886179 4754 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of c87ab9b579140b0a54235a7b7812977815f23e9829ef9fadcf2f4775443d3c8f is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/certified-operators-qzw56" podUID="0ca0de6e-d004-4b46-aee4-66eae716184f" containerName="registry-server" Jan 26 16:53:28 crc kubenswrapper[4754]: I0126 16:53:28.041014 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-qzw56_0ca0de6e-d004-4b46-aee4-66eae716184f/registry-server/0.log" Jan 26 16:53:28 crc kubenswrapper[4754]: I0126 16:53:28.042897 4754 generic.go:334] "Generic (PLEG): container finished" podID="0ca0de6e-d004-4b46-aee4-66eae716184f" containerID="c87ab9b579140b0a54235a7b7812977815f23e9829ef9fadcf2f4775443d3c8f" exitCode=-1 Jan 26 16:53:28 crc kubenswrapper[4754]: I0126 16:53:28.043002 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qzw56" event={"ID":"0ca0de6e-d004-4b46-aee4-66eae716184f","Type":"ContainerDied","Data":"c87ab9b579140b0a54235a7b7812977815f23e9829ef9fadcf2f4775443d3c8f"} Jan 26 16:53:28 crc kubenswrapper[4754]: E0126 16:53:28.402757 4754 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 2e25b7bf0c7d3598416895619dfd7bf1e691d76eb5e80aee438983164baf5bf0 is running failed: container process not found" containerID="2e25b7bf0c7d3598416895619dfd7bf1e691d76eb5e80aee438983164baf5bf0" cmd=["grpc_health_probe","-addr=:50051"] Jan 26 16:53:28 crc kubenswrapper[4754]: E0126 16:53:28.403402 4754 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 2e25b7bf0c7d3598416895619dfd7bf1e691d76eb5e80aee438983164baf5bf0 is running failed: container process not found" containerID="2e25b7bf0c7d3598416895619dfd7bf1e691d76eb5e80aee438983164baf5bf0" cmd=["grpc_health_probe","-addr=:50051"] Jan 26 16:53:28 crc kubenswrapper[4754]: E0126 16:53:28.403789 4754 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 2e25b7bf0c7d3598416895619dfd7bf1e691d76eb5e80aee438983164baf5bf0 is running failed: container process not found" containerID="2e25b7bf0c7d3598416895619dfd7bf1e691d76eb5e80aee438983164baf5bf0" cmd=["grpc_health_probe","-addr=:50051"] Jan 26 16:53:28 crc kubenswrapper[4754]: E0126 16:53:28.403820 4754 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 2e25b7bf0c7d3598416895619dfd7bf1e691d76eb5e80aee438983164baf5bf0 is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/redhat-marketplace-hhhn5" podUID="e36dc2d0-47de-46a6-aeba-52475754867e" containerName="registry-server" Jan 26 16:53:29 crc kubenswrapper[4754]: E0126 16:53:29.417423 4754 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of bcf0497d70be7f24baa4f82c8fe6e79184c60b7185ce37f72e6a06cb1f133a33 is running failed: container process not found" containerID="bcf0497d70be7f24baa4f82c8fe6e79184c60b7185ce37f72e6a06cb1f133a33" cmd=["grpc_health_probe","-addr=:50051"] Jan 26 16:53:29 crc kubenswrapper[4754]: E0126 16:53:29.418150 4754 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of bcf0497d70be7f24baa4f82c8fe6e79184c60b7185ce37f72e6a06cb1f133a33 is running failed: container process not found" containerID="bcf0497d70be7f24baa4f82c8fe6e79184c60b7185ce37f72e6a06cb1f133a33" cmd=["grpc_health_probe","-addr=:50051"] Jan 26 16:53:29 crc kubenswrapper[4754]: E0126 16:53:29.418640 4754 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of bcf0497d70be7f24baa4f82c8fe6e79184c60b7185ce37f72e6a06cb1f133a33 is running failed: container process not found" containerID="bcf0497d70be7f24baa4f82c8fe6e79184c60b7185ce37f72e6a06cb1f133a33" cmd=["grpc_health_probe","-addr=:50051"] Jan 26 16:53:29 crc kubenswrapper[4754]: E0126 16:53:29.418769 4754 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of bcf0497d70be7f24baa4f82c8fe6e79184c60b7185ce37f72e6a06cb1f133a33 is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/redhat-operators-5rvw2" podUID="b1b51a00-e6fe-428a-85ef-9d31badc8457" containerName="registry-server" Jan 26 16:53:30 crc kubenswrapper[4754]: I0126 16:53:30.340220 4754 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-7xsg2 container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.36:8080/healthz\": dial tcp 10.217.0.36:8080: connect: connection refused" start-of-body= Jan 26 16:53:30 crc kubenswrapper[4754]: I0126 16:53:30.340733 4754 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-7xsg2" podUID="f06be1d1-4134-4ad7-9233-b4099687bfe2" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.36:8080/healthz\": dial tcp 10.217.0.36:8080: connect: connection refused" Jan 26 16:53:31 crc kubenswrapper[4754]: I0126 16:53:31.784121 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-hhhn5_e36dc2d0-47de-46a6-aeba-52475754867e/registry-server/0.log" Jan 26 16:53:31 crc kubenswrapper[4754]: I0126 16:53:31.785177 4754 generic.go:334] "Generic (PLEG): container finished" podID="e36dc2d0-47de-46a6-aeba-52475754867e" containerID="2e25b7bf0c7d3598416895619dfd7bf1e691d76eb5e80aee438983164baf5bf0" exitCode=-1 Jan 26 16:53:31 crc kubenswrapper[4754]: I0126 16:53:31.785268 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hhhn5" event={"ID":"e36dc2d0-47de-46a6-aeba-52475754867e","Type":"ContainerDied","Data":"2e25b7bf0c7d3598416895619dfd7bf1e691d76eb5e80aee438983164baf5bf0"} Jan 26 16:53:34 crc kubenswrapper[4754]: I0126 16:53:34.429526 4754 patch_prober.go:28] interesting pod/controller-manager-d5469898f-fs6b7 container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.58:8443/healthz\": dial tcp 10.217.0.58:8443: connect: connection refused" start-of-body= Jan 26 16:53:34 crc kubenswrapper[4754]: I0126 16:53:34.429620 4754 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-d5469898f-fs6b7" podUID="7b233668-d1fb-40c4-bfd4-96374a7fbc81" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.58:8443/healthz\": dial tcp 10.217.0.58:8443: connect: connection refused" Jan 26 16:53:35 crc kubenswrapper[4754]: I0126 16:53:35.522428 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-7xsg2_f06be1d1-4134-4ad7-9233-b4099687bfe2/marketplace-operator/1.log" Jan 26 16:53:35 crc kubenswrapper[4754]: I0126 16:53:35.523433 4754 generic.go:334] "Generic (PLEG): container finished" podID="f06be1d1-4134-4ad7-9233-b4099687bfe2" containerID="57b65fbcda96a2d7c3ad5056427ff581fc02c615544c7b716f61866017c20b02" exitCode=-1 Jan 26 16:53:35 crc kubenswrapper[4754]: I0126 16:53:35.523504 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-7xsg2" event={"ID":"f06be1d1-4134-4ad7-9233-b4099687bfe2","Type":"ContainerDied","Data":"57b65fbcda96a2d7c3ad5056427ff581fc02c615544c7b716f61866017c20b02"} Jan 26 16:53:35 crc kubenswrapper[4754]: I0126 16:53:35.523653 4754 scope.go:117] "RemoveContainer" containerID="089c7bd3b8cd79c825b9c58787fb66cd12628763659360ce5e1b716c6f8fa7d0" Jan 26 16:53:36 crc kubenswrapper[4754]: E0126 16:53:36.393361 4754 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 3533e22f9c0731ad76ebd37e68266f9c0db47db9605ad038aa9b1e1df2d19b8f is running failed: container process not found" containerID="3533e22f9c0731ad76ebd37e68266f9c0db47db9605ad038aa9b1e1df2d19b8f" cmd=["grpc_health_probe","-addr=:50051"] Jan 26 16:53:36 crc kubenswrapper[4754]: E0126 16:53:36.393982 4754 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 3533e22f9c0731ad76ebd37e68266f9c0db47db9605ad038aa9b1e1df2d19b8f is running failed: container process not found" containerID="3533e22f9c0731ad76ebd37e68266f9c0db47db9605ad038aa9b1e1df2d19b8f" cmd=["grpc_health_probe","-addr=:50051"] Jan 26 16:53:36 crc kubenswrapper[4754]: E0126 16:53:36.394592 4754 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 3533e22f9c0731ad76ebd37e68266f9c0db47db9605ad038aa9b1e1df2d19b8f is running failed: container process not found" containerID="3533e22f9c0731ad76ebd37e68266f9c0db47db9605ad038aa9b1e1df2d19b8f" cmd=["grpc_health_probe","-addr=:50051"] Jan 26 16:53:36 crc kubenswrapper[4754]: E0126 16:53:36.394635 4754 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 3533e22f9c0731ad76ebd37e68266f9c0db47db9605ad038aa9b1e1df2d19b8f is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/community-operators-cgm62" podUID="c2962a85-523e-4893-8a19-b7baa1360c01" containerName="registry-server" Jan 26 16:53:36 crc kubenswrapper[4754]: E0126 16:53:36.884296 4754 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of c87ab9b579140b0a54235a7b7812977815f23e9829ef9fadcf2f4775443d3c8f is running failed: container process not found" containerID="c87ab9b579140b0a54235a7b7812977815f23e9829ef9fadcf2f4775443d3c8f" cmd=["grpc_health_probe","-addr=:50051"] Jan 26 16:53:36 crc kubenswrapper[4754]: E0126 16:53:36.885803 4754 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of c87ab9b579140b0a54235a7b7812977815f23e9829ef9fadcf2f4775443d3c8f is running failed: container process not found" containerID="c87ab9b579140b0a54235a7b7812977815f23e9829ef9fadcf2f4775443d3c8f" cmd=["grpc_health_probe","-addr=:50051"] Jan 26 16:53:36 crc kubenswrapper[4754]: E0126 16:53:36.886345 4754 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of c87ab9b579140b0a54235a7b7812977815f23e9829ef9fadcf2f4775443d3c8f is running failed: container process not found" containerID="c87ab9b579140b0a54235a7b7812977815f23e9829ef9fadcf2f4775443d3c8f" cmd=["grpc_health_probe","-addr=:50051"] Jan 26 16:53:36 crc kubenswrapper[4754]: E0126 16:53:36.886439 4754 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of c87ab9b579140b0a54235a7b7812977815f23e9829ef9fadcf2f4775443d3c8f is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/certified-operators-qzw56" podUID="0ca0de6e-d004-4b46-aee4-66eae716184f" containerName="registry-server" Jan 26 16:53:37 crc kubenswrapper[4754]: I0126 16:53:37.129157 4754 patch_prober.go:28] interesting pod/machine-config-daemon-x65wv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 26 16:53:37 crc kubenswrapper[4754]: I0126 16:53:37.129469 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 26 16:53:38 crc kubenswrapper[4754]: E0126 16:53:38.402927 4754 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 2e25b7bf0c7d3598416895619dfd7bf1e691d76eb5e80aee438983164baf5bf0 is running failed: container process not found" containerID="2e25b7bf0c7d3598416895619dfd7bf1e691d76eb5e80aee438983164baf5bf0" cmd=["grpc_health_probe","-addr=:50051"] Jan 26 16:53:38 crc kubenswrapper[4754]: E0126 16:53:38.403657 4754 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 2e25b7bf0c7d3598416895619dfd7bf1e691d76eb5e80aee438983164baf5bf0 is running failed: container process not found" containerID="2e25b7bf0c7d3598416895619dfd7bf1e691d76eb5e80aee438983164baf5bf0" cmd=["grpc_health_probe","-addr=:50051"] Jan 26 16:53:38 crc kubenswrapper[4754]: E0126 16:53:38.404165 4754 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 2e25b7bf0c7d3598416895619dfd7bf1e691d76eb5e80aee438983164baf5bf0 is running failed: container process not found" containerID="2e25b7bf0c7d3598416895619dfd7bf1e691d76eb5e80aee438983164baf5bf0" cmd=["grpc_health_probe","-addr=:50051"] Jan 26 16:53:38 crc kubenswrapper[4754]: E0126 16:53:38.404198 4754 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 2e25b7bf0c7d3598416895619dfd7bf1e691d76eb5e80aee438983164baf5bf0 is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/redhat-marketplace-hhhn5" podUID="e36dc2d0-47de-46a6-aeba-52475754867e" containerName="registry-server" Jan 26 16:53:39 crc kubenswrapper[4754]: I0126 16:53:39.257253 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-cgm62_c2962a85-523e-4893-8a19-b7baa1360c01/registry-server/0.log" Jan 26 16:53:39 crc kubenswrapper[4754]: I0126 16:53:39.258200 4754 generic.go:334] "Generic (PLEG): container finished" podID="c2962a85-523e-4893-8a19-b7baa1360c01" containerID="3533e22f9c0731ad76ebd37e68266f9c0db47db9605ad038aa9b1e1df2d19b8f" exitCode=-1 Jan 26 16:53:39 crc kubenswrapper[4754]: I0126 16:53:39.258259 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cgm62" event={"ID":"c2962a85-523e-4893-8a19-b7baa1360c01","Type":"ContainerDied","Data":"3533e22f9c0731ad76ebd37e68266f9c0db47db9605ad038aa9b1e1df2d19b8f"} Jan 26 16:53:39 crc kubenswrapper[4754]: E0126 16:53:39.416277 4754 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of bcf0497d70be7f24baa4f82c8fe6e79184c60b7185ce37f72e6a06cb1f133a33 is running failed: container process not found" containerID="bcf0497d70be7f24baa4f82c8fe6e79184c60b7185ce37f72e6a06cb1f133a33" cmd=["grpc_health_probe","-addr=:50051"] Jan 26 16:53:39 crc kubenswrapper[4754]: E0126 16:53:39.417528 4754 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of bcf0497d70be7f24baa4f82c8fe6e79184c60b7185ce37f72e6a06cb1f133a33 is running failed: container process not found" containerID="bcf0497d70be7f24baa4f82c8fe6e79184c60b7185ce37f72e6a06cb1f133a33" cmd=["grpc_health_probe","-addr=:50051"] Jan 26 16:53:39 crc kubenswrapper[4754]: E0126 16:53:39.418114 4754 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of bcf0497d70be7f24baa4f82c8fe6e79184c60b7185ce37f72e6a06cb1f133a33 is running failed: container process not found" containerID="bcf0497d70be7f24baa4f82c8fe6e79184c60b7185ce37f72e6a06cb1f133a33" cmd=["grpc_health_probe","-addr=:50051"] Jan 26 16:53:39 crc kubenswrapper[4754]: E0126 16:53:39.418281 4754 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of bcf0497d70be7f24baa4f82c8fe6e79184c60b7185ce37f72e6a06cb1f133a33 is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/redhat-operators-5rvw2" podUID="b1b51a00-e6fe-428a-85ef-9d31badc8457" containerName="registry-server" Jan 26 16:53:40 crc kubenswrapper[4754]: I0126 16:53:40.340148 4754 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-7xsg2 container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.36:8080/healthz\": dial tcp 10.217.0.36:8080: connect: connection refused" start-of-body= Jan 26 16:53:40 crc kubenswrapper[4754]: I0126 16:53:40.340210 4754 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-7xsg2" podUID="f06be1d1-4134-4ad7-9233-b4099687bfe2" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.36:8080/healthz\": dial tcp 10.217.0.36:8080: connect: connection refused" Jan 26 16:53:40 crc kubenswrapper[4754]: I0126 16:53:40.340281 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-7xsg2" Jan 26 16:53:41 crc kubenswrapper[4754]: I0126 16:53:41.033838 4754 patch_prober.go:28] interesting pod/console-operator-58897d9998-wjdgm container/console-operator namespace/openshift-console-operator: Liveness probe status=failure output="Get \"https://10.217.0.32:8443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Jan 26 16:53:41 crc kubenswrapper[4754]: I0126 16:53:41.033903 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console-operator/console-operator-58897d9998-wjdgm" podUID="07245d2d-bd41-4bca-9f4a-fab5f4a48eb9" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.32:8443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Jan 26 16:53:42 crc kubenswrapper[4754]: I0126 16:53:42.991364 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-5rvw2_b1b51a00-e6fe-428a-85ef-9d31badc8457/registry-server/0.log" Jan 26 16:53:42 crc kubenswrapper[4754]: I0126 16:53:42.993472 4754 generic.go:334] "Generic (PLEG): container finished" podID="b1b51a00-e6fe-428a-85ef-9d31badc8457" containerID="bcf0497d70be7f24baa4f82c8fe6e79184c60b7185ce37f72e6a06cb1f133a33" exitCode=-1 Jan 26 16:53:42 crc kubenswrapper[4754]: I0126 16:53:42.993534 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5rvw2" event={"ID":"b1b51a00-e6fe-428a-85ef-9d31badc8457","Type":"ContainerDied","Data":"bcf0497d70be7f24baa4f82c8fe6e79184c60b7185ce37f72e6a06cb1f133a33"} Jan 26 16:53:43 crc kubenswrapper[4754]: I0126 16:53:43.619981 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-mp4qz"] Jan 26 16:53:44 crc kubenswrapper[4754]: I0126 16:53:44.021158 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-mp4qz" event={"ID":"eedf4e24-8d2e-4ec6-9caa-f5af47592b89","Type":"ContainerStarted","Data":"cf42e17db284a74ed90034fbb608adc345ff21563421bede75a6d681e14940c8"} Jan 26 16:53:44 crc kubenswrapper[4754]: I0126 16:53:44.021463 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-mp4qz" event={"ID":"eedf4e24-8d2e-4ec6-9caa-f5af47592b89","Type":"ContainerStarted","Data":"13a73f23cfa9198eccecf1462d298c559c64fa01bee4faa8d9a96c037de75d3e"} Jan 26 16:53:44 crc kubenswrapper[4754]: I0126 16:53:44.022875 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-mp4qz" Jan 26 16:53:44 crc kubenswrapper[4754]: I0126 16:53:44.031709 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-d5469898f-fs6b7" event={"ID":"7b233668-d1fb-40c4-bfd4-96374a7fbc81","Type":"ContainerDied","Data":"e6f97fc7dfef63a671ab3aa6635822beba68d78daa551df3fc7bfad79342985e"} Jan 26 16:53:44 crc kubenswrapper[4754]: I0126 16:53:44.031751 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e6f97fc7dfef63a671ab3aa6635822beba68d78daa551df3fc7bfad79342985e" Jan 26 16:53:44 crc kubenswrapper[4754]: I0126 16:53:44.033916 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-8x46q" event={"ID":"e50abbad-382f-4b2f-bda0-4fd0e6a1c4e4","Type":"ContainerStarted","Data":"d1c5c5e6752bfb01d43694e953da2a1b8c6dca3fb2ce8187a0ccb600b2ba4429"} Jan 26 16:53:44 crc kubenswrapper[4754]: I0126 16:53:44.034876 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-8x46q" Jan 26 16:53:44 crc kubenswrapper[4754]: I0126 16:53:44.038116 4754 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-mp4qz container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.61:8080/healthz\": dial tcp 10.217.0.61:8080: connect: connection refused" start-of-body= Jan 26 16:53:44 crc kubenswrapper[4754]: I0126 16:53:44.038193 4754 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-mp4qz" podUID="eedf4e24-8d2e-4ec6-9caa-f5af47592b89" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.61:8080/healthz\": dial tcp 10.217.0.61:8080: connect: connection refused" Jan 26 16:53:44 crc kubenswrapper[4754]: I0126 16:53:44.046790 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-mp4qz" podStartSLOduration=25.046771009 podStartE2EDuration="25.046771009s" podCreationTimestamp="2026-01-26 16:53:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 16:53:44.040915535 +0000 UTC m=+390.565095979" watchObservedRunningTime="2026-01-26 16:53:44.046771009 +0000 UTC m=+390.570951443" Jan 26 16:53:44 crc kubenswrapper[4754]: I0126 16:53:44.075306 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-8x46q" podStartSLOduration=28.07528021 podStartE2EDuration="28.07528021s" podCreationTimestamp="2026-01-26 16:53:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 16:53:44.073099019 +0000 UTC m=+390.597279503" watchObservedRunningTime="2026-01-26 16:53:44.07528021 +0000 UTC m=+390.599460644" Jan 26 16:53:44 crc kubenswrapper[4754]: I0126 16:53:44.085308 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-d5469898f-fs6b7" Jan 26 16:53:44 crc kubenswrapper[4754]: I0126 16:53:44.098928 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-7xsg2" Jan 26 16:53:44 crc kubenswrapper[4754]: I0126 16:53:44.125274 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qzw56" Jan 26 16:53:44 crc kubenswrapper[4754]: I0126 16:53:44.128895 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-7748bf895c-v5p45"] Jan 26 16:53:44 crc kubenswrapper[4754]: E0126 16:53:44.129117 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f06be1d1-4134-4ad7-9233-b4099687bfe2" containerName="marketplace-operator" Jan 26 16:53:44 crc kubenswrapper[4754]: I0126 16:53:44.129133 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="f06be1d1-4134-4ad7-9233-b4099687bfe2" containerName="marketplace-operator" Jan 26 16:53:44 crc kubenswrapper[4754]: E0126 16:53:44.129150 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ca0de6e-d004-4b46-aee4-66eae716184f" containerName="extract-utilities" Jan 26 16:53:44 crc kubenswrapper[4754]: I0126 16:53:44.129157 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ca0de6e-d004-4b46-aee4-66eae716184f" containerName="extract-utilities" Jan 26 16:53:44 crc kubenswrapper[4754]: E0126 16:53:44.129175 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ca0de6e-d004-4b46-aee4-66eae716184f" containerName="extract-content" Jan 26 16:53:44 crc kubenswrapper[4754]: I0126 16:53:44.129183 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ca0de6e-d004-4b46-aee4-66eae716184f" containerName="extract-content" Jan 26 16:53:44 crc kubenswrapper[4754]: E0126 16:53:44.129194 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b233668-d1fb-40c4-bfd4-96374a7fbc81" containerName="controller-manager" Jan 26 16:53:44 crc kubenswrapper[4754]: I0126 16:53:44.129202 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b233668-d1fb-40c4-bfd4-96374a7fbc81" containerName="controller-manager" Jan 26 16:53:44 crc kubenswrapper[4754]: E0126 16:53:44.129209 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ca0de6e-d004-4b46-aee4-66eae716184f" containerName="registry-server" Jan 26 16:53:44 crc kubenswrapper[4754]: I0126 16:53:44.129215 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ca0de6e-d004-4b46-aee4-66eae716184f" containerName="registry-server" Jan 26 16:53:44 crc kubenswrapper[4754]: I0126 16:53:44.129336 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="f06be1d1-4134-4ad7-9233-b4099687bfe2" containerName="marketplace-operator" Jan 26 16:53:44 crc kubenswrapper[4754]: I0126 16:53:44.129349 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="0ca0de6e-d004-4b46-aee4-66eae716184f" containerName="registry-server" Jan 26 16:53:44 crc kubenswrapper[4754]: I0126 16:53:44.129359 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="f06be1d1-4134-4ad7-9233-b4099687bfe2" containerName="marketplace-operator" Jan 26 16:53:44 crc kubenswrapper[4754]: I0126 16:53:44.129370 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="7b233668-d1fb-40c4-bfd4-96374a7fbc81" containerName="controller-manager" Jan 26 16:53:44 crc kubenswrapper[4754]: I0126 16:53:44.129804 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7748bf895c-v5p45" Jan 26 16:53:44 crc kubenswrapper[4754]: I0126 16:53:44.132653 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-7748bf895c-v5p45"] Jan 26 16:53:44 crc kubenswrapper[4754]: I0126 16:53:44.139892 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5rvw2" Jan 26 16:53:44 crc kubenswrapper[4754]: I0126 16:53:44.145803 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cgm62" Jan 26 16:53:44 crc kubenswrapper[4754]: I0126 16:53:44.153330 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hhhn5" Jan 26 16:53:44 crc kubenswrapper[4754]: I0126 16:53:44.265353 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0ca0de6e-d004-4b46-aee4-66eae716184f-utilities\") pod \"0ca0de6e-d004-4b46-aee4-66eae716184f\" (UID: \"0ca0de6e-d004-4b46-aee4-66eae716184f\") " Jan 26 16:53:44 crc kubenswrapper[4754]: I0126 16:53:44.265401 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e36dc2d0-47de-46a6-aeba-52475754867e-catalog-content\") pod \"e36dc2d0-47de-46a6-aeba-52475754867e\" (UID: \"e36dc2d0-47de-46a6-aeba-52475754867e\") " Jan 26 16:53:44 crc kubenswrapper[4754]: I0126 16:53:44.265425 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0ca0de6e-d004-4b46-aee4-66eae716184f-catalog-content\") pod \"0ca0de6e-d004-4b46-aee4-66eae716184f\" (UID: \"0ca0de6e-d004-4b46-aee4-66eae716184f\") " Jan 26 16:53:44 crc kubenswrapper[4754]: I0126 16:53:44.265456 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zszpm\" (UniqueName: \"kubernetes.io/projected/f06be1d1-4134-4ad7-9233-b4099687bfe2-kube-api-access-zszpm\") pod \"f06be1d1-4134-4ad7-9233-b4099687bfe2\" (UID: \"f06be1d1-4134-4ad7-9233-b4099687bfe2\") " Jan 26 16:53:44 crc kubenswrapper[4754]: I0126 16:53:44.265477 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f06be1d1-4134-4ad7-9233-b4099687bfe2-marketplace-trusted-ca\") pod \"f06be1d1-4134-4ad7-9233-b4099687bfe2\" (UID: \"f06be1d1-4134-4ad7-9233-b4099687bfe2\") " Jan 26 16:53:44 crc kubenswrapper[4754]: I0126 16:53:44.265501 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z7pgd\" (UniqueName: \"kubernetes.io/projected/c2962a85-523e-4893-8a19-b7baa1360c01-kube-api-access-z7pgd\") pod \"c2962a85-523e-4893-8a19-b7baa1360c01\" (UID: \"c2962a85-523e-4893-8a19-b7baa1360c01\") " Jan 26 16:53:44 crc kubenswrapper[4754]: I0126 16:53:44.265528 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d55d7\" (UniqueName: \"kubernetes.io/projected/e36dc2d0-47de-46a6-aeba-52475754867e-kube-api-access-d55d7\") pod \"e36dc2d0-47de-46a6-aeba-52475754867e\" (UID: \"e36dc2d0-47de-46a6-aeba-52475754867e\") " Jan 26 16:53:44 crc kubenswrapper[4754]: I0126 16:53:44.265544 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b1b51a00-e6fe-428a-85ef-9d31badc8457-catalog-content\") pod \"b1b51a00-e6fe-428a-85ef-9d31badc8457\" (UID: \"b1b51a00-e6fe-428a-85ef-9d31badc8457\") " Jan 26 16:53:44 crc kubenswrapper[4754]: I0126 16:53:44.265561 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xwf62\" (UniqueName: \"kubernetes.io/projected/0ca0de6e-d004-4b46-aee4-66eae716184f-kube-api-access-xwf62\") pod \"0ca0de6e-d004-4b46-aee4-66eae716184f\" (UID: \"0ca0de6e-d004-4b46-aee4-66eae716184f\") " Jan 26 16:53:44 crc kubenswrapper[4754]: I0126 16:53:44.265576 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7b233668-d1fb-40c4-bfd4-96374a7fbc81-client-ca\") pod \"7b233668-d1fb-40c4-bfd4-96374a7fbc81\" (UID: \"7b233668-d1fb-40c4-bfd4-96374a7fbc81\") " Jan 26 16:53:44 crc kubenswrapper[4754]: I0126 16:53:44.265600 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c2962a85-523e-4893-8a19-b7baa1360c01-utilities\") pod \"c2962a85-523e-4893-8a19-b7baa1360c01\" (UID: \"c2962a85-523e-4893-8a19-b7baa1360c01\") " Jan 26 16:53:44 crc kubenswrapper[4754]: I0126 16:53:44.265623 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c2962a85-523e-4893-8a19-b7baa1360c01-catalog-content\") pod \"c2962a85-523e-4893-8a19-b7baa1360c01\" (UID: \"c2962a85-523e-4893-8a19-b7baa1360c01\") " Jan 26 16:53:44 crc kubenswrapper[4754]: I0126 16:53:44.265685 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7b233668-d1fb-40c4-bfd4-96374a7fbc81-proxy-ca-bundles\") pod \"7b233668-d1fb-40c4-bfd4-96374a7fbc81\" (UID: \"7b233668-d1fb-40c4-bfd4-96374a7fbc81\") " Jan 26 16:53:44 crc kubenswrapper[4754]: I0126 16:53:44.265708 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8j5j7\" (UniqueName: \"kubernetes.io/projected/b1b51a00-e6fe-428a-85ef-9d31badc8457-kube-api-access-8j5j7\") pod \"b1b51a00-e6fe-428a-85ef-9d31badc8457\" (UID: \"b1b51a00-e6fe-428a-85ef-9d31badc8457\") " Jan 26 16:53:44 crc kubenswrapper[4754]: I0126 16:53:44.265731 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e36dc2d0-47de-46a6-aeba-52475754867e-utilities\") pod \"e36dc2d0-47de-46a6-aeba-52475754867e\" (UID: \"e36dc2d0-47de-46a6-aeba-52475754867e\") " Jan 26 16:53:44 crc kubenswrapper[4754]: I0126 16:53:44.265759 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/f06be1d1-4134-4ad7-9233-b4099687bfe2-marketplace-operator-metrics\") pod \"f06be1d1-4134-4ad7-9233-b4099687bfe2\" (UID: \"f06be1d1-4134-4ad7-9233-b4099687bfe2\") " Jan 26 16:53:44 crc kubenswrapper[4754]: I0126 16:53:44.265778 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7b233668-d1fb-40c4-bfd4-96374a7fbc81-serving-cert\") pod \"7b233668-d1fb-40c4-bfd4-96374a7fbc81\" (UID: \"7b233668-d1fb-40c4-bfd4-96374a7fbc81\") " Jan 26 16:53:44 crc kubenswrapper[4754]: I0126 16:53:44.265799 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b1b51a00-e6fe-428a-85ef-9d31badc8457-utilities\") pod \"b1b51a00-e6fe-428a-85ef-9d31badc8457\" (UID: \"b1b51a00-e6fe-428a-85ef-9d31badc8457\") " Jan 26 16:53:44 crc kubenswrapper[4754]: I0126 16:53:44.265820 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d67wv\" (UniqueName: \"kubernetes.io/projected/7b233668-d1fb-40c4-bfd4-96374a7fbc81-kube-api-access-d67wv\") pod \"7b233668-d1fb-40c4-bfd4-96374a7fbc81\" (UID: \"7b233668-d1fb-40c4-bfd4-96374a7fbc81\") " Jan 26 16:53:44 crc kubenswrapper[4754]: I0126 16:53:44.265840 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7b233668-d1fb-40c4-bfd4-96374a7fbc81-config\") pod \"7b233668-d1fb-40c4-bfd4-96374a7fbc81\" (UID: \"7b233668-d1fb-40c4-bfd4-96374a7fbc81\") " Jan 26 16:53:44 crc kubenswrapper[4754]: I0126 16:53:44.265997 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/409f1ff8-a849-4e8e-a7e9-0bb57ed94cb4-config\") pod \"controller-manager-7748bf895c-v5p45\" (UID: \"409f1ff8-a849-4e8e-a7e9-0bb57ed94cb4\") " pod="openshift-controller-manager/controller-manager-7748bf895c-v5p45" Jan 26 16:53:44 crc kubenswrapper[4754]: I0126 16:53:44.266035 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/409f1ff8-a849-4e8e-a7e9-0bb57ed94cb4-proxy-ca-bundles\") pod \"controller-manager-7748bf895c-v5p45\" (UID: \"409f1ff8-a849-4e8e-a7e9-0bb57ed94cb4\") " pod="openshift-controller-manager/controller-manager-7748bf895c-v5p45" Jan 26 16:53:44 crc kubenswrapper[4754]: I0126 16:53:44.266066 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w7cxf\" (UniqueName: \"kubernetes.io/projected/409f1ff8-a849-4e8e-a7e9-0bb57ed94cb4-kube-api-access-w7cxf\") pod \"controller-manager-7748bf895c-v5p45\" (UID: \"409f1ff8-a849-4e8e-a7e9-0bb57ed94cb4\") " pod="openshift-controller-manager/controller-manager-7748bf895c-v5p45" Jan 26 16:53:44 crc kubenswrapper[4754]: I0126 16:53:44.266089 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/409f1ff8-a849-4e8e-a7e9-0bb57ed94cb4-serving-cert\") pod \"controller-manager-7748bf895c-v5p45\" (UID: \"409f1ff8-a849-4e8e-a7e9-0bb57ed94cb4\") " pod="openshift-controller-manager/controller-manager-7748bf895c-v5p45" Jan 26 16:53:44 crc kubenswrapper[4754]: I0126 16:53:44.266143 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/409f1ff8-a849-4e8e-a7e9-0bb57ed94cb4-client-ca\") pod \"controller-manager-7748bf895c-v5p45\" (UID: \"409f1ff8-a849-4e8e-a7e9-0bb57ed94cb4\") " pod="openshift-controller-manager/controller-manager-7748bf895c-v5p45" Jan 26 16:53:44 crc kubenswrapper[4754]: I0126 16:53:44.268951 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7b233668-d1fb-40c4-bfd4-96374a7fbc81-config" (OuterVolumeSpecName: "config") pod "7b233668-d1fb-40c4-bfd4-96374a7fbc81" (UID: "7b233668-d1fb-40c4-bfd4-96374a7fbc81"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 16:53:44 crc kubenswrapper[4754]: I0126 16:53:44.269721 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c2962a85-523e-4893-8a19-b7baa1360c01-utilities" (OuterVolumeSpecName: "utilities") pod "c2962a85-523e-4893-8a19-b7baa1360c01" (UID: "c2962a85-523e-4893-8a19-b7baa1360c01"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 16:53:44 crc kubenswrapper[4754]: I0126 16:53:44.270377 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7b233668-d1fb-40c4-bfd4-96374a7fbc81-client-ca" (OuterVolumeSpecName: "client-ca") pod "7b233668-d1fb-40c4-bfd4-96374a7fbc81" (UID: "7b233668-d1fb-40c4-bfd4-96374a7fbc81"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 16:53:44 crc kubenswrapper[4754]: I0126 16:53:44.271352 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f06be1d1-4134-4ad7-9233-b4099687bfe2-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "f06be1d1-4134-4ad7-9233-b4099687bfe2" (UID: "f06be1d1-4134-4ad7-9233-b4099687bfe2"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 16:53:44 crc kubenswrapper[4754]: I0126 16:53:44.273517 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b1b51a00-e6fe-428a-85ef-9d31badc8457-utilities" (OuterVolumeSpecName: "utilities") pod "b1b51a00-e6fe-428a-85ef-9d31badc8457" (UID: "b1b51a00-e6fe-428a-85ef-9d31badc8457"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 16:53:44 crc kubenswrapper[4754]: I0126 16:53:44.273578 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e36dc2d0-47de-46a6-aeba-52475754867e-utilities" (OuterVolumeSpecName: "utilities") pod "e36dc2d0-47de-46a6-aeba-52475754867e" (UID: "e36dc2d0-47de-46a6-aeba-52475754867e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 16:53:44 crc kubenswrapper[4754]: I0126 16:53:44.273853 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0ca0de6e-d004-4b46-aee4-66eae716184f-utilities" (OuterVolumeSpecName: "utilities") pod "0ca0de6e-d004-4b46-aee4-66eae716184f" (UID: "0ca0de6e-d004-4b46-aee4-66eae716184f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 16:53:44 crc kubenswrapper[4754]: I0126 16:53:44.274318 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7b233668-d1fb-40c4-bfd4-96374a7fbc81-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7b233668-d1fb-40c4-bfd4-96374a7fbc81" (UID: "7b233668-d1fb-40c4-bfd4-96374a7fbc81"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 16:53:44 crc kubenswrapper[4754]: I0126 16:53:44.284581 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e36dc2d0-47de-46a6-aeba-52475754867e-kube-api-access-d55d7" (OuterVolumeSpecName: "kube-api-access-d55d7") pod "e36dc2d0-47de-46a6-aeba-52475754867e" (UID: "e36dc2d0-47de-46a6-aeba-52475754867e"). InnerVolumeSpecName "kube-api-access-d55d7". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:53:44 crc kubenswrapper[4754]: I0126 16:53:44.284793 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f06be1d1-4134-4ad7-9233-b4099687bfe2-kube-api-access-zszpm" (OuterVolumeSpecName: "kube-api-access-zszpm") pod "f06be1d1-4134-4ad7-9233-b4099687bfe2" (UID: "f06be1d1-4134-4ad7-9233-b4099687bfe2"). InnerVolumeSpecName "kube-api-access-zszpm". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:53:44 crc kubenswrapper[4754]: I0126 16:53:44.294592 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7b233668-d1fb-40c4-bfd4-96374a7fbc81-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7b233668-d1fb-40c4-bfd4-96374a7fbc81" (UID: "7b233668-d1fb-40c4-bfd4-96374a7fbc81"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 16:53:44 crc kubenswrapper[4754]: I0126 16:53:44.294592 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0ca0de6e-d004-4b46-aee4-66eae716184f-kube-api-access-xwf62" (OuterVolumeSpecName: "kube-api-access-xwf62") pod "0ca0de6e-d004-4b46-aee4-66eae716184f" (UID: "0ca0de6e-d004-4b46-aee4-66eae716184f"). InnerVolumeSpecName "kube-api-access-xwf62". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:53:44 crc kubenswrapper[4754]: I0126 16:53:44.294692 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7b233668-d1fb-40c4-bfd4-96374a7fbc81-kube-api-access-d67wv" (OuterVolumeSpecName: "kube-api-access-d67wv") pod "7b233668-d1fb-40c4-bfd4-96374a7fbc81" (UID: "7b233668-d1fb-40c4-bfd4-96374a7fbc81"). InnerVolumeSpecName "kube-api-access-d67wv". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:53:44 crc kubenswrapper[4754]: I0126 16:53:44.294771 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b1b51a00-e6fe-428a-85ef-9d31badc8457-kube-api-access-8j5j7" (OuterVolumeSpecName: "kube-api-access-8j5j7") pod "b1b51a00-e6fe-428a-85ef-9d31badc8457" (UID: "b1b51a00-e6fe-428a-85ef-9d31badc8457"). InnerVolumeSpecName "kube-api-access-8j5j7". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:53:44 crc kubenswrapper[4754]: I0126 16:53:44.296640 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f06be1d1-4134-4ad7-9233-b4099687bfe2-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "f06be1d1-4134-4ad7-9233-b4099687bfe2" (UID: "f06be1d1-4134-4ad7-9233-b4099687bfe2"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 16:53:44 crc kubenswrapper[4754]: I0126 16:53:44.299877 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e36dc2d0-47de-46a6-aeba-52475754867e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e36dc2d0-47de-46a6-aeba-52475754867e" (UID: "e36dc2d0-47de-46a6-aeba-52475754867e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 16:53:44 crc kubenswrapper[4754]: I0126 16:53:44.301109 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c2962a85-523e-4893-8a19-b7baa1360c01-kube-api-access-z7pgd" (OuterVolumeSpecName: "kube-api-access-z7pgd") pod "c2962a85-523e-4893-8a19-b7baa1360c01" (UID: "c2962a85-523e-4893-8a19-b7baa1360c01"). InnerVolumeSpecName "kube-api-access-z7pgd". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:53:44 crc kubenswrapper[4754]: I0126 16:53:44.352043 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0ca0de6e-d004-4b46-aee4-66eae716184f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0ca0de6e-d004-4b46-aee4-66eae716184f" (UID: "0ca0de6e-d004-4b46-aee4-66eae716184f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 16:53:44 crc kubenswrapper[4754]: I0126 16:53:44.355648 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c2962a85-523e-4893-8a19-b7baa1360c01-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c2962a85-523e-4893-8a19-b7baa1360c01" (UID: "c2962a85-523e-4893-8a19-b7baa1360c01"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 16:53:44 crc kubenswrapper[4754]: I0126 16:53:44.367355 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/409f1ff8-a849-4e8e-a7e9-0bb57ed94cb4-client-ca\") pod \"controller-manager-7748bf895c-v5p45\" (UID: \"409f1ff8-a849-4e8e-a7e9-0bb57ed94cb4\") " pod="openshift-controller-manager/controller-manager-7748bf895c-v5p45" Jan 26 16:53:44 crc kubenswrapper[4754]: I0126 16:53:44.367415 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/409f1ff8-a849-4e8e-a7e9-0bb57ed94cb4-config\") pod \"controller-manager-7748bf895c-v5p45\" (UID: \"409f1ff8-a849-4e8e-a7e9-0bb57ed94cb4\") " pod="openshift-controller-manager/controller-manager-7748bf895c-v5p45" Jan 26 16:53:44 crc kubenswrapper[4754]: I0126 16:53:44.367460 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/409f1ff8-a849-4e8e-a7e9-0bb57ed94cb4-proxy-ca-bundles\") pod \"controller-manager-7748bf895c-v5p45\" (UID: \"409f1ff8-a849-4e8e-a7e9-0bb57ed94cb4\") " pod="openshift-controller-manager/controller-manager-7748bf895c-v5p45" Jan 26 16:53:44 crc kubenswrapper[4754]: I0126 16:53:44.367494 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w7cxf\" (UniqueName: \"kubernetes.io/projected/409f1ff8-a849-4e8e-a7e9-0bb57ed94cb4-kube-api-access-w7cxf\") pod \"controller-manager-7748bf895c-v5p45\" (UID: \"409f1ff8-a849-4e8e-a7e9-0bb57ed94cb4\") " pod="openshift-controller-manager/controller-manager-7748bf895c-v5p45" Jan 26 16:53:44 crc kubenswrapper[4754]: I0126 16:53:44.367518 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/409f1ff8-a849-4e8e-a7e9-0bb57ed94cb4-serving-cert\") pod \"controller-manager-7748bf895c-v5p45\" (UID: \"409f1ff8-a849-4e8e-a7e9-0bb57ed94cb4\") " pod="openshift-controller-manager/controller-manager-7748bf895c-v5p45" Jan 26 16:53:44 crc kubenswrapper[4754]: I0126 16:53:44.367589 4754 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e36dc2d0-47de-46a6-aeba-52475754867e-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 26 16:53:44 crc kubenswrapper[4754]: I0126 16:53:44.367602 4754 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0ca0de6e-d004-4b46-aee4-66eae716184f-utilities\") on node \"crc\" DevicePath \"\"" Jan 26 16:53:44 crc kubenswrapper[4754]: I0126 16:53:44.367614 4754 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0ca0de6e-d004-4b46-aee4-66eae716184f-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 26 16:53:44 crc kubenswrapper[4754]: I0126 16:53:44.367626 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zszpm\" (UniqueName: \"kubernetes.io/projected/f06be1d1-4134-4ad7-9233-b4099687bfe2-kube-api-access-zszpm\") on node \"crc\" DevicePath \"\"" Jan 26 16:53:44 crc kubenswrapper[4754]: I0126 16:53:44.367638 4754 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f06be1d1-4134-4ad7-9233-b4099687bfe2-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Jan 26 16:53:44 crc kubenswrapper[4754]: I0126 16:53:44.367649 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z7pgd\" (UniqueName: \"kubernetes.io/projected/c2962a85-523e-4893-8a19-b7baa1360c01-kube-api-access-z7pgd\") on node \"crc\" DevicePath \"\"" Jan 26 16:53:44 crc kubenswrapper[4754]: I0126 16:53:44.367660 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d55d7\" (UniqueName: \"kubernetes.io/projected/e36dc2d0-47de-46a6-aeba-52475754867e-kube-api-access-d55d7\") on node \"crc\" DevicePath \"\"" Jan 26 16:53:44 crc kubenswrapper[4754]: I0126 16:53:44.367692 4754 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7b233668-d1fb-40c4-bfd4-96374a7fbc81-client-ca\") on node \"crc\" DevicePath \"\"" Jan 26 16:53:44 crc kubenswrapper[4754]: I0126 16:53:44.367704 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xwf62\" (UniqueName: \"kubernetes.io/projected/0ca0de6e-d004-4b46-aee4-66eae716184f-kube-api-access-xwf62\") on node \"crc\" DevicePath \"\"" Jan 26 16:53:44 crc kubenswrapper[4754]: I0126 16:53:44.367715 4754 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c2962a85-523e-4893-8a19-b7baa1360c01-utilities\") on node \"crc\" DevicePath \"\"" Jan 26 16:53:44 crc kubenswrapper[4754]: I0126 16:53:44.367726 4754 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c2962a85-523e-4893-8a19-b7baa1360c01-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 26 16:53:44 crc kubenswrapper[4754]: I0126 16:53:44.367737 4754 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7b233668-d1fb-40c4-bfd4-96374a7fbc81-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Jan 26 16:53:44 crc kubenswrapper[4754]: I0126 16:53:44.367748 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8j5j7\" (UniqueName: \"kubernetes.io/projected/b1b51a00-e6fe-428a-85ef-9d31badc8457-kube-api-access-8j5j7\") on node \"crc\" DevicePath \"\"" Jan 26 16:53:44 crc kubenswrapper[4754]: I0126 16:53:44.367758 4754 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e36dc2d0-47de-46a6-aeba-52475754867e-utilities\") on node \"crc\" DevicePath \"\"" Jan 26 16:53:44 crc kubenswrapper[4754]: I0126 16:53:44.367768 4754 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/f06be1d1-4134-4ad7-9233-b4099687bfe2-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Jan 26 16:53:44 crc kubenswrapper[4754]: I0126 16:53:44.367780 4754 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7b233668-d1fb-40c4-bfd4-96374a7fbc81-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 26 16:53:44 crc kubenswrapper[4754]: I0126 16:53:44.367790 4754 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b1b51a00-e6fe-428a-85ef-9d31badc8457-utilities\") on node \"crc\" DevicePath \"\"" Jan 26 16:53:44 crc kubenswrapper[4754]: I0126 16:53:44.367801 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d67wv\" (UniqueName: \"kubernetes.io/projected/7b233668-d1fb-40c4-bfd4-96374a7fbc81-kube-api-access-d67wv\") on node \"crc\" DevicePath \"\"" Jan 26 16:53:44 crc kubenswrapper[4754]: I0126 16:53:44.367814 4754 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7b233668-d1fb-40c4-bfd4-96374a7fbc81-config\") on node \"crc\" DevicePath \"\"" Jan 26 16:53:44 crc kubenswrapper[4754]: I0126 16:53:44.368647 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/409f1ff8-a849-4e8e-a7e9-0bb57ed94cb4-client-ca\") pod \"controller-manager-7748bf895c-v5p45\" (UID: \"409f1ff8-a849-4e8e-a7e9-0bb57ed94cb4\") " pod="openshift-controller-manager/controller-manager-7748bf895c-v5p45" Jan 26 16:53:44 crc kubenswrapper[4754]: I0126 16:53:44.369562 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/409f1ff8-a849-4e8e-a7e9-0bb57ed94cb4-proxy-ca-bundles\") pod \"controller-manager-7748bf895c-v5p45\" (UID: \"409f1ff8-a849-4e8e-a7e9-0bb57ed94cb4\") " pod="openshift-controller-manager/controller-manager-7748bf895c-v5p45" Jan 26 16:53:44 crc kubenswrapper[4754]: I0126 16:53:44.370010 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/409f1ff8-a849-4e8e-a7e9-0bb57ed94cb4-config\") pod \"controller-manager-7748bf895c-v5p45\" (UID: \"409f1ff8-a849-4e8e-a7e9-0bb57ed94cb4\") " pod="openshift-controller-manager/controller-manager-7748bf895c-v5p45" Jan 26 16:53:44 crc kubenswrapper[4754]: I0126 16:53:44.383746 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/409f1ff8-a849-4e8e-a7e9-0bb57ed94cb4-serving-cert\") pod \"controller-manager-7748bf895c-v5p45\" (UID: \"409f1ff8-a849-4e8e-a7e9-0bb57ed94cb4\") " pod="openshift-controller-manager/controller-manager-7748bf895c-v5p45" Jan 26 16:53:44 crc kubenswrapper[4754]: I0126 16:53:44.389763 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w7cxf\" (UniqueName: \"kubernetes.io/projected/409f1ff8-a849-4e8e-a7e9-0bb57ed94cb4-kube-api-access-w7cxf\") pod \"controller-manager-7748bf895c-v5p45\" (UID: \"409f1ff8-a849-4e8e-a7e9-0bb57ed94cb4\") " pod="openshift-controller-manager/controller-manager-7748bf895c-v5p45" Jan 26 16:53:44 crc kubenswrapper[4754]: I0126 16:53:44.440122 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b1b51a00-e6fe-428a-85ef-9d31badc8457-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b1b51a00-e6fe-428a-85ef-9d31badc8457" (UID: "b1b51a00-e6fe-428a-85ef-9d31badc8457"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 16:53:44 crc kubenswrapper[4754]: I0126 16:53:44.468628 4754 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b1b51a00-e6fe-428a-85ef-9d31badc8457-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 26 16:53:44 crc kubenswrapper[4754]: I0126 16:53:44.471815 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7748bf895c-v5p45" Jan 26 16:53:44 crc kubenswrapper[4754]: I0126 16:53:44.663797 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-7748bf895c-v5p45"] Jan 26 16:53:45 crc kubenswrapper[4754]: I0126 16:53:45.043240 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cgm62" event={"ID":"c2962a85-523e-4893-8a19-b7baa1360c01","Type":"ContainerDied","Data":"883b049d0221704c56bf963e8d4d31fbcca1c49b4524942c567e53ec46248777"} Jan 26 16:53:45 crc kubenswrapper[4754]: I0126 16:53:45.043617 4754 scope.go:117] "RemoveContainer" containerID="3533e22f9c0731ad76ebd37e68266f9c0db47db9605ad038aa9b1e1df2d19b8f" Jan 26 16:53:45 crc kubenswrapper[4754]: I0126 16:53:45.043271 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cgm62" Jan 26 16:53:45 crc kubenswrapper[4754]: I0126 16:53:45.046133 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5rvw2" Jan 26 16:53:45 crc kubenswrapper[4754]: I0126 16:53:45.046127 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5rvw2" event={"ID":"b1b51a00-e6fe-428a-85ef-9d31badc8457","Type":"ContainerDied","Data":"d16b4b2bb0d8fe98b530a4f93ec01d1a0c5511b630a321508338e6bb65347d6b"} Jan 26 16:53:45 crc kubenswrapper[4754]: I0126 16:53:45.055285 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7748bf895c-v5p45" event={"ID":"409f1ff8-a849-4e8e-a7e9-0bb57ed94cb4","Type":"ContainerStarted","Data":"32c537bc8599a821e4d60b2dd1d9e0eadfa2109b63483ea12cf22ac59e0b7ce5"} Jan 26 16:53:45 crc kubenswrapper[4754]: I0126 16:53:45.055325 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7748bf895c-v5p45" event={"ID":"409f1ff8-a849-4e8e-a7e9-0bb57ed94cb4","Type":"ContainerStarted","Data":"6961f858a713b34ceda5911de7ea2b6b54c15ca4411025717781cfc213d7ff19"} Jan 26 16:53:45 crc kubenswrapper[4754]: I0126 16:53:45.056455 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-7748bf895c-v5p45" Jan 26 16:53:45 crc kubenswrapper[4754]: I0126 16:53:45.061249 4754 scope.go:117] "RemoveContainer" containerID="fb749d40f32892c54ca3fb9072c0e952b9194886ebca7c67687290380d14da66" Jan 26 16:53:45 crc kubenswrapper[4754]: I0126 16:53:45.064053 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qzw56" Jan 26 16:53:45 crc kubenswrapper[4754]: I0126 16:53:45.065038 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qzw56" event={"ID":"0ca0de6e-d004-4b46-aee4-66eae716184f","Type":"ContainerDied","Data":"cf4854adfd91381464448a0db0431004cb6acb1b31a263fa88df803c4200c819"} Jan 26 16:53:45 crc kubenswrapper[4754]: I0126 16:53:45.076375 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-7748bf895c-v5p45" Jan 26 16:53:45 crc kubenswrapper[4754]: I0126 16:53:45.080033 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hhhn5" event={"ID":"e36dc2d0-47de-46a6-aeba-52475754867e","Type":"ContainerDied","Data":"fc6f5c5f1c7db70e9c330d4135008ece5d08f6a86648d7e02cda040ca7905d48"} Jan 26 16:53:45 crc kubenswrapper[4754]: I0126 16:53:45.080184 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hhhn5" Jan 26 16:53:45 crc kubenswrapper[4754]: I0126 16:53:45.087009 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-7xsg2" event={"ID":"f06be1d1-4134-4ad7-9233-b4099687bfe2","Type":"ContainerDied","Data":"8450fa38eda3b5cba6c4dda738ea2aa9cad3994786cefe2ac842546f0b4269f0"} Jan 26 16:53:45 crc kubenswrapper[4754]: I0126 16:53:45.087805 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-7xsg2" Jan 26 16:53:45 crc kubenswrapper[4754]: I0126 16:53:45.088217 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-d5469898f-fs6b7" Jan 26 16:53:45 crc kubenswrapper[4754]: I0126 16:53:45.091456 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-mp4qz" Jan 26 16:53:45 crc kubenswrapper[4754]: I0126 16:53:45.103410 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-7748bf895c-v5p45" podStartSLOduration=52.103384469 podStartE2EDuration="52.103384469s" podCreationTimestamp="2026-01-26 16:52:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 16:53:45.082226855 +0000 UTC m=+391.606407289" watchObservedRunningTime="2026-01-26 16:53:45.103384469 +0000 UTC m=+391.627564903" Jan 26 16:53:45 crc kubenswrapper[4754]: I0126 16:53:45.110978 4754 scope.go:117] "RemoveContainer" containerID="4df8dfad1bfb4083b96d0aa561d538a881291c90d138da405053526c0dba8ee4" Jan 26 16:53:45 crc kubenswrapper[4754]: I0126 16:53:45.117242 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-cgm62"] Jan 26 16:53:45 crc kubenswrapper[4754]: I0126 16:53:45.122792 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-cgm62"] Jan 26 16:53:45 crc kubenswrapper[4754]: I0126 16:53:45.139316 4754 scope.go:117] "RemoveContainer" containerID="bcf0497d70be7f24baa4f82c8fe6e79184c60b7185ce37f72e6a06cb1f133a33" Jan 26 16:53:45 crc kubenswrapper[4754]: I0126 16:53:45.157203 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-5rvw2"] Jan 26 16:53:45 crc kubenswrapper[4754]: I0126 16:53:45.162307 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-5rvw2"] Jan 26 16:53:45 crc kubenswrapper[4754]: I0126 16:53:45.162795 4754 scope.go:117] "RemoveContainer" containerID="48dce29db1605144274c78c071b5a77b3561d6c5d566b2d0b375a641b11eeb33" Jan 26 16:53:45 crc kubenswrapper[4754]: I0126 16:53:45.185994 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-7xsg2"] Jan 26 16:53:45 crc kubenswrapper[4754]: I0126 16:53:45.189876 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-7xsg2"] Jan 26 16:53:45 crc kubenswrapper[4754]: I0126 16:53:45.193111 4754 scope.go:117] "RemoveContainer" containerID="9960a5f602fda10c53a46543fdf587ea8c2512b205a707973e1f2eb83783eb4e" Jan 26 16:53:45 crc kubenswrapper[4754]: I0126 16:53:45.207577 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-hhhn5"] Jan 26 16:53:45 crc kubenswrapper[4754]: I0126 16:53:45.214140 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-hhhn5"] Jan 26 16:53:45 crc kubenswrapper[4754]: I0126 16:53:45.216174 4754 scope.go:117] "RemoveContainer" containerID="c87ab9b579140b0a54235a7b7812977815f23e9829ef9fadcf2f4775443d3c8f" Jan 26 16:53:45 crc kubenswrapper[4754]: I0126 16:53:45.225859 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-d5469898f-fs6b7"] Jan 26 16:53:45 crc kubenswrapper[4754]: I0126 16:53:45.229394 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-d5469898f-fs6b7"] Jan 26 16:53:45 crc kubenswrapper[4754]: I0126 16:53:45.268320 4754 scope.go:117] "RemoveContainer" containerID="3773bafe0080ca3c474a93c1b34b4f446d36eb3f9c3e6cda5f3da21f5babf026" Jan 26 16:53:45 crc kubenswrapper[4754]: I0126 16:53:45.270403 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-qzw56"] Jan 26 16:53:45 crc kubenswrapper[4754]: I0126 16:53:45.277199 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-qzw56"] Jan 26 16:53:45 crc kubenswrapper[4754]: I0126 16:53:45.285678 4754 scope.go:117] "RemoveContainer" containerID="c6ef6f376b3de6a06cb247c19a2dde75e11dca8e32c9db492300d119a97e789b" Jan 26 16:53:45 crc kubenswrapper[4754]: I0126 16:53:45.303354 4754 scope.go:117] "RemoveContainer" containerID="2e25b7bf0c7d3598416895619dfd7bf1e691d76eb5e80aee438983164baf5bf0" Jan 26 16:53:45 crc kubenswrapper[4754]: I0126 16:53:45.325585 4754 scope.go:117] "RemoveContainer" containerID="d20015cf16e678be973812f50ecb7e568826bfca9eeb39c0d06deeda2df3a367" Jan 26 16:53:45 crc kubenswrapper[4754]: I0126 16:53:45.343514 4754 scope.go:117] "RemoveContainer" containerID="fc7875a2ec569f0ebeb226ca273f9f8ee7f570b1b8622e77565446d2a859eaa3" Jan 26 16:53:45 crc kubenswrapper[4754]: I0126 16:53:45.359344 4754 scope.go:117] "RemoveContainer" containerID="57b65fbcda96a2d7c3ad5056427ff581fc02c615544c7b716f61866017c20b02" Jan 26 16:53:45 crc kubenswrapper[4754]: I0126 16:53:45.774739 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0ca0de6e-d004-4b46-aee4-66eae716184f" path="/var/lib/kubelet/pods/0ca0de6e-d004-4b46-aee4-66eae716184f/volumes" Jan 26 16:53:45 crc kubenswrapper[4754]: I0126 16:53:45.775371 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7b233668-d1fb-40c4-bfd4-96374a7fbc81" path="/var/lib/kubelet/pods/7b233668-d1fb-40c4-bfd4-96374a7fbc81/volumes" Jan 26 16:53:45 crc kubenswrapper[4754]: I0126 16:53:45.775857 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b1b51a00-e6fe-428a-85ef-9d31badc8457" path="/var/lib/kubelet/pods/b1b51a00-e6fe-428a-85ef-9d31badc8457/volumes" Jan 26 16:53:45 crc kubenswrapper[4754]: I0126 16:53:45.776828 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c2962a85-523e-4893-8a19-b7baa1360c01" path="/var/lib/kubelet/pods/c2962a85-523e-4893-8a19-b7baa1360c01/volumes" Jan 26 16:53:45 crc kubenswrapper[4754]: I0126 16:53:45.777376 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e36dc2d0-47de-46a6-aeba-52475754867e" path="/var/lib/kubelet/pods/e36dc2d0-47de-46a6-aeba-52475754867e/volumes" Jan 26 16:53:45 crc kubenswrapper[4754]: I0126 16:53:45.778379 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f06be1d1-4134-4ad7-9233-b4099687bfe2" path="/var/lib/kubelet/pods/f06be1d1-4134-4ad7-9233-b4099687bfe2/volumes" Jan 26 16:53:46 crc kubenswrapper[4754]: I0126 16:53:46.322107 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-l482d"] Jan 26 16:53:46 crc kubenswrapper[4754]: E0126 16:53:46.322304 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c2962a85-523e-4893-8a19-b7baa1360c01" containerName="extract-content" Jan 26 16:53:46 crc kubenswrapper[4754]: I0126 16:53:46.322315 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="c2962a85-523e-4893-8a19-b7baa1360c01" containerName="extract-content" Jan 26 16:53:46 crc kubenswrapper[4754]: E0126 16:53:46.322322 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f06be1d1-4134-4ad7-9233-b4099687bfe2" containerName="marketplace-operator" Jan 26 16:53:46 crc kubenswrapper[4754]: I0126 16:53:46.322329 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="f06be1d1-4134-4ad7-9233-b4099687bfe2" containerName="marketplace-operator" Jan 26 16:53:46 crc kubenswrapper[4754]: E0126 16:53:46.322340 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b1b51a00-e6fe-428a-85ef-9d31badc8457" containerName="extract-utilities" Jan 26 16:53:46 crc kubenswrapper[4754]: I0126 16:53:46.322346 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="b1b51a00-e6fe-428a-85ef-9d31badc8457" containerName="extract-utilities" Jan 26 16:53:46 crc kubenswrapper[4754]: E0126 16:53:46.322354 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e36dc2d0-47de-46a6-aeba-52475754867e" containerName="extract-content" Jan 26 16:53:46 crc kubenswrapper[4754]: I0126 16:53:46.322360 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="e36dc2d0-47de-46a6-aeba-52475754867e" containerName="extract-content" Jan 26 16:53:46 crc kubenswrapper[4754]: E0126 16:53:46.322368 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c2962a85-523e-4893-8a19-b7baa1360c01" containerName="registry-server" Jan 26 16:53:46 crc kubenswrapper[4754]: I0126 16:53:46.322373 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="c2962a85-523e-4893-8a19-b7baa1360c01" containerName="registry-server" Jan 26 16:53:46 crc kubenswrapper[4754]: E0126 16:53:46.322383 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b1b51a00-e6fe-428a-85ef-9d31badc8457" containerName="extract-content" Jan 26 16:53:46 crc kubenswrapper[4754]: I0126 16:53:46.322389 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="b1b51a00-e6fe-428a-85ef-9d31badc8457" containerName="extract-content" Jan 26 16:53:46 crc kubenswrapper[4754]: E0126 16:53:46.322397 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c2962a85-523e-4893-8a19-b7baa1360c01" containerName="extract-utilities" Jan 26 16:53:46 crc kubenswrapper[4754]: I0126 16:53:46.322403 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="c2962a85-523e-4893-8a19-b7baa1360c01" containerName="extract-utilities" Jan 26 16:53:46 crc kubenswrapper[4754]: E0126 16:53:46.322410 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b1b51a00-e6fe-428a-85ef-9d31badc8457" containerName="registry-server" Jan 26 16:53:46 crc kubenswrapper[4754]: I0126 16:53:46.322416 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="b1b51a00-e6fe-428a-85ef-9d31badc8457" containerName="registry-server" Jan 26 16:53:46 crc kubenswrapper[4754]: E0126 16:53:46.322425 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e36dc2d0-47de-46a6-aeba-52475754867e" containerName="registry-server" Jan 26 16:53:46 crc kubenswrapper[4754]: I0126 16:53:46.322431 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="e36dc2d0-47de-46a6-aeba-52475754867e" containerName="registry-server" Jan 26 16:53:46 crc kubenswrapper[4754]: E0126 16:53:46.322438 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e36dc2d0-47de-46a6-aeba-52475754867e" containerName="extract-utilities" Jan 26 16:53:46 crc kubenswrapper[4754]: I0126 16:53:46.322444 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="e36dc2d0-47de-46a6-aeba-52475754867e" containerName="extract-utilities" Jan 26 16:53:46 crc kubenswrapper[4754]: I0126 16:53:46.322551 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="e36dc2d0-47de-46a6-aeba-52475754867e" containerName="registry-server" Jan 26 16:53:46 crc kubenswrapper[4754]: I0126 16:53:46.322563 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="b1b51a00-e6fe-428a-85ef-9d31badc8457" containerName="registry-server" Jan 26 16:53:46 crc kubenswrapper[4754]: I0126 16:53:46.322571 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="c2962a85-523e-4893-8a19-b7baa1360c01" containerName="registry-server" Jan 26 16:53:46 crc kubenswrapper[4754]: I0126 16:53:46.339460 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-l482d" Jan 26 16:53:46 crc kubenswrapper[4754]: I0126 16:53:46.344138 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Jan 26 16:53:46 crc kubenswrapper[4754]: I0126 16:53:46.345855 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-l482d"] Jan 26 16:53:46 crc kubenswrapper[4754]: I0126 16:53:46.496660 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6s9wg\" (UniqueName: \"kubernetes.io/projected/833110b5-684c-4097-9e52-d3d8d1838e73-kube-api-access-6s9wg\") pod \"certified-operators-l482d\" (UID: \"833110b5-684c-4097-9e52-d3d8d1838e73\") " pod="openshift-marketplace/certified-operators-l482d" Jan 26 16:53:46 crc kubenswrapper[4754]: I0126 16:53:46.497013 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/833110b5-684c-4097-9e52-d3d8d1838e73-catalog-content\") pod \"certified-operators-l482d\" (UID: \"833110b5-684c-4097-9e52-d3d8d1838e73\") " pod="openshift-marketplace/certified-operators-l482d" Jan 26 16:53:46 crc kubenswrapper[4754]: I0126 16:53:46.497030 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/833110b5-684c-4097-9e52-d3d8d1838e73-utilities\") pod \"certified-operators-l482d\" (UID: \"833110b5-684c-4097-9e52-d3d8d1838e73\") " pod="openshift-marketplace/certified-operators-l482d" Jan 26 16:53:46 crc kubenswrapper[4754]: I0126 16:53:46.522397 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-c44vn"] Jan 26 16:53:46 crc kubenswrapper[4754]: I0126 16:53:46.523336 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-c44vn" Jan 26 16:53:46 crc kubenswrapper[4754]: I0126 16:53:46.524856 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Jan 26 16:53:46 crc kubenswrapper[4754]: I0126 16:53:46.536371 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-c44vn"] Jan 26 16:53:46 crc kubenswrapper[4754]: I0126 16:53:46.598435 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/833110b5-684c-4097-9e52-d3d8d1838e73-utilities\") pod \"certified-operators-l482d\" (UID: \"833110b5-684c-4097-9e52-d3d8d1838e73\") " pod="openshift-marketplace/certified-operators-l482d" Jan 26 16:53:46 crc kubenswrapper[4754]: I0126 16:53:46.598492 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/833110b5-684c-4097-9e52-d3d8d1838e73-catalog-content\") pod \"certified-operators-l482d\" (UID: \"833110b5-684c-4097-9e52-d3d8d1838e73\") " pod="openshift-marketplace/certified-operators-l482d" Jan 26 16:53:46 crc kubenswrapper[4754]: I0126 16:53:46.598553 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6s9wg\" (UniqueName: \"kubernetes.io/projected/833110b5-684c-4097-9e52-d3d8d1838e73-kube-api-access-6s9wg\") pod \"certified-operators-l482d\" (UID: \"833110b5-684c-4097-9e52-d3d8d1838e73\") " pod="openshift-marketplace/certified-operators-l482d" Jan 26 16:53:46 crc kubenswrapper[4754]: I0126 16:53:46.599023 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/833110b5-684c-4097-9e52-d3d8d1838e73-utilities\") pod \"certified-operators-l482d\" (UID: \"833110b5-684c-4097-9e52-d3d8d1838e73\") " pod="openshift-marketplace/certified-operators-l482d" Jan 26 16:53:46 crc kubenswrapper[4754]: I0126 16:53:46.599099 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/833110b5-684c-4097-9e52-d3d8d1838e73-catalog-content\") pod \"certified-operators-l482d\" (UID: \"833110b5-684c-4097-9e52-d3d8d1838e73\") " pod="openshift-marketplace/certified-operators-l482d" Jan 26 16:53:46 crc kubenswrapper[4754]: I0126 16:53:46.619733 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6s9wg\" (UniqueName: \"kubernetes.io/projected/833110b5-684c-4097-9e52-d3d8d1838e73-kube-api-access-6s9wg\") pod \"certified-operators-l482d\" (UID: \"833110b5-684c-4097-9e52-d3d8d1838e73\") " pod="openshift-marketplace/certified-operators-l482d" Jan 26 16:53:46 crc kubenswrapper[4754]: I0126 16:53:46.659576 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-l482d" Jan 26 16:53:46 crc kubenswrapper[4754]: I0126 16:53:46.699182 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/02fabee4-adea-4bca-ba0a-e6b98f6e68dd-catalog-content\") pod \"community-operators-c44vn\" (UID: \"02fabee4-adea-4bca-ba0a-e6b98f6e68dd\") " pod="openshift-marketplace/community-operators-c44vn" Jan 26 16:53:46 crc kubenswrapper[4754]: I0126 16:53:46.699460 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/02fabee4-adea-4bca-ba0a-e6b98f6e68dd-utilities\") pod \"community-operators-c44vn\" (UID: \"02fabee4-adea-4bca-ba0a-e6b98f6e68dd\") " pod="openshift-marketplace/community-operators-c44vn" Jan 26 16:53:46 crc kubenswrapper[4754]: I0126 16:53:46.699592 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lb2fs\" (UniqueName: \"kubernetes.io/projected/02fabee4-adea-4bca-ba0a-e6b98f6e68dd-kube-api-access-lb2fs\") pod \"community-operators-c44vn\" (UID: \"02fabee4-adea-4bca-ba0a-e6b98f6e68dd\") " pod="openshift-marketplace/community-operators-c44vn" Jan 26 16:53:46 crc kubenswrapper[4754]: I0126 16:53:46.802204 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/02fabee4-adea-4bca-ba0a-e6b98f6e68dd-utilities\") pod \"community-operators-c44vn\" (UID: \"02fabee4-adea-4bca-ba0a-e6b98f6e68dd\") " pod="openshift-marketplace/community-operators-c44vn" Jan 26 16:53:46 crc kubenswrapper[4754]: I0126 16:53:46.802289 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lb2fs\" (UniqueName: \"kubernetes.io/projected/02fabee4-adea-4bca-ba0a-e6b98f6e68dd-kube-api-access-lb2fs\") pod \"community-operators-c44vn\" (UID: \"02fabee4-adea-4bca-ba0a-e6b98f6e68dd\") " pod="openshift-marketplace/community-operators-c44vn" Jan 26 16:53:46 crc kubenswrapper[4754]: I0126 16:53:46.802371 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/02fabee4-adea-4bca-ba0a-e6b98f6e68dd-catalog-content\") pod \"community-operators-c44vn\" (UID: \"02fabee4-adea-4bca-ba0a-e6b98f6e68dd\") " pod="openshift-marketplace/community-operators-c44vn" Jan 26 16:53:46 crc kubenswrapper[4754]: I0126 16:53:46.802949 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/02fabee4-adea-4bca-ba0a-e6b98f6e68dd-utilities\") pod \"community-operators-c44vn\" (UID: \"02fabee4-adea-4bca-ba0a-e6b98f6e68dd\") " pod="openshift-marketplace/community-operators-c44vn" Jan 26 16:53:46 crc kubenswrapper[4754]: I0126 16:53:46.803785 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/02fabee4-adea-4bca-ba0a-e6b98f6e68dd-catalog-content\") pod \"community-operators-c44vn\" (UID: \"02fabee4-adea-4bca-ba0a-e6b98f6e68dd\") " pod="openshift-marketplace/community-operators-c44vn" Jan 26 16:53:46 crc kubenswrapper[4754]: I0126 16:53:46.824074 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lb2fs\" (UniqueName: \"kubernetes.io/projected/02fabee4-adea-4bca-ba0a-e6b98f6e68dd-kube-api-access-lb2fs\") pod \"community-operators-c44vn\" (UID: \"02fabee4-adea-4bca-ba0a-e6b98f6e68dd\") " pod="openshift-marketplace/community-operators-c44vn" Jan 26 16:53:46 crc kubenswrapper[4754]: I0126 16:53:46.842013 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-c44vn" Jan 26 16:53:47 crc kubenswrapper[4754]: I0126 16:53:47.065724 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-l482d"] Jan 26 16:53:47 crc kubenswrapper[4754]: W0126 16:53:47.077881 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod833110b5_684c_4097_9e52_d3d8d1838e73.slice/crio-a9596fc361a04720af3a7646e6e2f9902add150a9947aeda0c6cbdf57d75b591 WatchSource:0}: Error finding container a9596fc361a04720af3a7646e6e2f9902add150a9947aeda0c6cbdf57d75b591: Status 404 returned error can't find the container with id a9596fc361a04720af3a7646e6e2f9902add150a9947aeda0c6cbdf57d75b591 Jan 26 16:53:47 crc kubenswrapper[4754]: I0126 16:53:47.105220 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-l482d" event={"ID":"833110b5-684c-4097-9e52-d3d8d1838e73","Type":"ContainerStarted","Data":"a9596fc361a04720af3a7646e6e2f9902add150a9947aeda0c6cbdf57d75b591"} Jan 26 16:53:47 crc kubenswrapper[4754]: I0126 16:53:47.228927 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-c44vn"] Jan 26 16:53:48 crc kubenswrapper[4754]: I0126 16:53:48.117574 4754 generic.go:334] "Generic (PLEG): container finished" podID="833110b5-684c-4097-9e52-d3d8d1838e73" containerID="cdafb1c82434540940dc94751916e14ab3f96474a9b0b50c0dfcee61e2b193fd" exitCode=0 Jan 26 16:53:48 crc kubenswrapper[4754]: I0126 16:53:48.117660 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-l482d" event={"ID":"833110b5-684c-4097-9e52-d3d8d1838e73","Type":"ContainerDied","Data":"cdafb1c82434540940dc94751916e14ab3f96474a9b0b50c0dfcee61e2b193fd"} Jan 26 16:53:48 crc kubenswrapper[4754]: I0126 16:53:48.120184 4754 generic.go:334] "Generic (PLEG): container finished" podID="02fabee4-adea-4bca-ba0a-e6b98f6e68dd" containerID="925fb011d02392bde0e28d49d484809b9f1b9b576ab95688a177d8d856380341" exitCode=0 Jan 26 16:53:48 crc kubenswrapper[4754]: I0126 16:53:48.120282 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-c44vn" event={"ID":"02fabee4-adea-4bca-ba0a-e6b98f6e68dd","Type":"ContainerDied","Data":"925fb011d02392bde0e28d49d484809b9f1b9b576ab95688a177d8d856380341"} Jan 26 16:53:48 crc kubenswrapper[4754]: I0126 16:53:48.120310 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-c44vn" event={"ID":"02fabee4-adea-4bca-ba0a-e6b98f6e68dd","Type":"ContainerStarted","Data":"b562999c6be5b50382dbebe637e35841519d0b712b255eed977cb5812542b5d5"} Jan 26 16:53:48 crc kubenswrapper[4754]: I0126 16:53:48.720315 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-xc89f"] Jan 26 16:53:48 crc kubenswrapper[4754]: I0126 16:53:48.721382 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xc89f" Jan 26 16:53:48 crc kubenswrapper[4754]: I0126 16:53:48.723362 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Jan 26 16:53:48 crc kubenswrapper[4754]: I0126 16:53:48.730880 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-xc89f"] Jan 26 16:53:48 crc kubenswrapper[4754]: I0126 16:53:48.828327 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/231baab8-113c-494b-b5d3-b169370901bf-catalog-content\") pod \"redhat-marketplace-xc89f\" (UID: \"231baab8-113c-494b-b5d3-b169370901bf\") " pod="openshift-marketplace/redhat-marketplace-xc89f" Jan 26 16:53:48 crc kubenswrapper[4754]: I0126 16:53:48.828474 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/231baab8-113c-494b-b5d3-b169370901bf-utilities\") pod \"redhat-marketplace-xc89f\" (UID: \"231baab8-113c-494b-b5d3-b169370901bf\") " pod="openshift-marketplace/redhat-marketplace-xc89f" Jan 26 16:53:48 crc kubenswrapper[4754]: I0126 16:53:48.828534 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6qskt\" (UniqueName: \"kubernetes.io/projected/231baab8-113c-494b-b5d3-b169370901bf-kube-api-access-6qskt\") pod \"redhat-marketplace-xc89f\" (UID: \"231baab8-113c-494b-b5d3-b169370901bf\") " pod="openshift-marketplace/redhat-marketplace-xc89f" Jan 26 16:53:48 crc kubenswrapper[4754]: I0126 16:53:48.921215 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-dn77h"] Jan 26 16:53:48 crc kubenswrapper[4754]: I0126 16:53:48.922644 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dn77h" Jan 26 16:53:48 crc kubenswrapper[4754]: I0126 16:53:48.925030 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Jan 26 16:53:48 crc kubenswrapper[4754]: I0126 16:53:48.929486 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/231baab8-113c-494b-b5d3-b169370901bf-utilities\") pod \"redhat-marketplace-xc89f\" (UID: \"231baab8-113c-494b-b5d3-b169370901bf\") " pod="openshift-marketplace/redhat-marketplace-xc89f" Jan 26 16:53:48 crc kubenswrapper[4754]: I0126 16:53:48.929556 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6qskt\" (UniqueName: \"kubernetes.io/projected/231baab8-113c-494b-b5d3-b169370901bf-kube-api-access-6qskt\") pod \"redhat-marketplace-xc89f\" (UID: \"231baab8-113c-494b-b5d3-b169370901bf\") " pod="openshift-marketplace/redhat-marketplace-xc89f" Jan 26 16:53:48 crc kubenswrapper[4754]: I0126 16:53:48.929595 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/231baab8-113c-494b-b5d3-b169370901bf-catalog-content\") pod \"redhat-marketplace-xc89f\" (UID: \"231baab8-113c-494b-b5d3-b169370901bf\") " pod="openshift-marketplace/redhat-marketplace-xc89f" Jan 26 16:53:48 crc kubenswrapper[4754]: I0126 16:53:48.930059 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/231baab8-113c-494b-b5d3-b169370901bf-catalog-content\") pod \"redhat-marketplace-xc89f\" (UID: \"231baab8-113c-494b-b5d3-b169370901bf\") " pod="openshift-marketplace/redhat-marketplace-xc89f" Jan 26 16:53:48 crc kubenswrapper[4754]: I0126 16:53:48.930363 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/231baab8-113c-494b-b5d3-b169370901bf-utilities\") pod \"redhat-marketplace-xc89f\" (UID: \"231baab8-113c-494b-b5d3-b169370901bf\") " pod="openshift-marketplace/redhat-marketplace-xc89f" Jan 26 16:53:48 crc kubenswrapper[4754]: I0126 16:53:48.931783 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-dn77h"] Jan 26 16:53:48 crc kubenswrapper[4754]: I0126 16:53:48.951037 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6qskt\" (UniqueName: \"kubernetes.io/projected/231baab8-113c-494b-b5d3-b169370901bf-kube-api-access-6qskt\") pod \"redhat-marketplace-xc89f\" (UID: \"231baab8-113c-494b-b5d3-b169370901bf\") " pod="openshift-marketplace/redhat-marketplace-xc89f" Jan 26 16:53:49 crc kubenswrapper[4754]: I0126 16:53:49.030372 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sgcrr\" (UniqueName: \"kubernetes.io/projected/2dfa533f-5161-4ef0-bc9b-397abae75b23-kube-api-access-sgcrr\") pod \"redhat-operators-dn77h\" (UID: \"2dfa533f-5161-4ef0-bc9b-397abae75b23\") " pod="openshift-marketplace/redhat-operators-dn77h" Jan 26 16:53:49 crc kubenswrapper[4754]: I0126 16:53:49.030422 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2dfa533f-5161-4ef0-bc9b-397abae75b23-catalog-content\") pod \"redhat-operators-dn77h\" (UID: \"2dfa533f-5161-4ef0-bc9b-397abae75b23\") " pod="openshift-marketplace/redhat-operators-dn77h" Jan 26 16:53:49 crc kubenswrapper[4754]: I0126 16:53:49.030484 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2dfa533f-5161-4ef0-bc9b-397abae75b23-utilities\") pod \"redhat-operators-dn77h\" (UID: \"2dfa533f-5161-4ef0-bc9b-397abae75b23\") " pod="openshift-marketplace/redhat-operators-dn77h" Jan 26 16:53:49 crc kubenswrapper[4754]: I0126 16:53:49.039424 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xc89f" Jan 26 16:53:49 crc kubenswrapper[4754]: I0126 16:53:49.131605 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sgcrr\" (UniqueName: \"kubernetes.io/projected/2dfa533f-5161-4ef0-bc9b-397abae75b23-kube-api-access-sgcrr\") pod \"redhat-operators-dn77h\" (UID: \"2dfa533f-5161-4ef0-bc9b-397abae75b23\") " pod="openshift-marketplace/redhat-operators-dn77h" Jan 26 16:53:49 crc kubenswrapper[4754]: I0126 16:53:49.131659 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2dfa533f-5161-4ef0-bc9b-397abae75b23-catalog-content\") pod \"redhat-operators-dn77h\" (UID: \"2dfa533f-5161-4ef0-bc9b-397abae75b23\") " pod="openshift-marketplace/redhat-operators-dn77h" Jan 26 16:53:49 crc kubenswrapper[4754]: I0126 16:53:49.131891 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2dfa533f-5161-4ef0-bc9b-397abae75b23-utilities\") pod \"redhat-operators-dn77h\" (UID: \"2dfa533f-5161-4ef0-bc9b-397abae75b23\") " pod="openshift-marketplace/redhat-operators-dn77h" Jan 26 16:53:49 crc kubenswrapper[4754]: I0126 16:53:49.132739 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2dfa533f-5161-4ef0-bc9b-397abae75b23-utilities\") pod \"redhat-operators-dn77h\" (UID: \"2dfa533f-5161-4ef0-bc9b-397abae75b23\") " pod="openshift-marketplace/redhat-operators-dn77h" Jan 26 16:53:49 crc kubenswrapper[4754]: I0126 16:53:49.133512 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2dfa533f-5161-4ef0-bc9b-397abae75b23-catalog-content\") pod \"redhat-operators-dn77h\" (UID: \"2dfa533f-5161-4ef0-bc9b-397abae75b23\") " pod="openshift-marketplace/redhat-operators-dn77h" Jan 26 16:53:49 crc kubenswrapper[4754]: I0126 16:53:49.160151 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sgcrr\" (UniqueName: \"kubernetes.io/projected/2dfa533f-5161-4ef0-bc9b-397abae75b23-kube-api-access-sgcrr\") pod \"redhat-operators-dn77h\" (UID: \"2dfa533f-5161-4ef0-bc9b-397abae75b23\") " pod="openshift-marketplace/redhat-operators-dn77h" Jan 26 16:53:49 crc kubenswrapper[4754]: I0126 16:53:49.237725 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dn77h" Jan 26 16:53:49 crc kubenswrapper[4754]: I0126 16:53:49.473930 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-xc89f"] Jan 26 16:53:49 crc kubenswrapper[4754]: I0126 16:53:49.645972 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-dn77h"] Jan 26 16:53:49 crc kubenswrapper[4754]: W0126 16:53:49.654281 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2dfa533f_5161_4ef0_bc9b_397abae75b23.slice/crio-54fb798dc06e6042d61ad63638ea0c7b76797d651dace81e82c62b3afdc111da WatchSource:0}: Error finding container 54fb798dc06e6042d61ad63638ea0c7b76797d651dace81e82c62b3afdc111da: Status 404 returned error can't find the container with id 54fb798dc06e6042d61ad63638ea0c7b76797d651dace81e82c62b3afdc111da Jan 26 16:53:50 crc kubenswrapper[4754]: I0126 16:53:50.141825 4754 generic.go:334] "Generic (PLEG): container finished" podID="02fabee4-adea-4bca-ba0a-e6b98f6e68dd" containerID="8620641dc513c3893075413f6dcf5fe4c24d5dc75a453ca93cfea7d584981377" exitCode=0 Jan 26 16:53:50 crc kubenswrapper[4754]: I0126 16:53:50.142139 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-c44vn" event={"ID":"02fabee4-adea-4bca-ba0a-e6b98f6e68dd","Type":"ContainerDied","Data":"8620641dc513c3893075413f6dcf5fe4c24d5dc75a453ca93cfea7d584981377"} Jan 26 16:53:50 crc kubenswrapper[4754]: I0126 16:53:50.147230 4754 generic.go:334] "Generic (PLEG): container finished" podID="231baab8-113c-494b-b5d3-b169370901bf" containerID="6ea4becaf6122b4a10f899d30011e0eac95621997405e503d8feb75922612aa8" exitCode=0 Jan 26 16:53:50 crc kubenswrapper[4754]: I0126 16:53:50.147344 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xc89f" event={"ID":"231baab8-113c-494b-b5d3-b169370901bf","Type":"ContainerDied","Data":"6ea4becaf6122b4a10f899d30011e0eac95621997405e503d8feb75922612aa8"} Jan 26 16:53:50 crc kubenswrapper[4754]: I0126 16:53:50.147383 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xc89f" event={"ID":"231baab8-113c-494b-b5d3-b169370901bf","Type":"ContainerStarted","Data":"3ebfa83a731a856ba08337b6e788ae8844d894f48770a85b5a57ec9111aa69d4"} Jan 26 16:53:50 crc kubenswrapper[4754]: I0126 16:53:50.154951 4754 generic.go:334] "Generic (PLEG): container finished" podID="2dfa533f-5161-4ef0-bc9b-397abae75b23" containerID="3724f201f0cbc6357edb0489ff7f5469b22571707f89b7efae371129d0feec76" exitCode=0 Jan 26 16:53:50 crc kubenswrapper[4754]: I0126 16:53:50.155041 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dn77h" event={"ID":"2dfa533f-5161-4ef0-bc9b-397abae75b23","Type":"ContainerDied","Data":"3724f201f0cbc6357edb0489ff7f5469b22571707f89b7efae371129d0feec76"} Jan 26 16:53:50 crc kubenswrapper[4754]: I0126 16:53:50.155092 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dn77h" event={"ID":"2dfa533f-5161-4ef0-bc9b-397abae75b23","Type":"ContainerStarted","Data":"54fb798dc06e6042d61ad63638ea0c7b76797d651dace81e82c62b3afdc111da"} Jan 26 16:53:50 crc kubenswrapper[4754]: I0126 16:53:50.161232 4754 generic.go:334] "Generic (PLEG): container finished" podID="833110b5-684c-4097-9e52-d3d8d1838e73" containerID="567bc05d44e76fe18b6a1e23914ff4f2bfb8eb85a95eecdfe15f9a5d80eb1ef2" exitCode=0 Jan 26 16:53:50 crc kubenswrapper[4754]: I0126 16:53:50.161271 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-l482d" event={"ID":"833110b5-684c-4097-9e52-d3d8d1838e73","Type":"ContainerDied","Data":"567bc05d44e76fe18b6a1e23914ff4f2bfb8eb85a95eecdfe15f9a5d80eb1ef2"} Jan 26 16:53:51 crc kubenswrapper[4754]: I0126 16:53:51.171928 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-l482d" event={"ID":"833110b5-684c-4097-9e52-d3d8d1838e73","Type":"ContainerStarted","Data":"9082e167aba36fe6b7a850694a1fb9c6d2a872460309547b98fd483c3424871f"} Jan 26 16:53:51 crc kubenswrapper[4754]: I0126 16:53:51.173412 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-c44vn" event={"ID":"02fabee4-adea-4bca-ba0a-e6b98f6e68dd","Type":"ContainerStarted","Data":"1cf0da491e537b59f07af6a6486ce12a1ab2151d7ab3afe00d3ccef630fb0534"} Jan 26 16:53:51 crc kubenswrapper[4754]: I0126 16:53:51.175862 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dn77h" event={"ID":"2dfa533f-5161-4ef0-bc9b-397abae75b23","Type":"ContainerStarted","Data":"2ec437b0b6a31d20450ca21a563883baf142e3eb3c4949f5e2478e0443aef7f1"} Jan 26 16:53:51 crc kubenswrapper[4754]: I0126 16:53:51.200196 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-l482d" podStartSLOduration=2.766763017 podStartE2EDuration="5.200180075s" podCreationTimestamp="2026-01-26 16:53:46 +0000 UTC" firstStartedPulling="2026-01-26 16:53:48.119169031 +0000 UTC m=+394.643349465" lastFinishedPulling="2026-01-26 16:53:50.552586089 +0000 UTC m=+397.076766523" observedRunningTime="2026-01-26 16:53:51.197843737 +0000 UTC m=+397.722024181" watchObservedRunningTime="2026-01-26 16:53:51.200180075 +0000 UTC m=+397.724360509" Jan 26 16:53:51 crc kubenswrapper[4754]: I0126 16:53:51.241014 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-c44vn" podStartSLOduration=2.643403421 podStartE2EDuration="5.240992086s" podCreationTimestamp="2026-01-26 16:53:46 +0000 UTC" firstStartedPulling="2026-01-26 16:53:48.121203972 +0000 UTC m=+394.645384416" lastFinishedPulling="2026-01-26 16:53:50.718792657 +0000 UTC m=+397.242973081" observedRunningTime="2026-01-26 16:53:51.237447439 +0000 UTC m=+397.761627893" watchObservedRunningTime="2026-01-26 16:53:51.240992086 +0000 UTC m=+397.765172520" Jan 26 16:53:52 crc kubenswrapper[4754]: I0126 16:53:52.184410 4754 generic.go:334] "Generic (PLEG): container finished" podID="231baab8-113c-494b-b5d3-b169370901bf" containerID="173fad659b5a84e1d302d02094b68c7801eaf468a6f6dbab00f625bca08b57db" exitCode=0 Jan 26 16:53:52 crc kubenswrapper[4754]: I0126 16:53:52.184705 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xc89f" event={"ID":"231baab8-113c-494b-b5d3-b169370901bf","Type":"ContainerDied","Data":"173fad659b5a84e1d302d02094b68c7801eaf468a6f6dbab00f625bca08b57db"} Jan 26 16:53:52 crc kubenswrapper[4754]: I0126 16:53:52.188590 4754 generic.go:334] "Generic (PLEG): container finished" podID="2dfa533f-5161-4ef0-bc9b-397abae75b23" containerID="2ec437b0b6a31d20450ca21a563883baf142e3eb3c4949f5e2478e0443aef7f1" exitCode=0 Jan 26 16:53:52 crc kubenswrapper[4754]: I0126 16:53:52.188969 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dn77h" event={"ID":"2dfa533f-5161-4ef0-bc9b-397abae75b23","Type":"ContainerDied","Data":"2ec437b0b6a31d20450ca21a563883baf142e3eb3c4949f5e2478e0443aef7f1"} Jan 26 16:53:53 crc kubenswrapper[4754]: I0126 16:53:53.204950 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xc89f" event={"ID":"231baab8-113c-494b-b5d3-b169370901bf","Type":"ContainerStarted","Data":"4cb9df5ec9b0b2595514b3cc038c99ed9bf3e913fa175e54d7db94646d7f9fa8"} Jan 26 16:53:53 crc kubenswrapper[4754]: I0126 16:53:53.206992 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dn77h" event={"ID":"2dfa533f-5161-4ef0-bc9b-397abae75b23","Type":"ContainerStarted","Data":"c07ff8eb15e512b80ca083542667b4fef4ab31ee43c4ac2611b5a117628981bc"} Jan 26 16:53:53 crc kubenswrapper[4754]: I0126 16:53:53.227914 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-xc89f" podStartSLOduration=2.637705748 podStartE2EDuration="5.22789597s" podCreationTimestamp="2026-01-26 16:53:48 +0000 UTC" firstStartedPulling="2026-01-26 16:53:50.152806013 +0000 UTC m=+396.676986447" lastFinishedPulling="2026-01-26 16:53:52.742996235 +0000 UTC m=+399.267176669" observedRunningTime="2026-01-26 16:53:53.224408633 +0000 UTC m=+399.748589087" watchObservedRunningTime="2026-01-26 16:53:53.22789597 +0000 UTC m=+399.752076404" Jan 26 16:53:53 crc kubenswrapper[4754]: I0126 16:53:53.248183 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-dn77h" podStartSLOduration=2.732978469 podStartE2EDuration="5.248148222s" podCreationTimestamp="2026-01-26 16:53:48 +0000 UTC" firstStartedPulling="2026-01-26 16:53:50.157254992 +0000 UTC m=+396.681435416" lastFinishedPulling="2026-01-26 16:53:52.672424735 +0000 UTC m=+399.196605169" observedRunningTime="2026-01-26 16:53:53.243062785 +0000 UTC m=+399.767243219" watchObservedRunningTime="2026-01-26 16:53:53.248148222 +0000 UTC m=+399.772328656" Jan 26 16:53:56 crc kubenswrapper[4754]: I0126 16:53:56.660516 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-l482d" Jan 26 16:53:56 crc kubenswrapper[4754]: I0126 16:53:56.661062 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-l482d" Jan 26 16:53:56 crc kubenswrapper[4754]: I0126 16:53:56.706830 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-l482d" Jan 26 16:53:56 crc kubenswrapper[4754]: I0126 16:53:56.842772 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-c44vn" Jan 26 16:53:56 crc kubenswrapper[4754]: I0126 16:53:56.842827 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-c44vn" Jan 26 16:53:56 crc kubenswrapper[4754]: I0126 16:53:56.878291 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-c44vn" Jan 26 16:53:57 crc kubenswrapper[4754]: I0126 16:53:57.272133 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-l482d" Jan 26 16:53:57 crc kubenswrapper[4754]: I0126 16:53:57.272899 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-c44vn" Jan 26 16:53:59 crc kubenswrapper[4754]: I0126 16:53:59.040511 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-xc89f" Jan 26 16:53:59 crc kubenswrapper[4754]: I0126 16:53:59.040571 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-xc89f" Jan 26 16:53:59 crc kubenswrapper[4754]: I0126 16:53:59.082527 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-xc89f" Jan 26 16:53:59 crc kubenswrapper[4754]: I0126 16:53:59.237914 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-dn77h" Jan 26 16:53:59 crc kubenswrapper[4754]: I0126 16:53:59.238013 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-dn77h" Jan 26 16:53:59 crc kubenswrapper[4754]: I0126 16:53:59.275985 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-dn77h" Jan 26 16:53:59 crc kubenswrapper[4754]: I0126 16:53:59.282475 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-xc89f" Jan 26 16:53:59 crc kubenswrapper[4754]: I0126 16:53:59.332130 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-dn77h" Jan 26 16:54:06 crc kubenswrapper[4754]: I0126 16:54:06.794479 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-8x46q" Jan 26 16:54:06 crc kubenswrapper[4754]: I0126 16:54:06.862930 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-47pnn"] Jan 26 16:54:07 crc kubenswrapper[4754]: I0126 16:54:07.130104 4754 patch_prober.go:28] interesting pod/machine-config-daemon-x65wv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 26 16:54:07 crc kubenswrapper[4754]: I0126 16:54:07.130165 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 26 16:54:07 crc kubenswrapper[4754]: I0126 16:54:07.130217 4754 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" Jan 26 16:54:07 crc kubenswrapper[4754]: I0126 16:54:07.130763 4754 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e4aaaab9b6e326f553465f494a4b68086742f0af0a269e13ef1c99f9395124a5"} pod="openshift-machine-config-operator/machine-config-daemon-x65wv" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 26 16:54:07 crc kubenswrapper[4754]: I0126 16:54:07.130821 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" containerName="machine-config-daemon" containerID="cri-o://e4aaaab9b6e326f553465f494a4b68086742f0af0a269e13ef1c99f9395124a5" gracePeriod=600 Jan 26 16:54:10 crc kubenswrapper[4754]: I0126 16:54:10.398442 4754 generic.go:334] "Generic (PLEG): container finished" podID="8c3718a4-f354-4284-92e0-fdfb45a692bd" containerID="e4aaaab9b6e326f553465f494a4b68086742f0af0a269e13ef1c99f9395124a5" exitCode=0 Jan 26 16:54:10 crc kubenswrapper[4754]: I0126 16:54:10.398558 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" event={"ID":"8c3718a4-f354-4284-92e0-fdfb45a692bd","Type":"ContainerDied","Data":"e4aaaab9b6e326f553465f494a4b68086742f0af0a269e13ef1c99f9395124a5"} Jan 26 16:54:10 crc kubenswrapper[4754]: I0126 16:54:10.398820 4754 scope.go:117] "RemoveContainer" containerID="0fffc03b536bfee3c7bc1bddd500e834e34b391896477a1501e3e06201a19374" Jan 26 16:54:11 crc kubenswrapper[4754]: I0126 16:54:11.406452 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" event={"ID":"8c3718a4-f354-4284-92e0-fdfb45a692bd","Type":"ContainerStarted","Data":"7ba829a1fdd77f6bde538408ab51fc99c1f39952e719ae1ed5b880b92ca7240c"} Jan 26 16:54:13 crc kubenswrapper[4754]: I0126 16:54:13.845784 4754 scope.go:117] "RemoveContainer" containerID="22ba7649a1308815bfad52016000987a7a9a5bc214d7ff61e137ba099b20fc6d" Jan 26 16:54:13 crc kubenswrapper[4754]: I0126 16:54:13.873263 4754 scope.go:117] "RemoveContainer" containerID="27b74e2e65a547e0e28fddd4f15feacb17f4d57596eeb8da82df9d39243c85e0" Jan 26 16:54:13 crc kubenswrapper[4754]: I0126 16:54:13.888585 4754 scope.go:117] "RemoveContainer" containerID="4f45445cbf1d180d6d4dd9588e85bf678f043bfce8cff0812b78ca3510904f9a" Jan 26 16:54:13 crc kubenswrapper[4754]: I0126 16:54:13.903794 4754 scope.go:117] "RemoveContainer" containerID="b9c5f40188fda14b14128514fe84bddfe8dcede3d6c560283ccb2ac0c66895fc" Jan 26 16:54:13 crc kubenswrapper[4754]: I0126 16:54:13.922212 4754 scope.go:117] "RemoveContainer" containerID="8c40e018a0979035827963efb9efc48a878a55b9df68ec1a9242d7cff64fe480" Jan 26 16:54:13 crc kubenswrapper[4754]: I0126 16:54:13.935888 4754 scope.go:117] "RemoveContainer" containerID="085fa7b8d4fef8b2722fac8c346107699b5a6bd33fbfc74ed35831cb8fdf9dad" Jan 26 16:54:31 crc kubenswrapper[4754]: I0126 16:54:31.902942 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-47pnn" podUID="d098371a-9920-44d2-b918-f21c8142fac2" containerName="registry" containerID="cri-o://d8f5c5806d081e327104a1e858d04ead57c27dca4287c0e64499580aaaf7f79e" gracePeriod=30 Jan 26 16:54:32 crc kubenswrapper[4754]: I0126 16:54:32.255552 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-47pnn" Jan 26 16:54:32 crc kubenswrapper[4754]: I0126 16:54:32.310499 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/d098371a-9920-44d2-b918-f21c8142fac2-ca-trust-extracted\") pod \"d098371a-9920-44d2-b918-f21c8142fac2\" (UID: \"d098371a-9920-44d2-b918-f21c8142fac2\") " Jan 26 16:54:32 crc kubenswrapper[4754]: I0126 16:54:32.310568 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d098371a-9920-44d2-b918-f21c8142fac2-bound-sa-token\") pod \"d098371a-9920-44d2-b918-f21c8142fac2\" (UID: \"d098371a-9920-44d2-b918-f21c8142fac2\") " Jan 26 16:54:32 crc kubenswrapper[4754]: I0126 16:54:32.310594 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/d098371a-9920-44d2-b918-f21c8142fac2-installation-pull-secrets\") pod \"d098371a-9920-44d2-b918-f21c8142fac2\" (UID: \"d098371a-9920-44d2-b918-f21c8142fac2\") " Jan 26 16:54:32 crc kubenswrapper[4754]: I0126 16:54:32.310828 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"d098371a-9920-44d2-b918-f21c8142fac2\" (UID: \"d098371a-9920-44d2-b918-f21c8142fac2\") " Jan 26 16:54:32 crc kubenswrapper[4754]: I0126 16:54:32.310854 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pmc9d\" (UniqueName: \"kubernetes.io/projected/d098371a-9920-44d2-b918-f21c8142fac2-kube-api-access-pmc9d\") pod \"d098371a-9920-44d2-b918-f21c8142fac2\" (UID: \"d098371a-9920-44d2-b918-f21c8142fac2\") " Jan 26 16:54:32 crc kubenswrapper[4754]: I0126 16:54:32.310884 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d098371a-9920-44d2-b918-f21c8142fac2-trusted-ca\") pod \"d098371a-9920-44d2-b918-f21c8142fac2\" (UID: \"d098371a-9920-44d2-b918-f21c8142fac2\") " Jan 26 16:54:32 crc kubenswrapper[4754]: I0126 16:54:32.310904 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/d098371a-9920-44d2-b918-f21c8142fac2-registry-tls\") pod \"d098371a-9920-44d2-b918-f21c8142fac2\" (UID: \"d098371a-9920-44d2-b918-f21c8142fac2\") " Jan 26 16:54:32 crc kubenswrapper[4754]: I0126 16:54:32.310924 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/d098371a-9920-44d2-b918-f21c8142fac2-registry-certificates\") pod \"d098371a-9920-44d2-b918-f21c8142fac2\" (UID: \"d098371a-9920-44d2-b918-f21c8142fac2\") " Jan 26 16:54:32 crc kubenswrapper[4754]: I0126 16:54:32.311651 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d098371a-9920-44d2-b918-f21c8142fac2-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "d098371a-9920-44d2-b918-f21c8142fac2" (UID: "d098371a-9920-44d2-b918-f21c8142fac2"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 16:54:32 crc kubenswrapper[4754]: I0126 16:54:32.311680 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d098371a-9920-44d2-b918-f21c8142fac2-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "d098371a-9920-44d2-b918-f21c8142fac2" (UID: "d098371a-9920-44d2-b918-f21c8142fac2"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 16:54:32 crc kubenswrapper[4754]: I0126 16:54:32.325128 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d098371a-9920-44d2-b918-f21c8142fac2-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "d098371a-9920-44d2-b918-f21c8142fac2" (UID: "d098371a-9920-44d2-b918-f21c8142fac2"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:54:32 crc kubenswrapper[4754]: I0126 16:54:32.325506 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d098371a-9920-44d2-b918-f21c8142fac2-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "d098371a-9920-44d2-b918-f21c8142fac2" (UID: "d098371a-9920-44d2-b918-f21c8142fac2"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:54:32 crc kubenswrapper[4754]: I0126 16:54:32.326745 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d098371a-9920-44d2-b918-f21c8142fac2-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "d098371a-9920-44d2-b918-f21c8142fac2" (UID: "d098371a-9920-44d2-b918-f21c8142fac2"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 16:54:32 crc kubenswrapper[4754]: I0126 16:54:32.327872 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d098371a-9920-44d2-b918-f21c8142fac2-kube-api-access-pmc9d" (OuterVolumeSpecName: "kube-api-access-pmc9d") pod "d098371a-9920-44d2-b918-f21c8142fac2" (UID: "d098371a-9920-44d2-b918-f21c8142fac2"). InnerVolumeSpecName "kube-api-access-pmc9d". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:54:32 crc kubenswrapper[4754]: I0126 16:54:32.332049 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d098371a-9920-44d2-b918-f21c8142fac2-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "d098371a-9920-44d2-b918-f21c8142fac2" (UID: "d098371a-9920-44d2-b918-f21c8142fac2"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 16:54:32 crc kubenswrapper[4754]: I0126 16:54:32.332473 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "d098371a-9920-44d2-b918-f21c8142fac2" (UID: "d098371a-9920-44d2-b918-f21c8142fac2"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Jan 26 16:54:32 crc kubenswrapper[4754]: I0126 16:54:32.411826 4754 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/d098371a-9920-44d2-b918-f21c8142fac2-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Jan 26 16:54:32 crc kubenswrapper[4754]: I0126 16:54:32.411860 4754 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d098371a-9920-44d2-b918-f21c8142fac2-bound-sa-token\") on node \"crc\" DevicePath \"\"" Jan 26 16:54:32 crc kubenswrapper[4754]: I0126 16:54:32.411870 4754 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/d098371a-9920-44d2-b918-f21c8142fac2-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Jan 26 16:54:32 crc kubenswrapper[4754]: I0126 16:54:32.411883 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pmc9d\" (UniqueName: \"kubernetes.io/projected/d098371a-9920-44d2-b918-f21c8142fac2-kube-api-access-pmc9d\") on node \"crc\" DevicePath \"\"" Jan 26 16:54:32 crc kubenswrapper[4754]: I0126 16:54:32.411892 4754 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d098371a-9920-44d2-b918-f21c8142fac2-trusted-ca\") on node \"crc\" DevicePath \"\"" Jan 26 16:54:32 crc kubenswrapper[4754]: I0126 16:54:32.411901 4754 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/d098371a-9920-44d2-b918-f21c8142fac2-registry-tls\") on node \"crc\" DevicePath \"\"" Jan 26 16:54:32 crc kubenswrapper[4754]: I0126 16:54:32.411910 4754 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/d098371a-9920-44d2-b918-f21c8142fac2-registry-certificates\") on node \"crc\" DevicePath \"\"" Jan 26 16:54:32 crc kubenswrapper[4754]: I0126 16:54:32.522109 4754 generic.go:334] "Generic (PLEG): container finished" podID="d098371a-9920-44d2-b918-f21c8142fac2" containerID="d8f5c5806d081e327104a1e858d04ead57c27dca4287c0e64499580aaaf7f79e" exitCode=0 Jan 26 16:54:32 crc kubenswrapper[4754]: I0126 16:54:32.522153 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-47pnn" event={"ID":"d098371a-9920-44d2-b918-f21c8142fac2","Type":"ContainerDied","Data":"d8f5c5806d081e327104a1e858d04ead57c27dca4287c0e64499580aaaf7f79e"} Jan 26 16:54:32 crc kubenswrapper[4754]: I0126 16:54:32.522182 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-47pnn" event={"ID":"d098371a-9920-44d2-b918-f21c8142fac2","Type":"ContainerDied","Data":"537130dd816455c11cb6fc9665a7a41fe83c79a3a56347ddf6eac6e8f83da252"} Jan 26 16:54:32 crc kubenswrapper[4754]: I0126 16:54:32.522190 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-47pnn" Jan 26 16:54:32 crc kubenswrapper[4754]: I0126 16:54:32.522213 4754 scope.go:117] "RemoveContainer" containerID="d8f5c5806d081e327104a1e858d04ead57c27dca4287c0e64499580aaaf7f79e" Jan 26 16:54:32 crc kubenswrapper[4754]: I0126 16:54:32.538392 4754 scope.go:117] "RemoveContainer" containerID="d8f5c5806d081e327104a1e858d04ead57c27dca4287c0e64499580aaaf7f79e" Jan 26 16:54:32 crc kubenswrapper[4754]: E0126 16:54:32.539006 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d8f5c5806d081e327104a1e858d04ead57c27dca4287c0e64499580aaaf7f79e\": container with ID starting with d8f5c5806d081e327104a1e858d04ead57c27dca4287c0e64499580aaaf7f79e not found: ID does not exist" containerID="d8f5c5806d081e327104a1e858d04ead57c27dca4287c0e64499580aaaf7f79e" Jan 26 16:54:32 crc kubenswrapper[4754]: I0126 16:54:32.539065 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d8f5c5806d081e327104a1e858d04ead57c27dca4287c0e64499580aaaf7f79e"} err="failed to get container status \"d8f5c5806d081e327104a1e858d04ead57c27dca4287c0e64499580aaaf7f79e\": rpc error: code = NotFound desc = could not find container \"d8f5c5806d081e327104a1e858d04ead57c27dca4287c0e64499580aaaf7f79e\": container with ID starting with d8f5c5806d081e327104a1e858d04ead57c27dca4287c0e64499580aaaf7f79e not found: ID does not exist" Jan 26 16:54:32 crc kubenswrapper[4754]: I0126 16:54:32.551833 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-47pnn"] Jan 26 16:54:32 crc kubenswrapper[4754]: I0126 16:54:32.556171 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-47pnn"] Jan 26 16:54:33 crc kubenswrapper[4754]: I0126 16:54:33.776448 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d098371a-9920-44d2-b918-f21c8142fac2" path="/var/lib/kubelet/pods/d098371a-9920-44d2-b918-f21c8142fac2/volumes" Jan 26 16:56:37 crc kubenswrapper[4754]: I0126 16:56:37.129651 4754 patch_prober.go:28] interesting pod/machine-config-daemon-x65wv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 26 16:56:37 crc kubenswrapper[4754]: I0126 16:56:37.130308 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 26 16:57:07 crc kubenswrapper[4754]: I0126 16:57:07.129378 4754 patch_prober.go:28] interesting pod/machine-config-daemon-x65wv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 26 16:57:07 crc kubenswrapper[4754]: I0126 16:57:07.131002 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 26 16:57:37 crc kubenswrapper[4754]: I0126 16:57:37.129157 4754 patch_prober.go:28] interesting pod/machine-config-daemon-x65wv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 26 16:57:37 crc kubenswrapper[4754]: I0126 16:57:37.130157 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 26 16:57:37 crc kubenswrapper[4754]: I0126 16:57:37.130228 4754 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" Jan 26 16:57:37 crc kubenswrapper[4754]: I0126 16:57:37.131021 4754 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"7ba829a1fdd77f6bde538408ab51fc99c1f39952e719ae1ed5b880b92ca7240c"} pod="openshift-machine-config-operator/machine-config-daemon-x65wv" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 26 16:57:37 crc kubenswrapper[4754]: I0126 16:57:37.131099 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" containerName="machine-config-daemon" containerID="cri-o://7ba829a1fdd77f6bde538408ab51fc99c1f39952e719ae1ed5b880b92ca7240c" gracePeriod=600 Jan 26 16:57:37 crc kubenswrapper[4754]: I0126 16:57:37.530361 4754 generic.go:334] "Generic (PLEG): container finished" podID="8c3718a4-f354-4284-92e0-fdfb45a692bd" containerID="7ba829a1fdd77f6bde538408ab51fc99c1f39952e719ae1ed5b880b92ca7240c" exitCode=0 Jan 26 16:57:37 crc kubenswrapper[4754]: I0126 16:57:37.530431 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" event={"ID":"8c3718a4-f354-4284-92e0-fdfb45a692bd","Type":"ContainerDied","Data":"7ba829a1fdd77f6bde538408ab51fc99c1f39952e719ae1ed5b880b92ca7240c"} Jan 26 16:57:37 crc kubenswrapper[4754]: I0126 16:57:37.530851 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" event={"ID":"8c3718a4-f354-4284-92e0-fdfb45a692bd","Type":"ContainerStarted","Data":"2258d22dac1a41453a2ed61fe65a0c8edeaa3d706e7abb2b8003f3eba6c44d91"} Jan 26 16:57:37 crc kubenswrapper[4754]: I0126 16:57:37.530873 4754 scope.go:117] "RemoveContainer" containerID="e4aaaab9b6e326f553465f494a4b68086742f0af0a269e13ef1c99f9395124a5" Jan 26 16:59:14 crc kubenswrapper[4754]: I0126 16:59:14.065703 4754 scope.go:117] "RemoveContainer" containerID="002543359d3f5099129099dc09e27d79540d55011c076232de8ebe011a200767" Jan 26 16:59:37 crc kubenswrapper[4754]: I0126 16:59:37.129039 4754 patch_prober.go:28] interesting pod/machine-config-daemon-x65wv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 26 16:59:37 crc kubenswrapper[4754]: I0126 16:59:37.129529 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 26 16:59:41 crc kubenswrapper[4754]: I0126 16:59:41.653441 4754 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Jan 26 16:59:49 crc kubenswrapper[4754]: I0126 16:59:49.549747 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-cf98fcc89-mksh5"] Jan 26 16:59:49 crc kubenswrapper[4754]: E0126 16:59:49.550346 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d098371a-9920-44d2-b918-f21c8142fac2" containerName="registry" Jan 26 16:59:49 crc kubenswrapper[4754]: I0126 16:59:49.550358 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="d098371a-9920-44d2-b918-f21c8142fac2" containerName="registry" Jan 26 16:59:49 crc kubenswrapper[4754]: I0126 16:59:49.550463 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="d098371a-9920-44d2-b918-f21c8142fac2" containerName="registry" Jan 26 16:59:49 crc kubenswrapper[4754]: I0126 16:59:49.550852 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-cf98fcc89-mksh5" Jan 26 16:59:49 crc kubenswrapper[4754]: I0126 16:59:49.555549 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Jan 26 16:59:49 crc kubenswrapper[4754]: I0126 16:59:49.555872 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Jan 26 16:59:49 crc kubenswrapper[4754]: I0126 16:59:49.556082 4754 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-9lnmt" Jan 26 16:59:49 crc kubenswrapper[4754]: I0126 16:59:49.574158 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-858654f9db-5n7pf"] Jan 26 16:59:49 crc kubenswrapper[4754]: I0126 16:59:49.574948 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-858654f9db-5n7pf" Jan 26 16:59:49 crc kubenswrapper[4754]: I0126 16:59:49.579103 4754 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-2f76l" Jan 26 16:59:49 crc kubenswrapper[4754]: I0126 16:59:49.579542 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-cf98fcc89-mksh5"] Jan 26 16:59:49 crc kubenswrapper[4754]: I0126 16:59:49.592053 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-687f57d79b-q26n7"] Jan 26 16:59:49 crc kubenswrapper[4754]: I0126 16:59:49.593071 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-687f57d79b-q26n7" Jan 26 16:59:49 crc kubenswrapper[4754]: I0126 16:59:49.594676 4754 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-vxjzc" Jan 26 16:59:49 crc kubenswrapper[4754]: I0126 16:59:49.602783 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-858654f9db-5n7pf"] Jan 26 16:59:49 crc kubenswrapper[4754]: I0126 16:59:49.605341 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-687f57d79b-q26n7"] Jan 26 16:59:49 crc kubenswrapper[4754]: I0126 16:59:49.670687 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sdvqk\" (UniqueName: \"kubernetes.io/projected/9bddf956-836c-41ef-9038-f889e71b6823-kube-api-access-sdvqk\") pod \"cert-manager-cainjector-cf98fcc89-mksh5\" (UID: \"9bddf956-836c-41ef-9038-f889e71b6823\") " pod="cert-manager/cert-manager-cainjector-cf98fcc89-mksh5" Jan 26 16:59:49 crc kubenswrapper[4754]: I0126 16:59:49.772072 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sd696\" (UniqueName: \"kubernetes.io/projected/b2d0aba8-8dab-483d-87cd-756af280d526-kube-api-access-sd696\") pod \"cert-manager-858654f9db-5n7pf\" (UID: \"b2d0aba8-8dab-483d-87cd-756af280d526\") " pod="cert-manager/cert-manager-858654f9db-5n7pf" Jan 26 16:59:49 crc kubenswrapper[4754]: I0126 16:59:49.772125 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sdvqk\" (UniqueName: \"kubernetes.io/projected/9bddf956-836c-41ef-9038-f889e71b6823-kube-api-access-sdvqk\") pod \"cert-manager-cainjector-cf98fcc89-mksh5\" (UID: \"9bddf956-836c-41ef-9038-f889e71b6823\") " pod="cert-manager/cert-manager-cainjector-cf98fcc89-mksh5" Jan 26 16:59:49 crc kubenswrapper[4754]: I0126 16:59:49.772823 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vxgzr\" (UniqueName: \"kubernetes.io/projected/d46f02a6-da6f-4128-89e8-669cdb8622c6-kube-api-access-vxgzr\") pod \"cert-manager-webhook-687f57d79b-q26n7\" (UID: \"d46f02a6-da6f-4128-89e8-669cdb8622c6\") " pod="cert-manager/cert-manager-webhook-687f57d79b-q26n7" Jan 26 16:59:49 crc kubenswrapper[4754]: I0126 16:59:49.796828 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sdvqk\" (UniqueName: \"kubernetes.io/projected/9bddf956-836c-41ef-9038-f889e71b6823-kube-api-access-sdvqk\") pod \"cert-manager-cainjector-cf98fcc89-mksh5\" (UID: \"9bddf956-836c-41ef-9038-f889e71b6823\") " pod="cert-manager/cert-manager-cainjector-cf98fcc89-mksh5" Jan 26 16:59:49 crc kubenswrapper[4754]: I0126 16:59:49.874045 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vxgzr\" (UniqueName: \"kubernetes.io/projected/d46f02a6-da6f-4128-89e8-669cdb8622c6-kube-api-access-vxgzr\") pod \"cert-manager-webhook-687f57d79b-q26n7\" (UID: \"d46f02a6-da6f-4128-89e8-669cdb8622c6\") " pod="cert-manager/cert-manager-webhook-687f57d79b-q26n7" Jan 26 16:59:49 crc kubenswrapper[4754]: I0126 16:59:49.874119 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sd696\" (UniqueName: \"kubernetes.io/projected/b2d0aba8-8dab-483d-87cd-756af280d526-kube-api-access-sd696\") pod \"cert-manager-858654f9db-5n7pf\" (UID: \"b2d0aba8-8dab-483d-87cd-756af280d526\") " pod="cert-manager/cert-manager-858654f9db-5n7pf" Jan 26 16:59:49 crc kubenswrapper[4754]: I0126 16:59:49.880886 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-cf98fcc89-mksh5" Jan 26 16:59:49 crc kubenswrapper[4754]: I0126 16:59:49.894108 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sd696\" (UniqueName: \"kubernetes.io/projected/b2d0aba8-8dab-483d-87cd-756af280d526-kube-api-access-sd696\") pod \"cert-manager-858654f9db-5n7pf\" (UID: \"b2d0aba8-8dab-483d-87cd-756af280d526\") " pod="cert-manager/cert-manager-858654f9db-5n7pf" Jan 26 16:59:49 crc kubenswrapper[4754]: I0126 16:59:49.895582 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vxgzr\" (UniqueName: \"kubernetes.io/projected/d46f02a6-da6f-4128-89e8-669cdb8622c6-kube-api-access-vxgzr\") pod \"cert-manager-webhook-687f57d79b-q26n7\" (UID: \"d46f02a6-da6f-4128-89e8-669cdb8622c6\") " pod="cert-manager/cert-manager-webhook-687f57d79b-q26n7" Jan 26 16:59:49 crc kubenswrapper[4754]: I0126 16:59:49.905929 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-687f57d79b-q26n7" Jan 26 16:59:50 crc kubenswrapper[4754]: I0126 16:59:50.116542 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-687f57d79b-q26n7"] Jan 26 16:59:50 crc kubenswrapper[4754]: I0126 16:59:50.122797 4754 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 26 16:59:50 crc kubenswrapper[4754]: I0126 16:59:50.192376 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-858654f9db-5n7pf" Jan 26 16:59:50 crc kubenswrapper[4754]: I0126 16:59:50.303326 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-cf98fcc89-mksh5"] Jan 26 16:59:50 crc kubenswrapper[4754]: W0126 16:59:50.307444 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9bddf956_836c_41ef_9038_f889e71b6823.slice/crio-b92a9eb1f61dbb7e1d887ea51a32a2bb00b028b18f125955f27e532e15608f91 WatchSource:0}: Error finding container b92a9eb1f61dbb7e1d887ea51a32a2bb00b028b18f125955f27e532e15608f91: Status 404 returned error can't find the container with id b92a9eb1f61dbb7e1d887ea51a32a2bb00b028b18f125955f27e532e15608f91 Jan 26 16:59:50 crc kubenswrapper[4754]: I0126 16:59:50.317636 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-687f57d79b-q26n7" event={"ID":"d46f02a6-da6f-4128-89e8-669cdb8622c6","Type":"ContainerStarted","Data":"3530606d58fd9106104383dc9acece675dd07db97dff044c32dfd6c6d10bcd87"} Jan 26 16:59:50 crc kubenswrapper[4754]: I0126 16:59:50.403457 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-858654f9db-5n7pf"] Jan 26 16:59:50 crc kubenswrapper[4754]: W0126 16:59:50.410273 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb2d0aba8_8dab_483d_87cd_756af280d526.slice/crio-5bb3ae0a392c587da4a88826c3d144717e844fbd8f8267c5ab6e9bb836150665 WatchSource:0}: Error finding container 5bb3ae0a392c587da4a88826c3d144717e844fbd8f8267c5ab6e9bb836150665: Status 404 returned error can't find the container with id 5bb3ae0a392c587da4a88826c3d144717e844fbd8f8267c5ab6e9bb836150665 Jan 26 16:59:50 crc kubenswrapper[4754]: I0126 16:59:50.862876 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-qln6t"] Jan 26 16:59:50 crc kubenswrapper[4754]: I0126 16:59:50.864436 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qln6t" Jan 26 16:59:50 crc kubenswrapper[4754]: I0126 16:59:50.869173 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-qln6t"] Jan 26 16:59:50 crc kubenswrapper[4754]: I0126 16:59:50.887558 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3927f5f7-ced1-41f1-8e55-ae354f995a78-utilities\") pod \"certified-operators-qln6t\" (UID: \"3927f5f7-ced1-41f1-8e55-ae354f995a78\") " pod="openshift-marketplace/certified-operators-qln6t" Jan 26 16:59:50 crc kubenswrapper[4754]: I0126 16:59:50.887683 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pff5t\" (UniqueName: \"kubernetes.io/projected/3927f5f7-ced1-41f1-8e55-ae354f995a78-kube-api-access-pff5t\") pod \"certified-operators-qln6t\" (UID: \"3927f5f7-ced1-41f1-8e55-ae354f995a78\") " pod="openshift-marketplace/certified-operators-qln6t" Jan 26 16:59:50 crc kubenswrapper[4754]: I0126 16:59:50.887720 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3927f5f7-ced1-41f1-8e55-ae354f995a78-catalog-content\") pod \"certified-operators-qln6t\" (UID: \"3927f5f7-ced1-41f1-8e55-ae354f995a78\") " pod="openshift-marketplace/certified-operators-qln6t" Jan 26 16:59:50 crc kubenswrapper[4754]: I0126 16:59:50.988479 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pff5t\" (UniqueName: \"kubernetes.io/projected/3927f5f7-ced1-41f1-8e55-ae354f995a78-kube-api-access-pff5t\") pod \"certified-operators-qln6t\" (UID: \"3927f5f7-ced1-41f1-8e55-ae354f995a78\") " pod="openshift-marketplace/certified-operators-qln6t" Jan 26 16:59:50 crc kubenswrapper[4754]: I0126 16:59:50.988533 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3927f5f7-ced1-41f1-8e55-ae354f995a78-catalog-content\") pod \"certified-operators-qln6t\" (UID: \"3927f5f7-ced1-41f1-8e55-ae354f995a78\") " pod="openshift-marketplace/certified-operators-qln6t" Jan 26 16:59:50 crc kubenswrapper[4754]: I0126 16:59:50.988594 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3927f5f7-ced1-41f1-8e55-ae354f995a78-utilities\") pod \"certified-operators-qln6t\" (UID: \"3927f5f7-ced1-41f1-8e55-ae354f995a78\") " pod="openshift-marketplace/certified-operators-qln6t" Jan 26 16:59:50 crc kubenswrapper[4754]: I0126 16:59:50.989076 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3927f5f7-ced1-41f1-8e55-ae354f995a78-utilities\") pod \"certified-operators-qln6t\" (UID: \"3927f5f7-ced1-41f1-8e55-ae354f995a78\") " pod="openshift-marketplace/certified-operators-qln6t" Jan 26 16:59:50 crc kubenswrapper[4754]: I0126 16:59:50.989145 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3927f5f7-ced1-41f1-8e55-ae354f995a78-catalog-content\") pod \"certified-operators-qln6t\" (UID: \"3927f5f7-ced1-41f1-8e55-ae354f995a78\") " pod="openshift-marketplace/certified-operators-qln6t" Jan 26 16:59:51 crc kubenswrapper[4754]: I0126 16:59:51.007479 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pff5t\" (UniqueName: \"kubernetes.io/projected/3927f5f7-ced1-41f1-8e55-ae354f995a78-kube-api-access-pff5t\") pod \"certified-operators-qln6t\" (UID: \"3927f5f7-ced1-41f1-8e55-ae354f995a78\") " pod="openshift-marketplace/certified-operators-qln6t" Jan 26 16:59:51 crc kubenswrapper[4754]: I0126 16:59:51.198491 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qln6t" Jan 26 16:59:51 crc kubenswrapper[4754]: I0126 16:59:51.327164 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-cf98fcc89-mksh5" event={"ID":"9bddf956-836c-41ef-9038-f889e71b6823","Type":"ContainerStarted","Data":"b92a9eb1f61dbb7e1d887ea51a32a2bb00b028b18f125955f27e532e15608f91"} Jan 26 16:59:51 crc kubenswrapper[4754]: I0126 16:59:51.332387 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-858654f9db-5n7pf" event={"ID":"b2d0aba8-8dab-483d-87cd-756af280d526","Type":"ContainerStarted","Data":"5bb3ae0a392c587da4a88826c3d144717e844fbd8f8267c5ab6e9bb836150665"} Jan 26 16:59:51 crc kubenswrapper[4754]: I0126 16:59:51.764250 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-qln6t"] Jan 26 16:59:52 crc kubenswrapper[4754]: I0126 16:59:52.339803 4754 generic.go:334] "Generic (PLEG): container finished" podID="3927f5f7-ced1-41f1-8e55-ae354f995a78" containerID="dcd2c65ef7e38d4880aefd61bcc73c9a7c22231366043739a5f0d92ac4b68325" exitCode=0 Jan 26 16:59:52 crc kubenswrapper[4754]: I0126 16:59:52.339884 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qln6t" event={"ID":"3927f5f7-ced1-41f1-8e55-ae354f995a78","Type":"ContainerDied","Data":"dcd2c65ef7e38d4880aefd61bcc73c9a7c22231366043739a5f0d92ac4b68325"} Jan 26 16:59:52 crc kubenswrapper[4754]: I0126 16:59:52.340357 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qln6t" event={"ID":"3927f5f7-ced1-41f1-8e55-ae354f995a78","Type":"ContainerStarted","Data":"db2ce6e1881136437a85d570238e84e48994a28fd7cb4a9609610ac10e59a19d"} Jan 26 16:59:54 crc kubenswrapper[4754]: I0126 16:59:54.357233 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-cf98fcc89-mksh5" event={"ID":"9bddf956-836c-41ef-9038-f889e71b6823","Type":"ContainerStarted","Data":"2cb52666fde92442cb5266169e8e7638ff07a3df8ad7c5c10affb4e2a83bc042"} Jan 26 16:59:54 crc kubenswrapper[4754]: I0126 16:59:54.359493 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-687f57d79b-q26n7" event={"ID":"d46f02a6-da6f-4128-89e8-669cdb8622c6","Type":"ContainerStarted","Data":"cabe36e40645759468e64ea0776fbb27a85ad14f1bceae25b85fc3940b35770c"} Jan 26 16:59:54 crc kubenswrapper[4754]: I0126 16:59:54.359696 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-687f57d79b-q26n7" Jan 26 16:59:54 crc kubenswrapper[4754]: I0126 16:59:54.362102 4754 generic.go:334] "Generic (PLEG): container finished" podID="3927f5f7-ced1-41f1-8e55-ae354f995a78" containerID="5d53a53d80c48d9e41a7a78baad8344f22cd0c60fbebbb433938bd4bb06088d0" exitCode=0 Jan 26 16:59:54 crc kubenswrapper[4754]: I0126 16:59:54.362206 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qln6t" event={"ID":"3927f5f7-ced1-41f1-8e55-ae354f995a78","Type":"ContainerDied","Data":"5d53a53d80c48d9e41a7a78baad8344f22cd0c60fbebbb433938bd4bb06088d0"} Jan 26 16:59:54 crc kubenswrapper[4754]: I0126 16:59:54.365170 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-858654f9db-5n7pf" event={"ID":"b2d0aba8-8dab-483d-87cd-756af280d526","Type":"ContainerStarted","Data":"3fc194da086cb644d3047f28b6d0613b2619015f37b80cdbf57efb8c9e703ee5"} Jan 26 16:59:54 crc kubenswrapper[4754]: I0126 16:59:54.374320 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-cf98fcc89-mksh5" podStartSLOduration=1.998281253 podStartE2EDuration="5.374302921s" podCreationTimestamp="2026-01-26 16:59:49 +0000 UTC" firstStartedPulling="2026-01-26 16:59:50.310150808 +0000 UTC m=+756.834331242" lastFinishedPulling="2026-01-26 16:59:53.686172446 +0000 UTC m=+760.210352910" observedRunningTime="2026-01-26 16:59:54.370367833 +0000 UTC m=+760.894548277" watchObservedRunningTime="2026-01-26 16:59:54.374302921 +0000 UTC m=+760.898483355" Jan 26 16:59:54 crc kubenswrapper[4754]: I0126 16:59:54.405124 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-858654f9db-5n7pf" podStartSLOduration=2.130905614 podStartE2EDuration="5.405101789s" podCreationTimestamp="2026-01-26 16:59:49 +0000 UTC" firstStartedPulling="2026-01-26 16:59:50.41233452 +0000 UTC m=+756.936514964" lastFinishedPulling="2026-01-26 16:59:53.686530695 +0000 UTC m=+760.210711139" observedRunningTime="2026-01-26 16:59:54.40221179 +0000 UTC m=+760.926392224" watchObservedRunningTime="2026-01-26 16:59:54.405101789 +0000 UTC m=+760.929282233" Jan 26 16:59:54 crc kubenswrapper[4754]: I0126 16:59:54.420436 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-687f57d79b-q26n7" podStartSLOduration=1.906765333 podStartE2EDuration="5.4204129s" podCreationTimestamp="2026-01-26 16:59:49 +0000 UTC" firstStartedPulling="2026-01-26 16:59:50.122500671 +0000 UTC m=+756.646681105" lastFinishedPulling="2026-01-26 16:59:53.636148228 +0000 UTC m=+760.160328672" observedRunningTime="2026-01-26 16:59:54.418623031 +0000 UTC m=+760.942803465" watchObservedRunningTime="2026-01-26 16:59:54.4204129 +0000 UTC m=+760.944593344" Jan 26 16:59:55 crc kubenswrapper[4754]: I0126 16:59:55.374249 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qln6t" event={"ID":"3927f5f7-ced1-41f1-8e55-ae354f995a78","Type":"ContainerStarted","Data":"ab293f35bec44c3b25b47a18935fc86d8b05d717ad8694b188b67cd23e850b56"} Jan 26 16:59:55 crc kubenswrapper[4754]: I0126 16:59:55.402241 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-qln6t" podStartSLOduration=3.017892017 podStartE2EDuration="5.402224372s" podCreationTimestamp="2026-01-26 16:59:50 +0000 UTC" firstStartedPulling="2026-01-26 16:59:52.354940654 +0000 UTC m=+758.879121088" lastFinishedPulling="2026-01-26 16:59:54.739273009 +0000 UTC m=+761.263453443" observedRunningTime="2026-01-26 16:59:55.398744596 +0000 UTC m=+761.922925050" watchObservedRunningTime="2026-01-26 16:59:55.402224372 +0000 UTC m=+761.926404806" Jan 26 16:59:59 crc kubenswrapper[4754]: I0126 16:59:59.108045 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-jsbxt"] Jan 26 16:59:59 crc kubenswrapper[4754]: I0126 16:59:59.109342 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-jsbxt" podUID="689569c4-93ee-4b82-93fc-9af0f97e6dc3" containerName="ovn-controller" containerID="cri-o://3fe4d9395447ac570eca1338e749dbefea76295d427a12487b63a142a7d6a2e4" gracePeriod=30 Jan 26 16:59:59 crc kubenswrapper[4754]: I0126 16:59:59.109383 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-jsbxt" podUID="689569c4-93ee-4b82-93fc-9af0f97e6dc3" containerName="sbdb" containerID="cri-o://07a43ff7860ff65faf51f4a5e8aeb326f06d6d7222677a92560ae2e3a0c81484" gracePeriod=30 Jan 26 16:59:59 crc kubenswrapper[4754]: I0126 16:59:59.109542 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-jsbxt" podUID="689569c4-93ee-4b82-93fc-9af0f97e6dc3" containerName="nbdb" containerID="cri-o://b1ac1f49bd564750fcff530a0f76657b0ed41573539acd020a922d19861ef834" gracePeriod=30 Jan 26 16:59:59 crc kubenswrapper[4754]: I0126 16:59:59.109648 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-jsbxt" podUID="689569c4-93ee-4b82-93fc-9af0f97e6dc3" containerName="northd" containerID="cri-o://a300a80ad87f46ba1b4fd7ad3cf6c25767fcfe966b93bdd86d729843f53a5cda" gracePeriod=30 Jan 26 16:59:59 crc kubenswrapper[4754]: I0126 16:59:59.109800 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-jsbxt" podUID="689569c4-93ee-4b82-93fc-9af0f97e6dc3" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://919572e731813a98c7d8a94879bba931f75c472c7cd681e86d5197ebab43815d" gracePeriod=30 Jan 26 16:59:59 crc kubenswrapper[4754]: I0126 16:59:59.109904 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-jsbxt" podUID="689569c4-93ee-4b82-93fc-9af0f97e6dc3" containerName="kube-rbac-proxy-node" containerID="cri-o://033f8b104d4e9ff6d5f87ca998665f12e0c56696b3d30e01ede05f4da540365d" gracePeriod=30 Jan 26 16:59:59 crc kubenswrapper[4754]: I0126 16:59:59.110030 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-jsbxt" podUID="689569c4-93ee-4b82-93fc-9af0f97e6dc3" containerName="ovn-acl-logging" containerID="cri-o://31dcab2f9a2d218b0e49dad1df46377ff4521592824a0fbac254d5b8e6d0632e" gracePeriod=30 Jan 26 16:59:59 crc kubenswrapper[4754]: I0126 16:59:59.139446 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-jsbxt" podUID="689569c4-93ee-4b82-93fc-9af0f97e6dc3" containerName="ovnkube-controller" containerID="cri-o://0d6f7aa12978cc78e269af0ca6f59f424d8898f692c6f8d07454a1868b0204cd" gracePeriod=30 Jan 26 16:59:59 crc kubenswrapper[4754]: I0126 16:59:59.397442 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-skh5n_b619d233-b592-4b05-a0b7-dc094c88471e/kube-multus/2.log" Jan 26 16:59:59 crc kubenswrapper[4754]: I0126 16:59:59.398490 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-skh5n_b619d233-b592-4b05-a0b7-dc094c88471e/kube-multus/1.log" Jan 26 16:59:59 crc kubenswrapper[4754]: I0126 16:59:59.398560 4754 generic.go:334] "Generic (PLEG): container finished" podID="b619d233-b592-4b05-a0b7-dc094c88471e" containerID="8a1283e9d0714480b093730068bfb9fc6a36c9982a88f87e70ec9fea74bb462c" exitCode=2 Jan 26 16:59:59 crc kubenswrapper[4754]: I0126 16:59:59.398606 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-skh5n" event={"ID":"b619d233-b592-4b05-a0b7-dc094c88471e","Type":"ContainerDied","Data":"8a1283e9d0714480b093730068bfb9fc6a36c9982a88f87e70ec9fea74bb462c"} Jan 26 16:59:59 crc kubenswrapper[4754]: I0126 16:59:59.398701 4754 scope.go:117] "RemoveContainer" containerID="c8d026fdb7ea7dc5c40add18606d8f1a7703865edf93ef19edaf811f598364c5" Jan 26 16:59:59 crc kubenswrapper[4754]: I0126 16:59:59.399121 4754 scope.go:117] "RemoveContainer" containerID="8a1283e9d0714480b093730068bfb9fc6a36c9982a88f87e70ec9fea74bb462c" Jan 26 16:59:59 crc kubenswrapper[4754]: I0126 16:59:59.402949 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-jsbxt_689569c4-93ee-4b82-93fc-9af0f97e6dc3/ovnkube-controller/3.log" Jan 26 16:59:59 crc kubenswrapper[4754]: I0126 16:59:59.406183 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-jsbxt_689569c4-93ee-4b82-93fc-9af0f97e6dc3/ovn-acl-logging/0.log" Jan 26 16:59:59 crc kubenswrapper[4754]: I0126 16:59:59.407009 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-jsbxt_689569c4-93ee-4b82-93fc-9af0f97e6dc3/ovn-controller/0.log" Jan 26 16:59:59 crc kubenswrapper[4754]: I0126 16:59:59.407545 4754 generic.go:334] "Generic (PLEG): container finished" podID="689569c4-93ee-4b82-93fc-9af0f97e6dc3" containerID="0d6f7aa12978cc78e269af0ca6f59f424d8898f692c6f8d07454a1868b0204cd" exitCode=0 Jan 26 16:59:59 crc kubenswrapper[4754]: I0126 16:59:59.407572 4754 generic.go:334] "Generic (PLEG): container finished" podID="689569c4-93ee-4b82-93fc-9af0f97e6dc3" containerID="919572e731813a98c7d8a94879bba931f75c472c7cd681e86d5197ebab43815d" exitCode=0 Jan 26 16:59:59 crc kubenswrapper[4754]: I0126 16:59:59.407583 4754 generic.go:334] "Generic (PLEG): container finished" podID="689569c4-93ee-4b82-93fc-9af0f97e6dc3" containerID="033f8b104d4e9ff6d5f87ca998665f12e0c56696b3d30e01ede05f4da540365d" exitCode=0 Jan 26 16:59:59 crc kubenswrapper[4754]: I0126 16:59:59.407591 4754 generic.go:334] "Generic (PLEG): container finished" podID="689569c4-93ee-4b82-93fc-9af0f97e6dc3" containerID="31dcab2f9a2d218b0e49dad1df46377ff4521592824a0fbac254d5b8e6d0632e" exitCode=143 Jan 26 16:59:59 crc kubenswrapper[4754]: I0126 16:59:59.407600 4754 generic.go:334] "Generic (PLEG): container finished" podID="689569c4-93ee-4b82-93fc-9af0f97e6dc3" containerID="3fe4d9395447ac570eca1338e749dbefea76295d427a12487b63a142a7d6a2e4" exitCode=143 Jan 26 16:59:59 crc kubenswrapper[4754]: I0126 16:59:59.407625 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jsbxt" event={"ID":"689569c4-93ee-4b82-93fc-9af0f97e6dc3","Type":"ContainerDied","Data":"0d6f7aa12978cc78e269af0ca6f59f424d8898f692c6f8d07454a1868b0204cd"} Jan 26 16:59:59 crc kubenswrapper[4754]: I0126 16:59:59.407657 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jsbxt" event={"ID":"689569c4-93ee-4b82-93fc-9af0f97e6dc3","Type":"ContainerDied","Data":"919572e731813a98c7d8a94879bba931f75c472c7cd681e86d5197ebab43815d"} Jan 26 16:59:59 crc kubenswrapper[4754]: I0126 16:59:59.407707 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jsbxt" event={"ID":"689569c4-93ee-4b82-93fc-9af0f97e6dc3","Type":"ContainerDied","Data":"033f8b104d4e9ff6d5f87ca998665f12e0c56696b3d30e01ede05f4da540365d"} Jan 26 16:59:59 crc kubenswrapper[4754]: I0126 16:59:59.407720 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jsbxt" event={"ID":"689569c4-93ee-4b82-93fc-9af0f97e6dc3","Type":"ContainerDied","Data":"31dcab2f9a2d218b0e49dad1df46377ff4521592824a0fbac254d5b8e6d0632e"} Jan 26 16:59:59 crc kubenswrapper[4754]: I0126 16:59:59.407731 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jsbxt" event={"ID":"689569c4-93ee-4b82-93fc-9af0f97e6dc3","Type":"ContainerDied","Data":"3fe4d9395447ac570eca1338e749dbefea76295d427a12487b63a142a7d6a2e4"} Jan 26 16:59:59 crc kubenswrapper[4754]: I0126 16:59:59.892396 4754 scope.go:117] "RemoveContainer" containerID="2c059b7d618442c8e541bce1d2e576de3e9d9f9b3bd82973f0039fc6c8429d6e" Jan 26 16:59:59 crc kubenswrapper[4754]: I0126 16:59:59.902463 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-jsbxt_689569c4-93ee-4b82-93fc-9af0f97e6dc3/ovn-acl-logging/0.log" Jan 26 16:59:59 crc kubenswrapper[4754]: I0126 16:59:59.904274 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-jsbxt_689569c4-93ee-4b82-93fc-9af0f97e6dc3/ovn-controller/0.log" Jan 26 16:59:59 crc kubenswrapper[4754]: I0126 16:59:59.904896 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-jsbxt" Jan 26 16:59:59 crc kubenswrapper[4754]: I0126 16:59:59.910978 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-687f57d79b-q26n7" Jan 26 16:59:59 crc kubenswrapper[4754]: I0126 16:59:59.978526 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-5tc6f"] Jan 26 16:59:59 crc kubenswrapper[4754]: E0126 16:59:59.978825 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="689569c4-93ee-4b82-93fc-9af0f97e6dc3" containerName="ovnkube-controller" Jan 26 16:59:59 crc kubenswrapper[4754]: I0126 16:59:59.978847 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="689569c4-93ee-4b82-93fc-9af0f97e6dc3" containerName="ovnkube-controller" Jan 26 16:59:59 crc kubenswrapper[4754]: E0126 16:59:59.978860 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="689569c4-93ee-4b82-93fc-9af0f97e6dc3" containerName="sbdb" Jan 26 16:59:59 crc kubenswrapper[4754]: I0126 16:59:59.978871 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="689569c4-93ee-4b82-93fc-9af0f97e6dc3" containerName="sbdb" Jan 26 16:59:59 crc kubenswrapper[4754]: E0126 16:59:59.978879 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="689569c4-93ee-4b82-93fc-9af0f97e6dc3" containerName="kube-rbac-proxy-node" Jan 26 16:59:59 crc kubenswrapper[4754]: I0126 16:59:59.978887 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="689569c4-93ee-4b82-93fc-9af0f97e6dc3" containerName="kube-rbac-proxy-node" Jan 26 16:59:59 crc kubenswrapper[4754]: E0126 16:59:59.978899 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="689569c4-93ee-4b82-93fc-9af0f97e6dc3" containerName="kubecfg-setup" Jan 26 16:59:59 crc kubenswrapper[4754]: I0126 16:59:59.978913 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="689569c4-93ee-4b82-93fc-9af0f97e6dc3" containerName="kubecfg-setup" Jan 26 16:59:59 crc kubenswrapper[4754]: E0126 16:59:59.978928 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="689569c4-93ee-4b82-93fc-9af0f97e6dc3" containerName="ovnkube-controller" Jan 26 16:59:59 crc kubenswrapper[4754]: I0126 16:59:59.978935 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="689569c4-93ee-4b82-93fc-9af0f97e6dc3" containerName="ovnkube-controller" Jan 26 16:59:59 crc kubenswrapper[4754]: E0126 16:59:59.978945 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="689569c4-93ee-4b82-93fc-9af0f97e6dc3" containerName="ovnkube-controller" Jan 26 16:59:59 crc kubenswrapper[4754]: I0126 16:59:59.978954 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="689569c4-93ee-4b82-93fc-9af0f97e6dc3" containerName="ovnkube-controller" Jan 26 16:59:59 crc kubenswrapper[4754]: E0126 16:59:59.978964 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="689569c4-93ee-4b82-93fc-9af0f97e6dc3" containerName="kube-rbac-proxy-ovn-metrics" Jan 26 16:59:59 crc kubenswrapper[4754]: I0126 16:59:59.978990 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="689569c4-93ee-4b82-93fc-9af0f97e6dc3" containerName="kube-rbac-proxy-ovn-metrics" Jan 26 16:59:59 crc kubenswrapper[4754]: E0126 16:59:59.979002 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="689569c4-93ee-4b82-93fc-9af0f97e6dc3" containerName="northd" Jan 26 16:59:59 crc kubenswrapper[4754]: I0126 16:59:59.979009 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="689569c4-93ee-4b82-93fc-9af0f97e6dc3" containerName="northd" Jan 26 16:59:59 crc kubenswrapper[4754]: E0126 16:59:59.979020 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="689569c4-93ee-4b82-93fc-9af0f97e6dc3" containerName="ovn-acl-logging" Jan 26 16:59:59 crc kubenswrapper[4754]: I0126 16:59:59.979028 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="689569c4-93ee-4b82-93fc-9af0f97e6dc3" containerName="ovn-acl-logging" Jan 26 16:59:59 crc kubenswrapper[4754]: E0126 16:59:59.979041 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="689569c4-93ee-4b82-93fc-9af0f97e6dc3" containerName="ovn-controller" Jan 26 16:59:59 crc kubenswrapper[4754]: I0126 16:59:59.979048 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="689569c4-93ee-4b82-93fc-9af0f97e6dc3" containerName="ovn-controller" Jan 26 16:59:59 crc kubenswrapper[4754]: E0126 16:59:59.979057 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="689569c4-93ee-4b82-93fc-9af0f97e6dc3" containerName="nbdb" Jan 26 16:59:59 crc kubenswrapper[4754]: I0126 16:59:59.979064 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="689569c4-93ee-4b82-93fc-9af0f97e6dc3" containerName="nbdb" Jan 26 16:59:59 crc kubenswrapper[4754]: E0126 16:59:59.979076 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="689569c4-93ee-4b82-93fc-9af0f97e6dc3" containerName="ovnkube-controller" Jan 26 16:59:59 crc kubenswrapper[4754]: I0126 16:59:59.979083 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="689569c4-93ee-4b82-93fc-9af0f97e6dc3" containerName="ovnkube-controller" Jan 26 16:59:59 crc kubenswrapper[4754]: I0126 16:59:59.979212 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="689569c4-93ee-4b82-93fc-9af0f97e6dc3" containerName="ovnkube-controller" Jan 26 16:59:59 crc kubenswrapper[4754]: I0126 16:59:59.979225 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="689569c4-93ee-4b82-93fc-9af0f97e6dc3" containerName="kube-rbac-proxy-node" Jan 26 16:59:59 crc kubenswrapper[4754]: I0126 16:59:59.979247 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="689569c4-93ee-4b82-93fc-9af0f97e6dc3" containerName="ovnkube-controller" Jan 26 16:59:59 crc kubenswrapper[4754]: I0126 16:59:59.979258 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="689569c4-93ee-4b82-93fc-9af0f97e6dc3" containerName="ovnkube-controller" Jan 26 16:59:59 crc kubenswrapper[4754]: I0126 16:59:59.979269 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="689569c4-93ee-4b82-93fc-9af0f97e6dc3" containerName="ovnkube-controller" Jan 26 16:59:59 crc kubenswrapper[4754]: I0126 16:59:59.979279 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="689569c4-93ee-4b82-93fc-9af0f97e6dc3" containerName="ovn-acl-logging" Jan 26 16:59:59 crc kubenswrapper[4754]: I0126 16:59:59.979287 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="689569c4-93ee-4b82-93fc-9af0f97e6dc3" containerName="northd" Jan 26 16:59:59 crc kubenswrapper[4754]: I0126 16:59:59.979297 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="689569c4-93ee-4b82-93fc-9af0f97e6dc3" containerName="ovnkube-controller" Jan 26 16:59:59 crc kubenswrapper[4754]: I0126 16:59:59.979309 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="689569c4-93ee-4b82-93fc-9af0f97e6dc3" containerName="nbdb" Jan 26 16:59:59 crc kubenswrapper[4754]: I0126 16:59:59.979317 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="689569c4-93ee-4b82-93fc-9af0f97e6dc3" containerName="sbdb" Jan 26 16:59:59 crc kubenswrapper[4754]: I0126 16:59:59.979329 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="689569c4-93ee-4b82-93fc-9af0f97e6dc3" containerName="ovn-controller" Jan 26 16:59:59 crc kubenswrapper[4754]: I0126 16:59:59.979338 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="689569c4-93ee-4b82-93fc-9af0f97e6dc3" containerName="kube-rbac-proxy-ovn-metrics" Jan 26 16:59:59 crc kubenswrapper[4754]: E0126 16:59:59.979457 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="689569c4-93ee-4b82-93fc-9af0f97e6dc3" containerName="ovnkube-controller" Jan 26 16:59:59 crc kubenswrapper[4754]: I0126 16:59:59.979467 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="689569c4-93ee-4b82-93fc-9af0f97e6dc3" containerName="ovnkube-controller" Jan 26 16:59:59 crc kubenswrapper[4754]: I0126 16:59:59.981492 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-5tc6f" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.107724 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/689569c4-93ee-4b82-93fc-9af0f97e6dc3-log-socket\") pod \"689569c4-93ee-4b82-93fc-9af0f97e6dc3\" (UID: \"689569c4-93ee-4b82-93fc-9af0f97e6dc3\") " Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.107767 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/689569c4-93ee-4b82-93fc-9af0f97e6dc3-host-slash\") pod \"689569c4-93ee-4b82-93fc-9af0f97e6dc3\" (UID: \"689569c4-93ee-4b82-93fc-9af0f97e6dc3\") " Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.107795 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/689569c4-93ee-4b82-93fc-9af0f97e6dc3-run-systemd\") pod \"689569c4-93ee-4b82-93fc-9af0f97e6dc3\" (UID: \"689569c4-93ee-4b82-93fc-9af0f97e6dc3\") " Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.107817 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/689569c4-93ee-4b82-93fc-9af0f97e6dc3-host-cni-netd\") pod \"689569c4-93ee-4b82-93fc-9af0f97e6dc3\" (UID: \"689569c4-93ee-4b82-93fc-9af0f97e6dc3\") " Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.107841 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/689569c4-93ee-4b82-93fc-9af0f97e6dc3-systemd-units\") pod \"689569c4-93ee-4b82-93fc-9af0f97e6dc3\" (UID: \"689569c4-93ee-4b82-93fc-9af0f97e6dc3\") " Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.107836 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/689569c4-93ee-4b82-93fc-9af0f97e6dc3-log-socket" (OuterVolumeSpecName: "log-socket") pod "689569c4-93ee-4b82-93fc-9af0f97e6dc3" (UID: "689569c4-93ee-4b82-93fc-9af0f97e6dc3"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.107886 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/689569c4-93ee-4b82-93fc-9af0f97e6dc3-env-overrides\") pod \"689569c4-93ee-4b82-93fc-9af0f97e6dc3\" (UID: \"689569c4-93ee-4b82-93fc-9af0f97e6dc3\") " Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.107914 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/689569c4-93ee-4b82-93fc-9af0f97e6dc3-node-log\") pod \"689569c4-93ee-4b82-93fc-9af0f97e6dc3\" (UID: \"689569c4-93ee-4b82-93fc-9af0f97e6dc3\") " Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.107946 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/689569c4-93ee-4b82-93fc-9af0f97e6dc3-run-ovn\") pod \"689569c4-93ee-4b82-93fc-9af0f97e6dc3\" (UID: \"689569c4-93ee-4b82-93fc-9af0f97e6dc3\") " Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.107981 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/689569c4-93ee-4b82-93fc-9af0f97e6dc3-ovn-node-metrics-cert\") pod \"689569c4-93ee-4b82-93fc-9af0f97e6dc3\" (UID: \"689569c4-93ee-4b82-93fc-9af0f97e6dc3\") " Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.108003 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/689569c4-93ee-4b82-93fc-9af0f97e6dc3-etc-openvswitch\") pod \"689569c4-93ee-4b82-93fc-9af0f97e6dc3\" (UID: \"689569c4-93ee-4b82-93fc-9af0f97e6dc3\") " Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.108022 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/689569c4-93ee-4b82-93fc-9af0f97e6dc3-var-lib-openvswitch\") pod \"689569c4-93ee-4b82-93fc-9af0f97e6dc3\" (UID: \"689569c4-93ee-4b82-93fc-9af0f97e6dc3\") " Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.108050 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/689569c4-93ee-4b82-93fc-9af0f97e6dc3-host-var-lib-cni-networks-ovn-kubernetes\") pod \"689569c4-93ee-4b82-93fc-9af0f97e6dc3\" (UID: \"689569c4-93ee-4b82-93fc-9af0f97e6dc3\") " Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.108074 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/689569c4-93ee-4b82-93fc-9af0f97e6dc3-ovnkube-config\") pod \"689569c4-93ee-4b82-93fc-9af0f97e6dc3\" (UID: \"689569c4-93ee-4b82-93fc-9af0f97e6dc3\") " Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.108093 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/689569c4-93ee-4b82-93fc-9af0f97e6dc3-host-run-netns\") pod \"689569c4-93ee-4b82-93fc-9af0f97e6dc3\" (UID: \"689569c4-93ee-4b82-93fc-9af0f97e6dc3\") " Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.108112 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/689569c4-93ee-4b82-93fc-9af0f97e6dc3-host-run-ovn-kubernetes\") pod \"689569c4-93ee-4b82-93fc-9af0f97e6dc3\" (UID: \"689569c4-93ee-4b82-93fc-9af0f97e6dc3\") " Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.108152 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cpxwt\" (UniqueName: \"kubernetes.io/projected/689569c4-93ee-4b82-93fc-9af0f97e6dc3-kube-api-access-cpxwt\") pod \"689569c4-93ee-4b82-93fc-9af0f97e6dc3\" (UID: \"689569c4-93ee-4b82-93fc-9af0f97e6dc3\") " Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.108178 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/689569c4-93ee-4b82-93fc-9af0f97e6dc3-ovnkube-script-lib\") pod \"689569c4-93ee-4b82-93fc-9af0f97e6dc3\" (UID: \"689569c4-93ee-4b82-93fc-9af0f97e6dc3\") " Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.108210 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/689569c4-93ee-4b82-93fc-9af0f97e6dc3-host-cni-bin\") pod \"689569c4-93ee-4b82-93fc-9af0f97e6dc3\" (UID: \"689569c4-93ee-4b82-93fc-9af0f97e6dc3\") " Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.108228 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/689569c4-93ee-4b82-93fc-9af0f97e6dc3-host-kubelet\") pod \"689569c4-93ee-4b82-93fc-9af0f97e6dc3\" (UID: \"689569c4-93ee-4b82-93fc-9af0f97e6dc3\") " Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.108258 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/689569c4-93ee-4b82-93fc-9af0f97e6dc3-run-openvswitch\") pod \"689569c4-93ee-4b82-93fc-9af0f97e6dc3\" (UID: \"689569c4-93ee-4b82-93fc-9af0f97e6dc3\") " Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.108387 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/d1ff24b0-9c9f-4ecc-a7f9-aeeb99fd7e8d-ovn-node-metrics-cert\") pod \"ovnkube-node-5tc6f\" (UID: \"d1ff24b0-9c9f-4ecc-a7f9-aeeb99fd7e8d\") " pod="openshift-ovn-kubernetes/ovnkube-node-5tc6f" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.108415 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/d1ff24b0-9c9f-4ecc-a7f9-aeeb99fd7e8d-run-ovn\") pod \"ovnkube-node-5tc6f\" (UID: \"d1ff24b0-9c9f-4ecc-a7f9-aeeb99fd7e8d\") " pod="openshift-ovn-kubernetes/ovnkube-node-5tc6f" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.108440 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/d1ff24b0-9c9f-4ecc-a7f9-aeeb99fd7e8d-host-run-netns\") pod \"ovnkube-node-5tc6f\" (UID: \"d1ff24b0-9c9f-4ecc-a7f9-aeeb99fd7e8d\") " pod="openshift-ovn-kubernetes/ovnkube-node-5tc6f" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.108460 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/d1ff24b0-9c9f-4ecc-a7f9-aeeb99fd7e8d-systemd-units\") pod \"ovnkube-node-5tc6f\" (UID: \"d1ff24b0-9c9f-4ecc-a7f9-aeeb99fd7e8d\") " pod="openshift-ovn-kubernetes/ovnkube-node-5tc6f" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.108478 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1ff24b0-9c9f-4ecc-a7f9-aeeb99fd7e8d-host-kubelet\") pod \"ovnkube-node-5tc6f\" (UID: \"d1ff24b0-9c9f-4ecc-a7f9-aeeb99fd7e8d\") " pod="openshift-ovn-kubernetes/ovnkube-node-5tc6f" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.108501 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d1ff24b0-9c9f-4ecc-a7f9-aeeb99fd7e8d-run-openvswitch\") pod \"ovnkube-node-5tc6f\" (UID: \"d1ff24b0-9c9f-4ecc-a7f9-aeeb99fd7e8d\") " pod="openshift-ovn-kubernetes/ovnkube-node-5tc6f" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.108526 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d1ff24b0-9c9f-4ecc-a7f9-aeeb99fd7e8d-host-run-ovn-kubernetes\") pod \"ovnkube-node-5tc6f\" (UID: \"d1ff24b0-9c9f-4ecc-a7f9-aeeb99fd7e8d\") " pod="openshift-ovn-kubernetes/ovnkube-node-5tc6f" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.108559 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d1ff24b0-9c9f-4ecc-a7f9-aeeb99fd7e8d-etc-openvswitch\") pod \"ovnkube-node-5tc6f\" (UID: \"d1ff24b0-9c9f-4ecc-a7f9-aeeb99fd7e8d\") " pod="openshift-ovn-kubernetes/ovnkube-node-5tc6f" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.108599 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/d1ff24b0-9c9f-4ecc-a7f9-aeeb99fd7e8d-host-cni-bin\") pod \"ovnkube-node-5tc6f\" (UID: \"d1ff24b0-9c9f-4ecc-a7f9-aeeb99fd7e8d\") " pod="openshift-ovn-kubernetes/ovnkube-node-5tc6f" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.108630 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/d1ff24b0-9c9f-4ecc-a7f9-aeeb99fd7e8d-env-overrides\") pod \"ovnkube-node-5tc6f\" (UID: \"d1ff24b0-9c9f-4ecc-a7f9-aeeb99fd7e8d\") " pod="openshift-ovn-kubernetes/ovnkube-node-5tc6f" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.108649 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/d1ff24b0-9c9f-4ecc-a7f9-aeeb99fd7e8d-run-systemd\") pod \"ovnkube-node-5tc6f\" (UID: \"d1ff24b0-9c9f-4ecc-a7f9-aeeb99fd7e8d\") " pod="openshift-ovn-kubernetes/ovnkube-node-5tc6f" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.108697 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kzdxh\" (UniqueName: \"kubernetes.io/projected/d1ff24b0-9c9f-4ecc-a7f9-aeeb99fd7e8d-kube-api-access-kzdxh\") pod \"ovnkube-node-5tc6f\" (UID: \"d1ff24b0-9c9f-4ecc-a7f9-aeeb99fd7e8d\") " pod="openshift-ovn-kubernetes/ovnkube-node-5tc6f" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.108732 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/d1ff24b0-9c9f-4ecc-a7f9-aeeb99fd7e8d-host-cni-netd\") pod \"ovnkube-node-5tc6f\" (UID: \"d1ff24b0-9c9f-4ecc-a7f9-aeeb99fd7e8d\") " pod="openshift-ovn-kubernetes/ovnkube-node-5tc6f" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.108758 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d1ff24b0-9c9f-4ecc-a7f9-aeeb99fd7e8d-host-slash\") pod \"ovnkube-node-5tc6f\" (UID: \"d1ff24b0-9c9f-4ecc-a7f9-aeeb99fd7e8d\") " pod="openshift-ovn-kubernetes/ovnkube-node-5tc6f" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.108788 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d1ff24b0-9c9f-4ecc-a7f9-aeeb99fd7e8d-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-5tc6f\" (UID: \"d1ff24b0-9c9f-4ecc-a7f9-aeeb99fd7e8d\") " pod="openshift-ovn-kubernetes/ovnkube-node-5tc6f" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.108812 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/d1ff24b0-9c9f-4ecc-a7f9-aeeb99fd7e8d-ovnkube-config\") pod \"ovnkube-node-5tc6f\" (UID: \"d1ff24b0-9c9f-4ecc-a7f9-aeeb99fd7e8d\") " pod="openshift-ovn-kubernetes/ovnkube-node-5tc6f" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.107912 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/689569c4-93ee-4b82-93fc-9af0f97e6dc3-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "689569c4-93ee-4b82-93fc-9af0f97e6dc3" (UID: "689569c4-93ee-4b82-93fc-9af0f97e6dc3"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.108916 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/d1ff24b0-9c9f-4ecc-a7f9-aeeb99fd7e8d-ovnkube-script-lib\") pod \"ovnkube-node-5tc6f\" (UID: \"d1ff24b0-9c9f-4ecc-a7f9-aeeb99fd7e8d\") " pod="openshift-ovn-kubernetes/ovnkube-node-5tc6f" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.108951 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/689569c4-93ee-4b82-93fc-9af0f97e6dc3-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "689569c4-93ee-4b82-93fc-9af0f97e6dc3" (UID: "689569c4-93ee-4b82-93fc-9af0f97e6dc3"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.108993 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d1ff24b0-9c9f-4ecc-a7f9-aeeb99fd7e8d-var-lib-openvswitch\") pod \"ovnkube-node-5tc6f\" (UID: \"d1ff24b0-9c9f-4ecc-a7f9-aeeb99fd7e8d\") " pod="openshift-ovn-kubernetes/ovnkube-node-5tc6f" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.109038 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/689569c4-93ee-4b82-93fc-9af0f97e6dc3-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "689569c4-93ee-4b82-93fc-9af0f97e6dc3" (UID: "689569c4-93ee-4b82-93fc-9af0f97e6dc3"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.109061 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/689569c4-93ee-4b82-93fc-9af0f97e6dc3-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "689569c4-93ee-4b82-93fc-9af0f97e6dc3" (UID: "689569c4-93ee-4b82-93fc-9af0f97e6dc3"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.108661 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/689569c4-93ee-4b82-93fc-9af0f97e6dc3-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "689569c4-93ee-4b82-93fc-9af0f97e6dc3" (UID: "689569c4-93ee-4b82-93fc-9af0f97e6dc3"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.108740 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/689569c4-93ee-4b82-93fc-9af0f97e6dc3-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "689569c4-93ee-4b82-93fc-9af0f97e6dc3" (UID: "689569c4-93ee-4b82-93fc-9af0f97e6dc3"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.108786 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/689569c4-93ee-4b82-93fc-9af0f97e6dc3-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "689569c4-93ee-4b82-93fc-9af0f97e6dc3" (UID: "689569c4-93ee-4b82-93fc-9af0f97e6dc3"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.109161 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/689569c4-93ee-4b82-93fc-9af0f97e6dc3-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "689569c4-93ee-4b82-93fc-9af0f97e6dc3" (UID: "689569c4-93ee-4b82-93fc-9af0f97e6dc3"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.109194 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/689569c4-93ee-4b82-93fc-9af0f97e6dc3-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "689569c4-93ee-4b82-93fc-9af0f97e6dc3" (UID: "689569c4-93ee-4b82-93fc-9af0f97e6dc3"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.109406 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/689569c4-93ee-4b82-93fc-9af0f97e6dc3-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "689569c4-93ee-4b82-93fc-9af0f97e6dc3" (UID: "689569c4-93ee-4b82-93fc-9af0f97e6dc3"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.109485 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/689569c4-93ee-4b82-93fc-9af0f97e6dc3-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "689569c4-93ee-4b82-93fc-9af0f97e6dc3" (UID: "689569c4-93ee-4b82-93fc-9af0f97e6dc3"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.109122 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/d1ff24b0-9c9f-4ecc-a7f9-aeeb99fd7e8d-node-log\") pod \"ovnkube-node-5tc6f\" (UID: \"d1ff24b0-9c9f-4ecc-a7f9-aeeb99fd7e8d\") " pod="openshift-ovn-kubernetes/ovnkube-node-5tc6f" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.107905 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/689569c4-93ee-4b82-93fc-9af0f97e6dc3-host-slash" (OuterVolumeSpecName: "host-slash") pod "689569c4-93ee-4b82-93fc-9af0f97e6dc3" (UID: "689569c4-93ee-4b82-93fc-9af0f97e6dc3"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.109100 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/689569c4-93ee-4b82-93fc-9af0f97e6dc3-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "689569c4-93ee-4b82-93fc-9af0f97e6dc3" (UID: "689569c4-93ee-4b82-93fc-9af0f97e6dc3"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.108812 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/689569c4-93ee-4b82-93fc-9af0f97e6dc3-node-log" (OuterVolumeSpecName: "node-log") pod "689569c4-93ee-4b82-93fc-9af0f97e6dc3" (UID: "689569c4-93ee-4b82-93fc-9af0f97e6dc3"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.109583 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/689569c4-93ee-4b82-93fc-9af0f97e6dc3-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "689569c4-93ee-4b82-93fc-9af0f97e6dc3" (UID: "689569c4-93ee-4b82-93fc-9af0f97e6dc3"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.109603 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/d1ff24b0-9c9f-4ecc-a7f9-aeeb99fd7e8d-log-socket\") pod \"ovnkube-node-5tc6f\" (UID: \"d1ff24b0-9c9f-4ecc-a7f9-aeeb99fd7e8d\") " pod="openshift-ovn-kubernetes/ovnkube-node-5tc6f" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.109786 4754 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/689569c4-93ee-4b82-93fc-9af0f97e6dc3-env-overrides\") on node \"crc\" DevicePath \"\"" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.109806 4754 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/689569c4-93ee-4b82-93fc-9af0f97e6dc3-node-log\") on node \"crc\" DevicePath \"\"" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.109821 4754 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/689569c4-93ee-4b82-93fc-9af0f97e6dc3-run-ovn\") on node \"crc\" DevicePath \"\"" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.109834 4754 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/689569c4-93ee-4b82-93fc-9af0f97e6dc3-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.109847 4754 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/689569c4-93ee-4b82-93fc-9af0f97e6dc3-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.109863 4754 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/689569c4-93ee-4b82-93fc-9af0f97e6dc3-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.109877 4754 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/689569c4-93ee-4b82-93fc-9af0f97e6dc3-host-run-netns\") on node \"crc\" DevicePath \"\"" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.109890 4754 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/689569c4-93ee-4b82-93fc-9af0f97e6dc3-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.109903 4754 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/689569c4-93ee-4b82-93fc-9af0f97e6dc3-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.109915 4754 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/689569c4-93ee-4b82-93fc-9af0f97e6dc3-host-cni-bin\") on node \"crc\" DevicePath \"\"" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.109926 4754 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/689569c4-93ee-4b82-93fc-9af0f97e6dc3-host-kubelet\") on node \"crc\" DevicePath \"\"" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.109937 4754 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/689569c4-93ee-4b82-93fc-9af0f97e6dc3-run-openvswitch\") on node \"crc\" DevicePath \"\"" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.109948 4754 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/689569c4-93ee-4b82-93fc-9af0f97e6dc3-log-socket\") on node \"crc\" DevicePath \"\"" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.109960 4754 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/689569c4-93ee-4b82-93fc-9af0f97e6dc3-host-slash\") on node \"crc\" DevicePath \"\"" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.109972 4754 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/689569c4-93ee-4b82-93fc-9af0f97e6dc3-host-cni-netd\") on node \"crc\" DevicePath \"\"" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.109984 4754 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/689569c4-93ee-4b82-93fc-9af0f97e6dc3-systemd-units\") on node \"crc\" DevicePath \"\"" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.110135 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/689569c4-93ee-4b82-93fc-9af0f97e6dc3-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "689569c4-93ee-4b82-93fc-9af0f97e6dc3" (UID: "689569c4-93ee-4b82-93fc-9af0f97e6dc3"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.114750 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/689569c4-93ee-4b82-93fc-9af0f97e6dc3-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "689569c4-93ee-4b82-93fc-9af0f97e6dc3" (UID: "689569c4-93ee-4b82-93fc-9af0f97e6dc3"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.115306 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/689569c4-93ee-4b82-93fc-9af0f97e6dc3-kube-api-access-cpxwt" (OuterVolumeSpecName: "kube-api-access-cpxwt") pod "689569c4-93ee-4b82-93fc-9af0f97e6dc3" (UID: "689569c4-93ee-4b82-93fc-9af0f97e6dc3"). InnerVolumeSpecName "kube-api-access-cpxwt". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.125246 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/689569c4-93ee-4b82-93fc-9af0f97e6dc3-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "689569c4-93ee-4b82-93fc-9af0f97e6dc3" (UID: "689569c4-93ee-4b82-93fc-9af0f97e6dc3"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.173984 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29490780-7wcg8"] Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.174638 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29490780-7wcg8" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.176847 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.177012 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.211372 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/d1ff24b0-9c9f-4ecc-a7f9-aeeb99fd7e8d-ovnkube-script-lib\") pod \"ovnkube-node-5tc6f\" (UID: \"d1ff24b0-9c9f-4ecc-a7f9-aeeb99fd7e8d\") " pod="openshift-ovn-kubernetes/ovnkube-node-5tc6f" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.211432 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d1ff24b0-9c9f-4ecc-a7f9-aeeb99fd7e8d-var-lib-openvswitch\") pod \"ovnkube-node-5tc6f\" (UID: \"d1ff24b0-9c9f-4ecc-a7f9-aeeb99fd7e8d\") " pod="openshift-ovn-kubernetes/ovnkube-node-5tc6f" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.211467 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/d1ff24b0-9c9f-4ecc-a7f9-aeeb99fd7e8d-node-log\") pod \"ovnkube-node-5tc6f\" (UID: \"d1ff24b0-9c9f-4ecc-a7f9-aeeb99fd7e8d\") " pod="openshift-ovn-kubernetes/ovnkube-node-5tc6f" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.211495 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/d1ff24b0-9c9f-4ecc-a7f9-aeeb99fd7e8d-log-socket\") pod \"ovnkube-node-5tc6f\" (UID: \"d1ff24b0-9c9f-4ecc-a7f9-aeeb99fd7e8d\") " pod="openshift-ovn-kubernetes/ovnkube-node-5tc6f" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.211521 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/d1ff24b0-9c9f-4ecc-a7f9-aeeb99fd7e8d-ovn-node-metrics-cert\") pod \"ovnkube-node-5tc6f\" (UID: \"d1ff24b0-9c9f-4ecc-a7f9-aeeb99fd7e8d\") " pod="openshift-ovn-kubernetes/ovnkube-node-5tc6f" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.211542 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/d1ff24b0-9c9f-4ecc-a7f9-aeeb99fd7e8d-run-ovn\") pod \"ovnkube-node-5tc6f\" (UID: \"d1ff24b0-9c9f-4ecc-a7f9-aeeb99fd7e8d\") " pod="openshift-ovn-kubernetes/ovnkube-node-5tc6f" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.211562 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/d1ff24b0-9c9f-4ecc-a7f9-aeeb99fd7e8d-host-run-netns\") pod \"ovnkube-node-5tc6f\" (UID: \"d1ff24b0-9c9f-4ecc-a7f9-aeeb99fd7e8d\") " pod="openshift-ovn-kubernetes/ovnkube-node-5tc6f" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.211582 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/d1ff24b0-9c9f-4ecc-a7f9-aeeb99fd7e8d-systemd-units\") pod \"ovnkube-node-5tc6f\" (UID: \"d1ff24b0-9c9f-4ecc-a7f9-aeeb99fd7e8d\") " pod="openshift-ovn-kubernetes/ovnkube-node-5tc6f" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.211604 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1ff24b0-9c9f-4ecc-a7f9-aeeb99fd7e8d-host-kubelet\") pod \"ovnkube-node-5tc6f\" (UID: \"d1ff24b0-9c9f-4ecc-a7f9-aeeb99fd7e8d\") " pod="openshift-ovn-kubernetes/ovnkube-node-5tc6f" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.211628 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d1ff24b0-9c9f-4ecc-a7f9-aeeb99fd7e8d-run-openvswitch\") pod \"ovnkube-node-5tc6f\" (UID: \"d1ff24b0-9c9f-4ecc-a7f9-aeeb99fd7e8d\") " pod="openshift-ovn-kubernetes/ovnkube-node-5tc6f" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.211651 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d1ff24b0-9c9f-4ecc-a7f9-aeeb99fd7e8d-host-run-ovn-kubernetes\") pod \"ovnkube-node-5tc6f\" (UID: \"d1ff24b0-9c9f-4ecc-a7f9-aeeb99fd7e8d\") " pod="openshift-ovn-kubernetes/ovnkube-node-5tc6f" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.211706 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d1ff24b0-9c9f-4ecc-a7f9-aeeb99fd7e8d-etc-openvswitch\") pod \"ovnkube-node-5tc6f\" (UID: \"d1ff24b0-9c9f-4ecc-a7f9-aeeb99fd7e8d\") " pod="openshift-ovn-kubernetes/ovnkube-node-5tc6f" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.211735 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/d1ff24b0-9c9f-4ecc-a7f9-aeeb99fd7e8d-log-socket\") pod \"ovnkube-node-5tc6f\" (UID: \"d1ff24b0-9c9f-4ecc-a7f9-aeeb99fd7e8d\") " pod="openshift-ovn-kubernetes/ovnkube-node-5tc6f" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.211747 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/d1ff24b0-9c9f-4ecc-a7f9-aeeb99fd7e8d-host-cni-bin\") pod \"ovnkube-node-5tc6f\" (UID: \"d1ff24b0-9c9f-4ecc-a7f9-aeeb99fd7e8d\") " pod="openshift-ovn-kubernetes/ovnkube-node-5tc6f" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.211799 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/d1ff24b0-9c9f-4ecc-a7f9-aeeb99fd7e8d-host-cni-bin\") pod \"ovnkube-node-5tc6f\" (UID: \"d1ff24b0-9c9f-4ecc-a7f9-aeeb99fd7e8d\") " pod="openshift-ovn-kubernetes/ovnkube-node-5tc6f" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.211811 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/d1ff24b0-9c9f-4ecc-a7f9-aeeb99fd7e8d-node-log\") pod \"ovnkube-node-5tc6f\" (UID: \"d1ff24b0-9c9f-4ecc-a7f9-aeeb99fd7e8d\") " pod="openshift-ovn-kubernetes/ovnkube-node-5tc6f" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.211948 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/d1ff24b0-9c9f-4ecc-a7f9-aeeb99fd7e8d-systemd-units\") pod \"ovnkube-node-5tc6f\" (UID: \"d1ff24b0-9c9f-4ecc-a7f9-aeeb99fd7e8d\") " pod="openshift-ovn-kubernetes/ovnkube-node-5tc6f" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.211982 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d1ff24b0-9c9f-4ecc-a7f9-aeeb99fd7e8d-etc-openvswitch\") pod \"ovnkube-node-5tc6f\" (UID: \"d1ff24b0-9c9f-4ecc-a7f9-aeeb99fd7e8d\") " pod="openshift-ovn-kubernetes/ovnkube-node-5tc6f" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.211986 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1ff24b0-9c9f-4ecc-a7f9-aeeb99fd7e8d-host-kubelet\") pod \"ovnkube-node-5tc6f\" (UID: \"d1ff24b0-9c9f-4ecc-a7f9-aeeb99fd7e8d\") " pod="openshift-ovn-kubernetes/ovnkube-node-5tc6f" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.211998 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d1ff24b0-9c9f-4ecc-a7f9-aeeb99fd7e8d-host-run-ovn-kubernetes\") pod \"ovnkube-node-5tc6f\" (UID: \"d1ff24b0-9c9f-4ecc-a7f9-aeeb99fd7e8d\") " pod="openshift-ovn-kubernetes/ovnkube-node-5tc6f" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.212068 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/d1ff24b0-9c9f-4ecc-a7f9-aeeb99fd7e8d-run-systemd\") pod \"ovnkube-node-5tc6f\" (UID: \"d1ff24b0-9c9f-4ecc-a7f9-aeeb99fd7e8d\") " pod="openshift-ovn-kubernetes/ovnkube-node-5tc6f" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.212069 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/d1ff24b0-9c9f-4ecc-a7f9-aeeb99fd7e8d-run-ovn\") pod \"ovnkube-node-5tc6f\" (UID: \"d1ff24b0-9c9f-4ecc-a7f9-aeeb99fd7e8d\") " pod="openshift-ovn-kubernetes/ovnkube-node-5tc6f" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.212086 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/d1ff24b0-9c9f-4ecc-a7f9-aeeb99fd7e8d-host-run-netns\") pod \"ovnkube-node-5tc6f\" (UID: \"d1ff24b0-9c9f-4ecc-a7f9-aeeb99fd7e8d\") " pod="openshift-ovn-kubernetes/ovnkube-node-5tc6f" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.212106 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/d1ff24b0-9c9f-4ecc-a7f9-aeeb99fd7e8d-run-systemd\") pod \"ovnkube-node-5tc6f\" (UID: \"d1ff24b0-9c9f-4ecc-a7f9-aeeb99fd7e8d\") " pod="openshift-ovn-kubernetes/ovnkube-node-5tc6f" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.212116 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d1ff24b0-9c9f-4ecc-a7f9-aeeb99fd7e8d-var-lib-openvswitch\") pod \"ovnkube-node-5tc6f\" (UID: \"d1ff24b0-9c9f-4ecc-a7f9-aeeb99fd7e8d\") " pod="openshift-ovn-kubernetes/ovnkube-node-5tc6f" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.212140 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d1ff24b0-9c9f-4ecc-a7f9-aeeb99fd7e8d-run-openvswitch\") pod \"ovnkube-node-5tc6f\" (UID: \"d1ff24b0-9c9f-4ecc-a7f9-aeeb99fd7e8d\") " pod="openshift-ovn-kubernetes/ovnkube-node-5tc6f" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.212156 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/d1ff24b0-9c9f-4ecc-a7f9-aeeb99fd7e8d-env-overrides\") pod \"ovnkube-node-5tc6f\" (UID: \"d1ff24b0-9c9f-4ecc-a7f9-aeeb99fd7e8d\") " pod="openshift-ovn-kubernetes/ovnkube-node-5tc6f" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.212250 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kzdxh\" (UniqueName: \"kubernetes.io/projected/d1ff24b0-9c9f-4ecc-a7f9-aeeb99fd7e8d-kube-api-access-kzdxh\") pod \"ovnkube-node-5tc6f\" (UID: \"d1ff24b0-9c9f-4ecc-a7f9-aeeb99fd7e8d\") " pod="openshift-ovn-kubernetes/ovnkube-node-5tc6f" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.212282 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/d1ff24b0-9c9f-4ecc-a7f9-aeeb99fd7e8d-host-cni-netd\") pod \"ovnkube-node-5tc6f\" (UID: \"d1ff24b0-9c9f-4ecc-a7f9-aeeb99fd7e8d\") " pod="openshift-ovn-kubernetes/ovnkube-node-5tc6f" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.212307 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d1ff24b0-9c9f-4ecc-a7f9-aeeb99fd7e8d-host-slash\") pod \"ovnkube-node-5tc6f\" (UID: \"d1ff24b0-9c9f-4ecc-a7f9-aeeb99fd7e8d\") " pod="openshift-ovn-kubernetes/ovnkube-node-5tc6f" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.212351 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d1ff24b0-9c9f-4ecc-a7f9-aeeb99fd7e8d-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-5tc6f\" (UID: \"d1ff24b0-9c9f-4ecc-a7f9-aeeb99fd7e8d\") " pod="openshift-ovn-kubernetes/ovnkube-node-5tc6f" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.212380 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/d1ff24b0-9c9f-4ecc-a7f9-aeeb99fd7e8d-ovnkube-config\") pod \"ovnkube-node-5tc6f\" (UID: \"d1ff24b0-9c9f-4ecc-a7f9-aeeb99fd7e8d\") " pod="openshift-ovn-kubernetes/ovnkube-node-5tc6f" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.212426 4754 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/689569c4-93ee-4b82-93fc-9af0f97e6dc3-run-systemd\") on node \"crc\" DevicePath \"\"" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.212444 4754 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/689569c4-93ee-4b82-93fc-9af0f97e6dc3-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.212431 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/d1ff24b0-9c9f-4ecc-a7f9-aeeb99fd7e8d-host-cni-netd\") pod \"ovnkube-node-5tc6f\" (UID: \"d1ff24b0-9c9f-4ecc-a7f9-aeeb99fd7e8d\") " pod="openshift-ovn-kubernetes/ovnkube-node-5tc6f" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.212497 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d1ff24b0-9c9f-4ecc-a7f9-aeeb99fd7e8d-host-slash\") pod \"ovnkube-node-5tc6f\" (UID: \"d1ff24b0-9c9f-4ecc-a7f9-aeeb99fd7e8d\") " pod="openshift-ovn-kubernetes/ovnkube-node-5tc6f" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.212528 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d1ff24b0-9c9f-4ecc-a7f9-aeeb99fd7e8d-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-5tc6f\" (UID: \"d1ff24b0-9c9f-4ecc-a7f9-aeeb99fd7e8d\") " pod="openshift-ovn-kubernetes/ovnkube-node-5tc6f" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.212458 4754 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/689569c4-93ee-4b82-93fc-9af0f97e6dc3-ovnkube-config\") on node \"crc\" DevicePath \"\"" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.212557 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cpxwt\" (UniqueName: \"kubernetes.io/projected/689569c4-93ee-4b82-93fc-9af0f97e6dc3-kube-api-access-cpxwt\") on node \"crc\" DevicePath \"\"" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.213730 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/d1ff24b0-9c9f-4ecc-a7f9-aeeb99fd7e8d-env-overrides\") pod \"ovnkube-node-5tc6f\" (UID: \"d1ff24b0-9c9f-4ecc-a7f9-aeeb99fd7e8d\") " pod="openshift-ovn-kubernetes/ovnkube-node-5tc6f" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.213874 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/d1ff24b0-9c9f-4ecc-a7f9-aeeb99fd7e8d-ovnkube-config\") pod \"ovnkube-node-5tc6f\" (UID: \"d1ff24b0-9c9f-4ecc-a7f9-aeeb99fd7e8d\") " pod="openshift-ovn-kubernetes/ovnkube-node-5tc6f" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.214705 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/d1ff24b0-9c9f-4ecc-a7f9-aeeb99fd7e8d-ovnkube-script-lib\") pod \"ovnkube-node-5tc6f\" (UID: \"d1ff24b0-9c9f-4ecc-a7f9-aeeb99fd7e8d\") " pod="openshift-ovn-kubernetes/ovnkube-node-5tc6f" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.215998 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/d1ff24b0-9c9f-4ecc-a7f9-aeeb99fd7e8d-ovn-node-metrics-cert\") pod \"ovnkube-node-5tc6f\" (UID: \"d1ff24b0-9c9f-4ecc-a7f9-aeeb99fd7e8d\") " pod="openshift-ovn-kubernetes/ovnkube-node-5tc6f" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.239434 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kzdxh\" (UniqueName: \"kubernetes.io/projected/d1ff24b0-9c9f-4ecc-a7f9-aeeb99fd7e8d-kube-api-access-kzdxh\") pod \"ovnkube-node-5tc6f\" (UID: \"d1ff24b0-9c9f-4ecc-a7f9-aeeb99fd7e8d\") " pod="openshift-ovn-kubernetes/ovnkube-node-5tc6f" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.295040 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-5tc6f" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.313928 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4f2d7a6e-760a-4cd5-b9d0-7483d4f12933-config-volume\") pod \"collect-profiles-29490780-7wcg8\" (UID: \"4f2d7a6e-760a-4cd5-b9d0-7483d4f12933\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29490780-7wcg8" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.313974 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4f2d7a6e-760a-4cd5-b9d0-7483d4f12933-secret-volume\") pod \"collect-profiles-29490780-7wcg8\" (UID: \"4f2d7a6e-760a-4cd5-b9d0-7483d4f12933\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29490780-7wcg8" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.314013 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6dkbk\" (UniqueName: \"kubernetes.io/projected/4f2d7a6e-760a-4cd5-b9d0-7483d4f12933-kube-api-access-6dkbk\") pod \"collect-profiles-29490780-7wcg8\" (UID: \"4f2d7a6e-760a-4cd5-b9d0-7483d4f12933\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29490780-7wcg8" Jan 26 17:00:00 crc kubenswrapper[4754]: W0126 17:00:00.316834 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1ff24b0_9c9f_4ecc_a7f9_aeeb99fd7e8d.slice/crio-ec26d8c205bbed6f555291d0e10cd64ac2049b4c9929f479b84c65cf8effa8d3 WatchSource:0}: Error finding container ec26d8c205bbed6f555291d0e10cd64ac2049b4c9929f479b84c65cf8effa8d3: Status 404 returned error can't find the container with id ec26d8c205bbed6f555291d0e10cd64ac2049b4c9929f479b84c65cf8effa8d3 Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.414778 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4f2d7a6e-760a-4cd5-b9d0-7483d4f12933-secret-volume\") pod \"collect-profiles-29490780-7wcg8\" (UID: \"4f2d7a6e-760a-4cd5-b9d0-7483d4f12933\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29490780-7wcg8" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.415186 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6dkbk\" (UniqueName: \"kubernetes.io/projected/4f2d7a6e-760a-4cd5-b9d0-7483d4f12933-kube-api-access-6dkbk\") pod \"collect-profiles-29490780-7wcg8\" (UID: \"4f2d7a6e-760a-4cd5-b9d0-7483d4f12933\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29490780-7wcg8" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.415450 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4f2d7a6e-760a-4cd5-b9d0-7483d4f12933-config-volume\") pod \"collect-profiles-29490780-7wcg8\" (UID: \"4f2d7a6e-760a-4cd5-b9d0-7483d4f12933\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29490780-7wcg8" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.417071 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4f2d7a6e-760a-4cd5-b9d0-7483d4f12933-config-volume\") pod \"collect-profiles-29490780-7wcg8\" (UID: \"4f2d7a6e-760a-4cd5-b9d0-7483d4f12933\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29490780-7wcg8" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.418828 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-jsbxt_689569c4-93ee-4b82-93fc-9af0f97e6dc3/ovn-acl-logging/0.log" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.419586 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-jsbxt_689569c4-93ee-4b82-93fc-9af0f97e6dc3/ovn-controller/0.log" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.420152 4754 generic.go:334] "Generic (PLEG): container finished" podID="689569c4-93ee-4b82-93fc-9af0f97e6dc3" containerID="07a43ff7860ff65faf51f4a5e8aeb326f06d6d7222677a92560ae2e3a0c81484" exitCode=0 Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.420215 4754 generic.go:334] "Generic (PLEG): container finished" podID="689569c4-93ee-4b82-93fc-9af0f97e6dc3" containerID="b1ac1f49bd564750fcff530a0f76657b0ed41573539acd020a922d19861ef834" exitCode=0 Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.420238 4754 generic.go:334] "Generic (PLEG): container finished" podID="689569c4-93ee-4b82-93fc-9af0f97e6dc3" containerID="a300a80ad87f46ba1b4fd7ad3cf6c25767fcfe966b93bdd86d729843f53a5cda" exitCode=0 Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.420266 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jsbxt" event={"ID":"689569c4-93ee-4b82-93fc-9af0f97e6dc3","Type":"ContainerDied","Data":"07a43ff7860ff65faf51f4a5e8aeb326f06d6d7222677a92560ae2e3a0c81484"} Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.420328 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jsbxt" event={"ID":"689569c4-93ee-4b82-93fc-9af0f97e6dc3","Type":"ContainerDied","Data":"b1ac1f49bd564750fcff530a0f76657b0ed41573539acd020a922d19861ef834"} Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.420347 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jsbxt" event={"ID":"689569c4-93ee-4b82-93fc-9af0f97e6dc3","Type":"ContainerDied","Data":"a300a80ad87f46ba1b4fd7ad3cf6c25767fcfe966b93bdd86d729843f53a5cda"} Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.420361 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jsbxt" event={"ID":"689569c4-93ee-4b82-93fc-9af0f97e6dc3","Type":"ContainerDied","Data":"25d5c4accfdda0e7dca5e9aac838cceced3c6028b8e097f54e22904a0c839824"} Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.420383 4754 scope.go:117] "RemoveContainer" containerID="0d6f7aa12978cc78e269af0ca6f59f424d8898f692c6f8d07454a1868b0204cd" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.420656 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-jsbxt" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.422702 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5tc6f" event={"ID":"d1ff24b0-9c9f-4ecc-a7f9-aeeb99fd7e8d","Type":"ContainerStarted","Data":"ec26d8c205bbed6f555291d0e10cd64ac2049b4c9929f479b84c65cf8effa8d3"} Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.422968 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4f2d7a6e-760a-4cd5-b9d0-7483d4f12933-secret-volume\") pod \"collect-profiles-29490780-7wcg8\" (UID: \"4f2d7a6e-760a-4cd5-b9d0-7483d4f12933\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29490780-7wcg8" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.425014 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-skh5n_b619d233-b592-4b05-a0b7-dc094c88471e/kube-multus/2.log" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.425062 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-skh5n" event={"ID":"b619d233-b592-4b05-a0b7-dc094c88471e","Type":"ContainerStarted","Data":"f8c097655f185d83c1ca55d7d3182749ad2be40b1d51a15282a5b978a389a800"} Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.438552 4754 scope.go:117] "RemoveContainer" containerID="07a43ff7860ff65faf51f4a5e8aeb326f06d6d7222677a92560ae2e3a0c81484" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.442914 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6dkbk\" (UniqueName: \"kubernetes.io/projected/4f2d7a6e-760a-4cd5-b9d0-7483d4f12933-kube-api-access-6dkbk\") pod \"collect-profiles-29490780-7wcg8\" (UID: \"4f2d7a6e-760a-4cd5-b9d0-7483d4f12933\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29490780-7wcg8" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.452988 4754 scope.go:117] "RemoveContainer" containerID="b1ac1f49bd564750fcff530a0f76657b0ed41573539acd020a922d19861ef834" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.470361 4754 scope.go:117] "RemoveContainer" containerID="a300a80ad87f46ba1b4fd7ad3cf6c25767fcfe966b93bdd86d729843f53a5cda" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.488306 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-jsbxt"] Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.488970 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-jsbxt"] Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.496183 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29490780-7wcg8" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.498533 4754 scope.go:117] "RemoveContainer" containerID="919572e731813a98c7d8a94879bba931f75c472c7cd681e86d5197ebab43815d" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.514148 4754 scope.go:117] "RemoveContainer" containerID="033f8b104d4e9ff6d5f87ca998665f12e0c56696b3d30e01ede05f4da540365d" Jan 26 17:00:00 crc kubenswrapper[4754]: E0126 17:00:00.525233 4754 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_collect-profiles-29490780-7wcg8_openshift-operator-lifecycle-manager_4f2d7a6e-760a-4cd5-b9d0-7483d4f12933_0(48267e05e17f3b55968ccd5d65d3f6a0935b88bd26b890e00f2be05c463a63e7): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Jan 26 17:00:00 crc kubenswrapper[4754]: E0126 17:00:00.525300 4754 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_collect-profiles-29490780-7wcg8_openshift-operator-lifecycle-manager_4f2d7a6e-760a-4cd5-b9d0-7483d4f12933_0(48267e05e17f3b55968ccd5d65d3f6a0935b88bd26b890e00f2be05c463a63e7): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operator-lifecycle-manager/collect-profiles-29490780-7wcg8" Jan 26 17:00:00 crc kubenswrapper[4754]: E0126 17:00:00.525321 4754 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_collect-profiles-29490780-7wcg8_openshift-operator-lifecycle-manager_4f2d7a6e-760a-4cd5-b9d0-7483d4f12933_0(48267e05e17f3b55968ccd5d65d3f6a0935b88bd26b890e00f2be05c463a63e7): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operator-lifecycle-manager/collect-profiles-29490780-7wcg8" Jan 26 17:00:00 crc kubenswrapper[4754]: E0126 17:00:00.525365 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"collect-profiles-29490780-7wcg8_openshift-operator-lifecycle-manager(4f2d7a6e-760a-4cd5-b9d0-7483d4f12933)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"collect-profiles-29490780-7wcg8_openshift-operator-lifecycle-manager(4f2d7a6e-760a-4cd5-b9d0-7483d4f12933)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_collect-profiles-29490780-7wcg8_openshift-operator-lifecycle-manager_4f2d7a6e-760a-4cd5-b9d0-7483d4f12933_0(48267e05e17f3b55968ccd5d65d3f6a0935b88bd26b890e00f2be05c463a63e7): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operator-lifecycle-manager/collect-profiles-29490780-7wcg8" podUID="4f2d7a6e-760a-4cd5-b9d0-7483d4f12933" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.532333 4754 scope.go:117] "RemoveContainer" containerID="31dcab2f9a2d218b0e49dad1df46377ff4521592824a0fbac254d5b8e6d0632e" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.578738 4754 scope.go:117] "RemoveContainer" containerID="3fe4d9395447ac570eca1338e749dbefea76295d427a12487b63a142a7d6a2e4" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.597926 4754 scope.go:117] "RemoveContainer" containerID="8d9a7bafabd44f8b88a6f62cbe45599af8563bc45d9d7ad3a7669f3d8d11eb9c" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.611637 4754 scope.go:117] "RemoveContainer" containerID="0d6f7aa12978cc78e269af0ca6f59f424d8898f692c6f8d07454a1868b0204cd" Jan 26 17:00:00 crc kubenswrapper[4754]: E0126 17:00:00.612096 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0d6f7aa12978cc78e269af0ca6f59f424d8898f692c6f8d07454a1868b0204cd\": container with ID starting with 0d6f7aa12978cc78e269af0ca6f59f424d8898f692c6f8d07454a1868b0204cd not found: ID does not exist" containerID="0d6f7aa12978cc78e269af0ca6f59f424d8898f692c6f8d07454a1868b0204cd" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.612137 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0d6f7aa12978cc78e269af0ca6f59f424d8898f692c6f8d07454a1868b0204cd"} err="failed to get container status \"0d6f7aa12978cc78e269af0ca6f59f424d8898f692c6f8d07454a1868b0204cd\": rpc error: code = NotFound desc = could not find container \"0d6f7aa12978cc78e269af0ca6f59f424d8898f692c6f8d07454a1868b0204cd\": container with ID starting with 0d6f7aa12978cc78e269af0ca6f59f424d8898f692c6f8d07454a1868b0204cd not found: ID does not exist" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.612164 4754 scope.go:117] "RemoveContainer" containerID="07a43ff7860ff65faf51f4a5e8aeb326f06d6d7222677a92560ae2e3a0c81484" Jan 26 17:00:00 crc kubenswrapper[4754]: E0126 17:00:00.612616 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"07a43ff7860ff65faf51f4a5e8aeb326f06d6d7222677a92560ae2e3a0c81484\": container with ID starting with 07a43ff7860ff65faf51f4a5e8aeb326f06d6d7222677a92560ae2e3a0c81484 not found: ID does not exist" containerID="07a43ff7860ff65faf51f4a5e8aeb326f06d6d7222677a92560ae2e3a0c81484" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.612647 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"07a43ff7860ff65faf51f4a5e8aeb326f06d6d7222677a92560ae2e3a0c81484"} err="failed to get container status \"07a43ff7860ff65faf51f4a5e8aeb326f06d6d7222677a92560ae2e3a0c81484\": rpc error: code = NotFound desc = could not find container \"07a43ff7860ff65faf51f4a5e8aeb326f06d6d7222677a92560ae2e3a0c81484\": container with ID starting with 07a43ff7860ff65faf51f4a5e8aeb326f06d6d7222677a92560ae2e3a0c81484 not found: ID does not exist" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.612682 4754 scope.go:117] "RemoveContainer" containerID="b1ac1f49bd564750fcff530a0f76657b0ed41573539acd020a922d19861ef834" Jan 26 17:00:00 crc kubenswrapper[4754]: E0126 17:00:00.612996 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b1ac1f49bd564750fcff530a0f76657b0ed41573539acd020a922d19861ef834\": container with ID starting with b1ac1f49bd564750fcff530a0f76657b0ed41573539acd020a922d19861ef834 not found: ID does not exist" containerID="b1ac1f49bd564750fcff530a0f76657b0ed41573539acd020a922d19861ef834" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.613016 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b1ac1f49bd564750fcff530a0f76657b0ed41573539acd020a922d19861ef834"} err="failed to get container status \"b1ac1f49bd564750fcff530a0f76657b0ed41573539acd020a922d19861ef834\": rpc error: code = NotFound desc = could not find container \"b1ac1f49bd564750fcff530a0f76657b0ed41573539acd020a922d19861ef834\": container with ID starting with b1ac1f49bd564750fcff530a0f76657b0ed41573539acd020a922d19861ef834 not found: ID does not exist" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.613031 4754 scope.go:117] "RemoveContainer" containerID="a300a80ad87f46ba1b4fd7ad3cf6c25767fcfe966b93bdd86d729843f53a5cda" Jan 26 17:00:00 crc kubenswrapper[4754]: E0126 17:00:00.613315 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a300a80ad87f46ba1b4fd7ad3cf6c25767fcfe966b93bdd86d729843f53a5cda\": container with ID starting with a300a80ad87f46ba1b4fd7ad3cf6c25767fcfe966b93bdd86d729843f53a5cda not found: ID does not exist" containerID="a300a80ad87f46ba1b4fd7ad3cf6c25767fcfe966b93bdd86d729843f53a5cda" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.613342 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a300a80ad87f46ba1b4fd7ad3cf6c25767fcfe966b93bdd86d729843f53a5cda"} err="failed to get container status \"a300a80ad87f46ba1b4fd7ad3cf6c25767fcfe966b93bdd86d729843f53a5cda\": rpc error: code = NotFound desc = could not find container \"a300a80ad87f46ba1b4fd7ad3cf6c25767fcfe966b93bdd86d729843f53a5cda\": container with ID starting with a300a80ad87f46ba1b4fd7ad3cf6c25767fcfe966b93bdd86d729843f53a5cda not found: ID does not exist" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.613360 4754 scope.go:117] "RemoveContainer" containerID="919572e731813a98c7d8a94879bba931f75c472c7cd681e86d5197ebab43815d" Jan 26 17:00:00 crc kubenswrapper[4754]: E0126 17:00:00.613707 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"919572e731813a98c7d8a94879bba931f75c472c7cd681e86d5197ebab43815d\": container with ID starting with 919572e731813a98c7d8a94879bba931f75c472c7cd681e86d5197ebab43815d not found: ID does not exist" containerID="919572e731813a98c7d8a94879bba931f75c472c7cd681e86d5197ebab43815d" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.613737 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"919572e731813a98c7d8a94879bba931f75c472c7cd681e86d5197ebab43815d"} err="failed to get container status \"919572e731813a98c7d8a94879bba931f75c472c7cd681e86d5197ebab43815d\": rpc error: code = NotFound desc = could not find container \"919572e731813a98c7d8a94879bba931f75c472c7cd681e86d5197ebab43815d\": container with ID starting with 919572e731813a98c7d8a94879bba931f75c472c7cd681e86d5197ebab43815d not found: ID does not exist" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.613756 4754 scope.go:117] "RemoveContainer" containerID="033f8b104d4e9ff6d5f87ca998665f12e0c56696b3d30e01ede05f4da540365d" Jan 26 17:00:00 crc kubenswrapper[4754]: E0126 17:00:00.614107 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"033f8b104d4e9ff6d5f87ca998665f12e0c56696b3d30e01ede05f4da540365d\": container with ID starting with 033f8b104d4e9ff6d5f87ca998665f12e0c56696b3d30e01ede05f4da540365d not found: ID does not exist" containerID="033f8b104d4e9ff6d5f87ca998665f12e0c56696b3d30e01ede05f4da540365d" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.614143 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"033f8b104d4e9ff6d5f87ca998665f12e0c56696b3d30e01ede05f4da540365d"} err="failed to get container status \"033f8b104d4e9ff6d5f87ca998665f12e0c56696b3d30e01ede05f4da540365d\": rpc error: code = NotFound desc = could not find container \"033f8b104d4e9ff6d5f87ca998665f12e0c56696b3d30e01ede05f4da540365d\": container with ID starting with 033f8b104d4e9ff6d5f87ca998665f12e0c56696b3d30e01ede05f4da540365d not found: ID does not exist" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.614156 4754 scope.go:117] "RemoveContainer" containerID="31dcab2f9a2d218b0e49dad1df46377ff4521592824a0fbac254d5b8e6d0632e" Jan 26 17:00:00 crc kubenswrapper[4754]: E0126 17:00:00.614583 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"31dcab2f9a2d218b0e49dad1df46377ff4521592824a0fbac254d5b8e6d0632e\": container with ID starting with 31dcab2f9a2d218b0e49dad1df46377ff4521592824a0fbac254d5b8e6d0632e not found: ID does not exist" containerID="31dcab2f9a2d218b0e49dad1df46377ff4521592824a0fbac254d5b8e6d0632e" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.614603 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"31dcab2f9a2d218b0e49dad1df46377ff4521592824a0fbac254d5b8e6d0632e"} err="failed to get container status \"31dcab2f9a2d218b0e49dad1df46377ff4521592824a0fbac254d5b8e6d0632e\": rpc error: code = NotFound desc = could not find container \"31dcab2f9a2d218b0e49dad1df46377ff4521592824a0fbac254d5b8e6d0632e\": container with ID starting with 31dcab2f9a2d218b0e49dad1df46377ff4521592824a0fbac254d5b8e6d0632e not found: ID does not exist" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.614615 4754 scope.go:117] "RemoveContainer" containerID="3fe4d9395447ac570eca1338e749dbefea76295d427a12487b63a142a7d6a2e4" Jan 26 17:00:00 crc kubenswrapper[4754]: E0126 17:00:00.614858 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3fe4d9395447ac570eca1338e749dbefea76295d427a12487b63a142a7d6a2e4\": container with ID starting with 3fe4d9395447ac570eca1338e749dbefea76295d427a12487b63a142a7d6a2e4 not found: ID does not exist" containerID="3fe4d9395447ac570eca1338e749dbefea76295d427a12487b63a142a7d6a2e4" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.614886 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3fe4d9395447ac570eca1338e749dbefea76295d427a12487b63a142a7d6a2e4"} err="failed to get container status \"3fe4d9395447ac570eca1338e749dbefea76295d427a12487b63a142a7d6a2e4\": rpc error: code = NotFound desc = could not find container \"3fe4d9395447ac570eca1338e749dbefea76295d427a12487b63a142a7d6a2e4\": container with ID starting with 3fe4d9395447ac570eca1338e749dbefea76295d427a12487b63a142a7d6a2e4 not found: ID does not exist" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.614918 4754 scope.go:117] "RemoveContainer" containerID="8d9a7bafabd44f8b88a6f62cbe45599af8563bc45d9d7ad3a7669f3d8d11eb9c" Jan 26 17:00:00 crc kubenswrapper[4754]: E0126 17:00:00.615135 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8d9a7bafabd44f8b88a6f62cbe45599af8563bc45d9d7ad3a7669f3d8d11eb9c\": container with ID starting with 8d9a7bafabd44f8b88a6f62cbe45599af8563bc45d9d7ad3a7669f3d8d11eb9c not found: ID does not exist" containerID="8d9a7bafabd44f8b88a6f62cbe45599af8563bc45d9d7ad3a7669f3d8d11eb9c" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.615169 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8d9a7bafabd44f8b88a6f62cbe45599af8563bc45d9d7ad3a7669f3d8d11eb9c"} err="failed to get container status \"8d9a7bafabd44f8b88a6f62cbe45599af8563bc45d9d7ad3a7669f3d8d11eb9c\": rpc error: code = NotFound desc = could not find container \"8d9a7bafabd44f8b88a6f62cbe45599af8563bc45d9d7ad3a7669f3d8d11eb9c\": container with ID starting with 8d9a7bafabd44f8b88a6f62cbe45599af8563bc45d9d7ad3a7669f3d8d11eb9c not found: ID does not exist" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.615183 4754 scope.go:117] "RemoveContainer" containerID="0d6f7aa12978cc78e269af0ca6f59f424d8898f692c6f8d07454a1868b0204cd" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.615433 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0d6f7aa12978cc78e269af0ca6f59f424d8898f692c6f8d07454a1868b0204cd"} err="failed to get container status \"0d6f7aa12978cc78e269af0ca6f59f424d8898f692c6f8d07454a1868b0204cd\": rpc error: code = NotFound desc = could not find container \"0d6f7aa12978cc78e269af0ca6f59f424d8898f692c6f8d07454a1868b0204cd\": container with ID starting with 0d6f7aa12978cc78e269af0ca6f59f424d8898f692c6f8d07454a1868b0204cd not found: ID does not exist" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.615459 4754 scope.go:117] "RemoveContainer" containerID="07a43ff7860ff65faf51f4a5e8aeb326f06d6d7222677a92560ae2e3a0c81484" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.615836 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"07a43ff7860ff65faf51f4a5e8aeb326f06d6d7222677a92560ae2e3a0c81484"} err="failed to get container status \"07a43ff7860ff65faf51f4a5e8aeb326f06d6d7222677a92560ae2e3a0c81484\": rpc error: code = NotFound desc = could not find container \"07a43ff7860ff65faf51f4a5e8aeb326f06d6d7222677a92560ae2e3a0c81484\": container with ID starting with 07a43ff7860ff65faf51f4a5e8aeb326f06d6d7222677a92560ae2e3a0c81484 not found: ID does not exist" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.615863 4754 scope.go:117] "RemoveContainer" containerID="b1ac1f49bd564750fcff530a0f76657b0ed41573539acd020a922d19861ef834" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.616122 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b1ac1f49bd564750fcff530a0f76657b0ed41573539acd020a922d19861ef834"} err="failed to get container status \"b1ac1f49bd564750fcff530a0f76657b0ed41573539acd020a922d19861ef834\": rpc error: code = NotFound desc = could not find container \"b1ac1f49bd564750fcff530a0f76657b0ed41573539acd020a922d19861ef834\": container with ID starting with b1ac1f49bd564750fcff530a0f76657b0ed41573539acd020a922d19861ef834 not found: ID does not exist" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.616142 4754 scope.go:117] "RemoveContainer" containerID="a300a80ad87f46ba1b4fd7ad3cf6c25767fcfe966b93bdd86d729843f53a5cda" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.616361 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a300a80ad87f46ba1b4fd7ad3cf6c25767fcfe966b93bdd86d729843f53a5cda"} err="failed to get container status \"a300a80ad87f46ba1b4fd7ad3cf6c25767fcfe966b93bdd86d729843f53a5cda\": rpc error: code = NotFound desc = could not find container \"a300a80ad87f46ba1b4fd7ad3cf6c25767fcfe966b93bdd86d729843f53a5cda\": container with ID starting with a300a80ad87f46ba1b4fd7ad3cf6c25767fcfe966b93bdd86d729843f53a5cda not found: ID does not exist" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.616389 4754 scope.go:117] "RemoveContainer" containerID="919572e731813a98c7d8a94879bba931f75c472c7cd681e86d5197ebab43815d" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.616718 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"919572e731813a98c7d8a94879bba931f75c472c7cd681e86d5197ebab43815d"} err="failed to get container status \"919572e731813a98c7d8a94879bba931f75c472c7cd681e86d5197ebab43815d\": rpc error: code = NotFound desc = could not find container \"919572e731813a98c7d8a94879bba931f75c472c7cd681e86d5197ebab43815d\": container with ID starting with 919572e731813a98c7d8a94879bba931f75c472c7cd681e86d5197ebab43815d not found: ID does not exist" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.616741 4754 scope.go:117] "RemoveContainer" containerID="033f8b104d4e9ff6d5f87ca998665f12e0c56696b3d30e01ede05f4da540365d" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.616984 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"033f8b104d4e9ff6d5f87ca998665f12e0c56696b3d30e01ede05f4da540365d"} err="failed to get container status \"033f8b104d4e9ff6d5f87ca998665f12e0c56696b3d30e01ede05f4da540365d\": rpc error: code = NotFound desc = could not find container \"033f8b104d4e9ff6d5f87ca998665f12e0c56696b3d30e01ede05f4da540365d\": container with ID starting with 033f8b104d4e9ff6d5f87ca998665f12e0c56696b3d30e01ede05f4da540365d not found: ID does not exist" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.617009 4754 scope.go:117] "RemoveContainer" containerID="31dcab2f9a2d218b0e49dad1df46377ff4521592824a0fbac254d5b8e6d0632e" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.617354 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"31dcab2f9a2d218b0e49dad1df46377ff4521592824a0fbac254d5b8e6d0632e"} err="failed to get container status \"31dcab2f9a2d218b0e49dad1df46377ff4521592824a0fbac254d5b8e6d0632e\": rpc error: code = NotFound desc = could not find container \"31dcab2f9a2d218b0e49dad1df46377ff4521592824a0fbac254d5b8e6d0632e\": container with ID starting with 31dcab2f9a2d218b0e49dad1df46377ff4521592824a0fbac254d5b8e6d0632e not found: ID does not exist" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.617376 4754 scope.go:117] "RemoveContainer" containerID="3fe4d9395447ac570eca1338e749dbefea76295d427a12487b63a142a7d6a2e4" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.617684 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3fe4d9395447ac570eca1338e749dbefea76295d427a12487b63a142a7d6a2e4"} err="failed to get container status \"3fe4d9395447ac570eca1338e749dbefea76295d427a12487b63a142a7d6a2e4\": rpc error: code = NotFound desc = could not find container \"3fe4d9395447ac570eca1338e749dbefea76295d427a12487b63a142a7d6a2e4\": container with ID starting with 3fe4d9395447ac570eca1338e749dbefea76295d427a12487b63a142a7d6a2e4 not found: ID does not exist" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.617703 4754 scope.go:117] "RemoveContainer" containerID="8d9a7bafabd44f8b88a6f62cbe45599af8563bc45d9d7ad3a7669f3d8d11eb9c" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.617957 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8d9a7bafabd44f8b88a6f62cbe45599af8563bc45d9d7ad3a7669f3d8d11eb9c"} err="failed to get container status \"8d9a7bafabd44f8b88a6f62cbe45599af8563bc45d9d7ad3a7669f3d8d11eb9c\": rpc error: code = NotFound desc = could not find container \"8d9a7bafabd44f8b88a6f62cbe45599af8563bc45d9d7ad3a7669f3d8d11eb9c\": container with ID starting with 8d9a7bafabd44f8b88a6f62cbe45599af8563bc45d9d7ad3a7669f3d8d11eb9c not found: ID does not exist" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.617984 4754 scope.go:117] "RemoveContainer" containerID="0d6f7aa12978cc78e269af0ca6f59f424d8898f692c6f8d07454a1868b0204cd" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.618323 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0d6f7aa12978cc78e269af0ca6f59f424d8898f692c6f8d07454a1868b0204cd"} err="failed to get container status \"0d6f7aa12978cc78e269af0ca6f59f424d8898f692c6f8d07454a1868b0204cd\": rpc error: code = NotFound desc = could not find container \"0d6f7aa12978cc78e269af0ca6f59f424d8898f692c6f8d07454a1868b0204cd\": container with ID starting with 0d6f7aa12978cc78e269af0ca6f59f424d8898f692c6f8d07454a1868b0204cd not found: ID does not exist" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.618346 4754 scope.go:117] "RemoveContainer" containerID="07a43ff7860ff65faf51f4a5e8aeb326f06d6d7222677a92560ae2e3a0c81484" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.618717 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"07a43ff7860ff65faf51f4a5e8aeb326f06d6d7222677a92560ae2e3a0c81484"} err="failed to get container status \"07a43ff7860ff65faf51f4a5e8aeb326f06d6d7222677a92560ae2e3a0c81484\": rpc error: code = NotFound desc = could not find container \"07a43ff7860ff65faf51f4a5e8aeb326f06d6d7222677a92560ae2e3a0c81484\": container with ID starting with 07a43ff7860ff65faf51f4a5e8aeb326f06d6d7222677a92560ae2e3a0c81484 not found: ID does not exist" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.618751 4754 scope.go:117] "RemoveContainer" containerID="b1ac1f49bd564750fcff530a0f76657b0ed41573539acd020a922d19861ef834" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.619064 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b1ac1f49bd564750fcff530a0f76657b0ed41573539acd020a922d19861ef834"} err="failed to get container status \"b1ac1f49bd564750fcff530a0f76657b0ed41573539acd020a922d19861ef834\": rpc error: code = NotFound desc = could not find container \"b1ac1f49bd564750fcff530a0f76657b0ed41573539acd020a922d19861ef834\": container with ID starting with b1ac1f49bd564750fcff530a0f76657b0ed41573539acd020a922d19861ef834 not found: ID does not exist" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.619084 4754 scope.go:117] "RemoveContainer" containerID="a300a80ad87f46ba1b4fd7ad3cf6c25767fcfe966b93bdd86d729843f53a5cda" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.619332 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a300a80ad87f46ba1b4fd7ad3cf6c25767fcfe966b93bdd86d729843f53a5cda"} err="failed to get container status \"a300a80ad87f46ba1b4fd7ad3cf6c25767fcfe966b93bdd86d729843f53a5cda\": rpc error: code = NotFound desc = could not find container \"a300a80ad87f46ba1b4fd7ad3cf6c25767fcfe966b93bdd86d729843f53a5cda\": container with ID starting with a300a80ad87f46ba1b4fd7ad3cf6c25767fcfe966b93bdd86d729843f53a5cda not found: ID does not exist" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.619362 4754 scope.go:117] "RemoveContainer" containerID="919572e731813a98c7d8a94879bba931f75c472c7cd681e86d5197ebab43815d" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.619654 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"919572e731813a98c7d8a94879bba931f75c472c7cd681e86d5197ebab43815d"} err="failed to get container status \"919572e731813a98c7d8a94879bba931f75c472c7cd681e86d5197ebab43815d\": rpc error: code = NotFound desc = could not find container \"919572e731813a98c7d8a94879bba931f75c472c7cd681e86d5197ebab43815d\": container with ID starting with 919572e731813a98c7d8a94879bba931f75c472c7cd681e86d5197ebab43815d not found: ID does not exist" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.619687 4754 scope.go:117] "RemoveContainer" containerID="033f8b104d4e9ff6d5f87ca998665f12e0c56696b3d30e01ede05f4da540365d" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.619929 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"033f8b104d4e9ff6d5f87ca998665f12e0c56696b3d30e01ede05f4da540365d"} err="failed to get container status \"033f8b104d4e9ff6d5f87ca998665f12e0c56696b3d30e01ede05f4da540365d\": rpc error: code = NotFound desc = could not find container \"033f8b104d4e9ff6d5f87ca998665f12e0c56696b3d30e01ede05f4da540365d\": container with ID starting with 033f8b104d4e9ff6d5f87ca998665f12e0c56696b3d30e01ede05f4da540365d not found: ID does not exist" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.619958 4754 scope.go:117] "RemoveContainer" containerID="31dcab2f9a2d218b0e49dad1df46377ff4521592824a0fbac254d5b8e6d0632e" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.620220 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"31dcab2f9a2d218b0e49dad1df46377ff4521592824a0fbac254d5b8e6d0632e"} err="failed to get container status \"31dcab2f9a2d218b0e49dad1df46377ff4521592824a0fbac254d5b8e6d0632e\": rpc error: code = NotFound desc = could not find container \"31dcab2f9a2d218b0e49dad1df46377ff4521592824a0fbac254d5b8e6d0632e\": container with ID starting with 31dcab2f9a2d218b0e49dad1df46377ff4521592824a0fbac254d5b8e6d0632e not found: ID does not exist" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.620248 4754 scope.go:117] "RemoveContainer" containerID="3fe4d9395447ac570eca1338e749dbefea76295d427a12487b63a142a7d6a2e4" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.620569 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3fe4d9395447ac570eca1338e749dbefea76295d427a12487b63a142a7d6a2e4"} err="failed to get container status \"3fe4d9395447ac570eca1338e749dbefea76295d427a12487b63a142a7d6a2e4\": rpc error: code = NotFound desc = could not find container \"3fe4d9395447ac570eca1338e749dbefea76295d427a12487b63a142a7d6a2e4\": container with ID starting with 3fe4d9395447ac570eca1338e749dbefea76295d427a12487b63a142a7d6a2e4 not found: ID does not exist" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.620590 4754 scope.go:117] "RemoveContainer" containerID="8d9a7bafabd44f8b88a6f62cbe45599af8563bc45d9d7ad3a7669f3d8d11eb9c" Jan 26 17:00:00 crc kubenswrapper[4754]: I0126 17:00:00.620852 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8d9a7bafabd44f8b88a6f62cbe45599af8563bc45d9d7ad3a7669f3d8d11eb9c"} err="failed to get container status \"8d9a7bafabd44f8b88a6f62cbe45599af8563bc45d9d7ad3a7669f3d8d11eb9c\": rpc error: code = NotFound desc = could not find container \"8d9a7bafabd44f8b88a6f62cbe45599af8563bc45d9d7ad3a7669f3d8d11eb9c\": container with ID starting with 8d9a7bafabd44f8b88a6f62cbe45599af8563bc45d9d7ad3a7669f3d8d11eb9c not found: ID does not exist" Jan 26 17:00:01 crc kubenswrapper[4754]: I0126 17:00:01.199487 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-qln6t" Jan 26 17:00:01 crc kubenswrapper[4754]: I0126 17:00:01.199807 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-qln6t" Jan 26 17:00:01 crc kubenswrapper[4754]: I0126 17:00:01.277369 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-qln6t" Jan 26 17:00:01 crc kubenswrapper[4754]: I0126 17:00:01.434449 4754 generic.go:334] "Generic (PLEG): container finished" podID="d1ff24b0-9c9f-4ecc-a7f9-aeeb99fd7e8d" containerID="d92a7e32b06c277770987ef8f41491d7316e9669453046829e09d6c1195a3764" exitCode=0 Jan 26 17:00:01 crc kubenswrapper[4754]: I0126 17:00:01.434566 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5tc6f" event={"ID":"d1ff24b0-9c9f-4ecc-a7f9-aeeb99fd7e8d","Type":"ContainerDied","Data":"d92a7e32b06c277770987ef8f41491d7316e9669453046829e09d6c1195a3764"} Jan 26 17:00:01 crc kubenswrapper[4754]: I0126 17:00:01.496172 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-qln6t" Jan 26 17:00:01 crc kubenswrapper[4754]: I0126 17:00:01.539046 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-qln6t"] Jan 26 17:00:01 crc kubenswrapper[4754]: I0126 17:00:01.775310 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="689569c4-93ee-4b82-93fc-9af0f97e6dc3" path="/var/lib/kubelet/pods/689569c4-93ee-4b82-93fc-9af0f97e6dc3/volumes" Jan 26 17:00:02 crc kubenswrapper[4754]: I0126 17:00:02.445939 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5tc6f" event={"ID":"d1ff24b0-9c9f-4ecc-a7f9-aeeb99fd7e8d","Type":"ContainerStarted","Data":"8615ede87398349e7ff9f695c09ab2507966cf4ef66ba5d2dff2709d88cb0d09"} Jan 26 17:00:02 crc kubenswrapper[4754]: I0126 17:00:02.446275 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5tc6f" event={"ID":"d1ff24b0-9c9f-4ecc-a7f9-aeeb99fd7e8d","Type":"ContainerStarted","Data":"5abdaf8e74bed1c39d282550562f9ac75ea5784f1104c72eb49d5b2d7bda3bee"} Jan 26 17:00:02 crc kubenswrapper[4754]: I0126 17:00:02.446300 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5tc6f" event={"ID":"d1ff24b0-9c9f-4ecc-a7f9-aeeb99fd7e8d","Type":"ContainerStarted","Data":"0a5c7508ff4e96bb428044444cfac977cae9477cbc6efaaac7ef22befbfba200"} Jan 26 17:00:02 crc kubenswrapper[4754]: I0126 17:00:02.446315 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5tc6f" event={"ID":"d1ff24b0-9c9f-4ecc-a7f9-aeeb99fd7e8d","Type":"ContainerStarted","Data":"dd29b495cf62d6a5b85f3ad8c981c094f45e2ea9300c2d75109f863062d24ee3"} Jan 26 17:00:02 crc kubenswrapper[4754]: I0126 17:00:02.446334 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5tc6f" event={"ID":"d1ff24b0-9c9f-4ecc-a7f9-aeeb99fd7e8d","Type":"ContainerStarted","Data":"0a2a6322b7ccad3f1f41aacab5dcbffab3f8f5a0320ac70d5289d5c38d4bf552"} Jan 26 17:00:02 crc kubenswrapper[4754]: I0126 17:00:02.446348 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5tc6f" event={"ID":"d1ff24b0-9c9f-4ecc-a7f9-aeeb99fd7e8d","Type":"ContainerStarted","Data":"f38e6ed65ad1d0761152141eb1d419aae9427018a066121432b4d1dd046aeb76"} Jan 26 17:00:03 crc kubenswrapper[4754]: I0126 17:00:03.454196 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-qln6t" podUID="3927f5f7-ced1-41f1-8e55-ae354f995a78" containerName="registry-server" containerID="cri-o://ab293f35bec44c3b25b47a18935fc86d8b05d717ad8694b188b67cd23e850b56" gracePeriod=2 Jan 26 17:00:03 crc kubenswrapper[4754]: I0126 17:00:03.658946 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qln6t" Jan 26 17:00:03 crc kubenswrapper[4754]: I0126 17:00:03.858244 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pff5t\" (UniqueName: \"kubernetes.io/projected/3927f5f7-ced1-41f1-8e55-ae354f995a78-kube-api-access-pff5t\") pod \"3927f5f7-ced1-41f1-8e55-ae354f995a78\" (UID: \"3927f5f7-ced1-41f1-8e55-ae354f995a78\") " Jan 26 17:00:03 crc kubenswrapper[4754]: I0126 17:00:03.858321 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3927f5f7-ced1-41f1-8e55-ae354f995a78-catalog-content\") pod \"3927f5f7-ced1-41f1-8e55-ae354f995a78\" (UID: \"3927f5f7-ced1-41f1-8e55-ae354f995a78\") " Jan 26 17:00:03 crc kubenswrapper[4754]: I0126 17:00:03.858357 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3927f5f7-ced1-41f1-8e55-ae354f995a78-utilities\") pod \"3927f5f7-ced1-41f1-8e55-ae354f995a78\" (UID: \"3927f5f7-ced1-41f1-8e55-ae354f995a78\") " Jan 26 17:00:03 crc kubenswrapper[4754]: I0126 17:00:03.860077 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3927f5f7-ced1-41f1-8e55-ae354f995a78-utilities" (OuterVolumeSpecName: "utilities") pod "3927f5f7-ced1-41f1-8e55-ae354f995a78" (UID: "3927f5f7-ced1-41f1-8e55-ae354f995a78"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:00:03 crc kubenswrapper[4754]: I0126 17:00:03.867143 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3927f5f7-ced1-41f1-8e55-ae354f995a78-kube-api-access-pff5t" (OuterVolumeSpecName: "kube-api-access-pff5t") pod "3927f5f7-ced1-41f1-8e55-ae354f995a78" (UID: "3927f5f7-ced1-41f1-8e55-ae354f995a78"). InnerVolumeSpecName "kube-api-access-pff5t". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:00:03 crc kubenswrapper[4754]: I0126 17:00:03.935876 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3927f5f7-ced1-41f1-8e55-ae354f995a78-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3927f5f7-ced1-41f1-8e55-ae354f995a78" (UID: "3927f5f7-ced1-41f1-8e55-ae354f995a78"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:00:03 crc kubenswrapper[4754]: I0126 17:00:03.960359 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pff5t\" (UniqueName: \"kubernetes.io/projected/3927f5f7-ced1-41f1-8e55-ae354f995a78-kube-api-access-pff5t\") on node \"crc\" DevicePath \"\"" Jan 26 17:00:03 crc kubenswrapper[4754]: I0126 17:00:03.960540 4754 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3927f5f7-ced1-41f1-8e55-ae354f995a78-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 26 17:00:03 crc kubenswrapper[4754]: I0126 17:00:03.960730 4754 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3927f5f7-ced1-41f1-8e55-ae354f995a78-utilities\") on node \"crc\" DevicePath \"\"" Jan 26 17:00:04 crc kubenswrapper[4754]: I0126 17:00:04.132550 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-9qhgb"] Jan 26 17:00:04 crc kubenswrapper[4754]: E0126 17:00:04.133121 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3927f5f7-ced1-41f1-8e55-ae354f995a78" containerName="extract-utilities" Jan 26 17:00:04 crc kubenswrapper[4754]: I0126 17:00:04.133164 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="3927f5f7-ced1-41f1-8e55-ae354f995a78" containerName="extract-utilities" Jan 26 17:00:04 crc kubenswrapper[4754]: E0126 17:00:04.133183 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3927f5f7-ced1-41f1-8e55-ae354f995a78" containerName="extract-content" Jan 26 17:00:04 crc kubenswrapper[4754]: I0126 17:00:04.133197 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="3927f5f7-ced1-41f1-8e55-ae354f995a78" containerName="extract-content" Jan 26 17:00:04 crc kubenswrapper[4754]: E0126 17:00:04.133224 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3927f5f7-ced1-41f1-8e55-ae354f995a78" containerName="registry-server" Jan 26 17:00:04 crc kubenswrapper[4754]: I0126 17:00:04.133236 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="3927f5f7-ced1-41f1-8e55-ae354f995a78" containerName="registry-server" Jan 26 17:00:04 crc kubenswrapper[4754]: I0126 17:00:04.133436 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="3927f5f7-ced1-41f1-8e55-ae354f995a78" containerName="registry-server" Jan 26 17:00:04 crc kubenswrapper[4754]: I0126 17:00:04.134762 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9qhgb" Jan 26 17:00:04 crc kubenswrapper[4754]: I0126 17:00:04.163056 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xfgjp\" (UniqueName: \"kubernetes.io/projected/fa3e2ba5-d178-49a5-9d14-a0d4c94897ac-kube-api-access-xfgjp\") pod \"community-operators-9qhgb\" (UID: \"fa3e2ba5-d178-49a5-9d14-a0d4c94897ac\") " pod="openshift-marketplace/community-operators-9qhgb" Jan 26 17:00:04 crc kubenswrapper[4754]: I0126 17:00:04.163121 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fa3e2ba5-d178-49a5-9d14-a0d4c94897ac-catalog-content\") pod \"community-operators-9qhgb\" (UID: \"fa3e2ba5-d178-49a5-9d14-a0d4c94897ac\") " pod="openshift-marketplace/community-operators-9qhgb" Jan 26 17:00:04 crc kubenswrapper[4754]: I0126 17:00:04.163357 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fa3e2ba5-d178-49a5-9d14-a0d4c94897ac-utilities\") pod \"community-operators-9qhgb\" (UID: \"fa3e2ba5-d178-49a5-9d14-a0d4c94897ac\") " pod="openshift-marketplace/community-operators-9qhgb" Jan 26 17:00:04 crc kubenswrapper[4754]: I0126 17:00:04.263948 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fa3e2ba5-d178-49a5-9d14-a0d4c94897ac-utilities\") pod \"community-operators-9qhgb\" (UID: \"fa3e2ba5-d178-49a5-9d14-a0d4c94897ac\") " pod="openshift-marketplace/community-operators-9qhgb" Jan 26 17:00:04 crc kubenswrapper[4754]: I0126 17:00:04.264021 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xfgjp\" (UniqueName: \"kubernetes.io/projected/fa3e2ba5-d178-49a5-9d14-a0d4c94897ac-kube-api-access-xfgjp\") pod \"community-operators-9qhgb\" (UID: \"fa3e2ba5-d178-49a5-9d14-a0d4c94897ac\") " pod="openshift-marketplace/community-operators-9qhgb" Jan 26 17:00:04 crc kubenswrapper[4754]: I0126 17:00:04.264623 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fa3e2ba5-d178-49a5-9d14-a0d4c94897ac-catalog-content\") pod \"community-operators-9qhgb\" (UID: \"fa3e2ba5-d178-49a5-9d14-a0d4c94897ac\") " pod="openshift-marketplace/community-operators-9qhgb" Jan 26 17:00:04 crc kubenswrapper[4754]: I0126 17:00:04.264615 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fa3e2ba5-d178-49a5-9d14-a0d4c94897ac-utilities\") pod \"community-operators-9qhgb\" (UID: \"fa3e2ba5-d178-49a5-9d14-a0d4c94897ac\") " pod="openshift-marketplace/community-operators-9qhgb" Jan 26 17:00:04 crc kubenswrapper[4754]: I0126 17:00:04.265102 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fa3e2ba5-d178-49a5-9d14-a0d4c94897ac-catalog-content\") pod \"community-operators-9qhgb\" (UID: \"fa3e2ba5-d178-49a5-9d14-a0d4c94897ac\") " pod="openshift-marketplace/community-operators-9qhgb" Jan 26 17:00:04 crc kubenswrapper[4754]: I0126 17:00:04.283706 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xfgjp\" (UniqueName: \"kubernetes.io/projected/fa3e2ba5-d178-49a5-9d14-a0d4c94897ac-kube-api-access-xfgjp\") pod \"community-operators-9qhgb\" (UID: \"fa3e2ba5-d178-49a5-9d14-a0d4c94897ac\") " pod="openshift-marketplace/community-operators-9qhgb" Jan 26 17:00:04 crc kubenswrapper[4754]: I0126 17:00:04.457603 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9qhgb" Jan 26 17:00:04 crc kubenswrapper[4754]: I0126 17:00:04.469124 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5tc6f" event={"ID":"d1ff24b0-9c9f-4ecc-a7f9-aeeb99fd7e8d","Type":"ContainerStarted","Data":"95e10b41fffa7feb7115db7e07438cadd0b9e26fd520ff7552601a81f6d31b57"} Jan 26 17:00:04 crc kubenswrapper[4754]: I0126 17:00:04.471988 4754 generic.go:334] "Generic (PLEG): container finished" podID="3927f5f7-ced1-41f1-8e55-ae354f995a78" containerID="ab293f35bec44c3b25b47a18935fc86d8b05d717ad8694b188b67cd23e850b56" exitCode=0 Jan 26 17:00:04 crc kubenswrapper[4754]: I0126 17:00:04.472038 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qln6t" event={"ID":"3927f5f7-ced1-41f1-8e55-ae354f995a78","Type":"ContainerDied","Data":"ab293f35bec44c3b25b47a18935fc86d8b05d717ad8694b188b67cd23e850b56"} Jan 26 17:00:04 crc kubenswrapper[4754]: I0126 17:00:04.472072 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qln6t" event={"ID":"3927f5f7-ced1-41f1-8e55-ae354f995a78","Type":"ContainerDied","Data":"db2ce6e1881136437a85d570238e84e48994a28fd7cb4a9609610ac10e59a19d"} Jan 26 17:00:04 crc kubenswrapper[4754]: I0126 17:00:04.472080 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qln6t" Jan 26 17:00:04 crc kubenswrapper[4754]: I0126 17:00:04.472117 4754 scope.go:117] "RemoveContainer" containerID="ab293f35bec44c3b25b47a18935fc86d8b05d717ad8694b188b67cd23e850b56" Jan 26 17:00:04 crc kubenswrapper[4754]: I0126 17:00:04.503772 4754 scope.go:117] "RemoveContainer" containerID="5d53a53d80c48d9e41a7a78baad8344f22cd0c60fbebbb433938bd4bb06088d0" Jan 26 17:00:04 crc kubenswrapper[4754]: E0126 17:00:04.514892 4754 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_community-operators-9qhgb_openshift-marketplace_fa3e2ba5-d178-49a5-9d14-a0d4c94897ac_0(bd3d7ee60a1a21ac5b9d02dd8539c969628c47ce6da8f99abb3fae12e23cf0b1): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Jan 26 17:00:04 crc kubenswrapper[4754]: E0126 17:00:04.515065 4754 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_community-operators-9qhgb_openshift-marketplace_fa3e2ba5-d178-49a5-9d14-a0d4c94897ac_0(bd3d7ee60a1a21ac5b9d02dd8539c969628c47ce6da8f99abb3fae12e23cf0b1): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-marketplace/community-operators-9qhgb" Jan 26 17:00:04 crc kubenswrapper[4754]: E0126 17:00:04.515179 4754 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_community-operators-9qhgb_openshift-marketplace_fa3e2ba5-d178-49a5-9d14-a0d4c94897ac_0(bd3d7ee60a1a21ac5b9d02dd8539c969628c47ce6da8f99abb3fae12e23cf0b1): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-marketplace/community-operators-9qhgb" Jan 26 17:00:04 crc kubenswrapper[4754]: E0126 17:00:04.515304 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"community-operators-9qhgb_openshift-marketplace(fa3e2ba5-d178-49a5-9d14-a0d4c94897ac)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"community-operators-9qhgb_openshift-marketplace(fa3e2ba5-d178-49a5-9d14-a0d4c94897ac)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_community-operators-9qhgb_openshift-marketplace_fa3e2ba5-d178-49a5-9d14-a0d4c94897ac_0(bd3d7ee60a1a21ac5b9d02dd8539c969628c47ce6da8f99abb3fae12e23cf0b1): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-marketplace/community-operators-9qhgb" podUID="fa3e2ba5-d178-49a5-9d14-a0d4c94897ac" Jan 26 17:00:04 crc kubenswrapper[4754]: I0126 17:00:04.525863 4754 scope.go:117] "RemoveContainer" containerID="dcd2c65ef7e38d4880aefd61bcc73c9a7c22231366043739a5f0d92ac4b68325" Jan 26 17:00:04 crc kubenswrapper[4754]: I0126 17:00:04.525877 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-qln6t"] Jan 26 17:00:04 crc kubenswrapper[4754]: I0126 17:00:04.529752 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-qln6t"] Jan 26 17:00:04 crc kubenswrapper[4754]: I0126 17:00:04.548049 4754 scope.go:117] "RemoveContainer" containerID="ab293f35bec44c3b25b47a18935fc86d8b05d717ad8694b188b67cd23e850b56" Jan 26 17:00:04 crc kubenswrapper[4754]: E0126 17:00:04.550091 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ab293f35bec44c3b25b47a18935fc86d8b05d717ad8694b188b67cd23e850b56\": container with ID starting with ab293f35bec44c3b25b47a18935fc86d8b05d717ad8694b188b67cd23e850b56 not found: ID does not exist" containerID="ab293f35bec44c3b25b47a18935fc86d8b05d717ad8694b188b67cd23e850b56" Jan 26 17:00:04 crc kubenswrapper[4754]: I0126 17:00:04.550198 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ab293f35bec44c3b25b47a18935fc86d8b05d717ad8694b188b67cd23e850b56"} err="failed to get container status \"ab293f35bec44c3b25b47a18935fc86d8b05d717ad8694b188b67cd23e850b56\": rpc error: code = NotFound desc = could not find container \"ab293f35bec44c3b25b47a18935fc86d8b05d717ad8694b188b67cd23e850b56\": container with ID starting with ab293f35bec44c3b25b47a18935fc86d8b05d717ad8694b188b67cd23e850b56 not found: ID does not exist" Jan 26 17:00:04 crc kubenswrapper[4754]: I0126 17:00:04.550297 4754 scope.go:117] "RemoveContainer" containerID="5d53a53d80c48d9e41a7a78baad8344f22cd0c60fbebbb433938bd4bb06088d0" Jan 26 17:00:04 crc kubenswrapper[4754]: E0126 17:00:04.550712 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5d53a53d80c48d9e41a7a78baad8344f22cd0c60fbebbb433938bd4bb06088d0\": container with ID starting with 5d53a53d80c48d9e41a7a78baad8344f22cd0c60fbebbb433938bd4bb06088d0 not found: ID does not exist" containerID="5d53a53d80c48d9e41a7a78baad8344f22cd0c60fbebbb433938bd4bb06088d0" Jan 26 17:00:04 crc kubenswrapper[4754]: I0126 17:00:04.550757 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5d53a53d80c48d9e41a7a78baad8344f22cd0c60fbebbb433938bd4bb06088d0"} err="failed to get container status \"5d53a53d80c48d9e41a7a78baad8344f22cd0c60fbebbb433938bd4bb06088d0\": rpc error: code = NotFound desc = could not find container \"5d53a53d80c48d9e41a7a78baad8344f22cd0c60fbebbb433938bd4bb06088d0\": container with ID starting with 5d53a53d80c48d9e41a7a78baad8344f22cd0c60fbebbb433938bd4bb06088d0 not found: ID does not exist" Jan 26 17:00:04 crc kubenswrapper[4754]: I0126 17:00:04.550783 4754 scope.go:117] "RemoveContainer" containerID="dcd2c65ef7e38d4880aefd61bcc73c9a7c22231366043739a5f0d92ac4b68325" Jan 26 17:00:04 crc kubenswrapper[4754]: E0126 17:00:04.551116 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dcd2c65ef7e38d4880aefd61bcc73c9a7c22231366043739a5f0d92ac4b68325\": container with ID starting with dcd2c65ef7e38d4880aefd61bcc73c9a7c22231366043739a5f0d92ac4b68325 not found: ID does not exist" containerID="dcd2c65ef7e38d4880aefd61bcc73c9a7c22231366043739a5f0d92ac4b68325" Jan 26 17:00:04 crc kubenswrapper[4754]: I0126 17:00:04.551217 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dcd2c65ef7e38d4880aefd61bcc73c9a7c22231366043739a5f0d92ac4b68325"} err="failed to get container status \"dcd2c65ef7e38d4880aefd61bcc73c9a7c22231366043739a5f0d92ac4b68325\": rpc error: code = NotFound desc = could not find container \"dcd2c65ef7e38d4880aefd61bcc73c9a7c22231366043739a5f0d92ac4b68325\": container with ID starting with dcd2c65ef7e38d4880aefd61bcc73c9a7c22231366043739a5f0d92ac4b68325 not found: ID does not exist" Jan 26 17:00:05 crc kubenswrapper[4754]: I0126 17:00:05.778361 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3927f5f7-ced1-41f1-8e55-ae354f995a78" path="/var/lib/kubelet/pods/3927f5f7-ced1-41f1-8e55-ae354f995a78/volumes" Jan 26 17:00:07 crc kubenswrapper[4754]: I0126 17:00:07.129770 4754 patch_prober.go:28] interesting pod/machine-config-daemon-x65wv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 26 17:00:07 crc kubenswrapper[4754]: I0126 17:00:07.130063 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 26 17:00:08 crc kubenswrapper[4754]: I0126 17:00:08.500459 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-9qhgb"] Jan 26 17:00:08 crc kubenswrapper[4754]: I0126 17:00:08.501042 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9qhgb" Jan 26 17:00:08 crc kubenswrapper[4754]: I0126 17:00:08.501716 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9qhgb" Jan 26 17:00:08 crc kubenswrapper[4754]: I0126 17:00:08.517417 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29490780-7wcg8"] Jan 26 17:00:08 crc kubenswrapper[4754]: I0126 17:00:08.517577 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29490780-7wcg8" Jan 26 17:00:08 crc kubenswrapper[4754]: I0126 17:00:08.518181 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29490780-7wcg8" Jan 26 17:00:08 crc kubenswrapper[4754]: I0126 17:00:08.548442 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5tc6f" event={"ID":"d1ff24b0-9c9f-4ecc-a7f9-aeeb99fd7e8d","Type":"ContainerStarted","Data":"bcafbe6e696b4be2ff373948176dfcaf787fb01123ac62c1487c50c51183b2f4"} Jan 26 17:00:08 crc kubenswrapper[4754]: I0126 17:00:08.549687 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-5tc6f" Jan 26 17:00:08 crc kubenswrapper[4754]: I0126 17:00:08.549725 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-5tc6f" Jan 26 17:00:08 crc kubenswrapper[4754]: I0126 17:00:08.549779 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-5tc6f" Jan 26 17:00:08 crc kubenswrapper[4754]: E0126 17:00:08.581953 4754 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_community-operators-9qhgb_openshift-marketplace_fa3e2ba5-d178-49a5-9d14-a0d4c94897ac_0(e21c9ed6195dc529b023a96f89cef2762aedcf00e9afa0c6719752c156abdad2): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Jan 26 17:00:08 crc kubenswrapper[4754]: E0126 17:00:08.582035 4754 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_community-operators-9qhgb_openshift-marketplace_fa3e2ba5-d178-49a5-9d14-a0d4c94897ac_0(e21c9ed6195dc529b023a96f89cef2762aedcf00e9afa0c6719752c156abdad2): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-marketplace/community-operators-9qhgb" Jan 26 17:00:08 crc kubenswrapper[4754]: E0126 17:00:08.582062 4754 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_community-operators-9qhgb_openshift-marketplace_fa3e2ba5-d178-49a5-9d14-a0d4c94897ac_0(e21c9ed6195dc529b023a96f89cef2762aedcf00e9afa0c6719752c156abdad2): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-marketplace/community-operators-9qhgb" Jan 26 17:00:08 crc kubenswrapper[4754]: E0126 17:00:08.582115 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"community-operators-9qhgb_openshift-marketplace(fa3e2ba5-d178-49a5-9d14-a0d4c94897ac)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"community-operators-9qhgb_openshift-marketplace(fa3e2ba5-d178-49a5-9d14-a0d4c94897ac)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_community-operators-9qhgb_openshift-marketplace_fa3e2ba5-d178-49a5-9d14-a0d4c94897ac_0(e21c9ed6195dc529b023a96f89cef2762aedcf00e9afa0c6719752c156abdad2): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-marketplace/community-operators-9qhgb" podUID="fa3e2ba5-d178-49a5-9d14-a0d4c94897ac" Jan 26 17:00:08 crc kubenswrapper[4754]: E0126 17:00:08.613041 4754 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_collect-profiles-29490780-7wcg8_openshift-operator-lifecycle-manager_4f2d7a6e-760a-4cd5-b9d0-7483d4f12933_0(e44fa03d29814cb6ee053561e5c3e2fe1ad18e3725e16ac8ca978810e84b9912): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Jan 26 17:00:08 crc kubenswrapper[4754]: E0126 17:00:08.613117 4754 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_collect-profiles-29490780-7wcg8_openshift-operator-lifecycle-manager_4f2d7a6e-760a-4cd5-b9d0-7483d4f12933_0(e44fa03d29814cb6ee053561e5c3e2fe1ad18e3725e16ac8ca978810e84b9912): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operator-lifecycle-manager/collect-profiles-29490780-7wcg8" Jan 26 17:00:08 crc kubenswrapper[4754]: E0126 17:00:08.613145 4754 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_collect-profiles-29490780-7wcg8_openshift-operator-lifecycle-manager_4f2d7a6e-760a-4cd5-b9d0-7483d4f12933_0(e44fa03d29814cb6ee053561e5c3e2fe1ad18e3725e16ac8ca978810e84b9912): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operator-lifecycle-manager/collect-profiles-29490780-7wcg8" Jan 26 17:00:08 crc kubenswrapper[4754]: E0126 17:00:08.613194 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"collect-profiles-29490780-7wcg8_openshift-operator-lifecycle-manager(4f2d7a6e-760a-4cd5-b9d0-7483d4f12933)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"collect-profiles-29490780-7wcg8_openshift-operator-lifecycle-manager(4f2d7a6e-760a-4cd5-b9d0-7483d4f12933)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_collect-profiles-29490780-7wcg8_openshift-operator-lifecycle-manager_4f2d7a6e-760a-4cd5-b9d0-7483d4f12933_0(e44fa03d29814cb6ee053561e5c3e2fe1ad18e3725e16ac8ca978810e84b9912): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operator-lifecycle-manager/collect-profiles-29490780-7wcg8" podUID="4f2d7a6e-760a-4cd5-b9d0-7483d4f12933" Jan 26 17:00:08 crc kubenswrapper[4754]: I0126 17:00:08.615114 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-5tc6f" podStartSLOduration=9.615080577 podStartE2EDuration="9.615080577s" podCreationTimestamp="2026-01-26 16:59:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 17:00:08.613801152 +0000 UTC m=+775.137981606" watchObservedRunningTime="2026-01-26 17:00:08.615080577 +0000 UTC m=+775.139261021" Jan 26 17:00:08 crc kubenswrapper[4754]: I0126 17:00:08.619185 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-5tc6f" Jan 26 17:00:08 crc kubenswrapper[4754]: I0126 17:00:08.626521 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-5tc6f" Jan 26 17:00:22 crc kubenswrapper[4754]: I0126 17:00:22.766378 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9qhgb" Jan 26 17:00:22 crc kubenswrapper[4754]: I0126 17:00:22.766626 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29490780-7wcg8" Jan 26 17:00:22 crc kubenswrapper[4754]: I0126 17:00:22.767714 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9qhgb" Jan 26 17:00:22 crc kubenswrapper[4754]: I0126 17:00:22.767717 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29490780-7wcg8" Jan 26 17:00:22 crc kubenswrapper[4754]: I0126 17:00:22.995752 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29490780-7wcg8"] Jan 26 17:00:23 crc kubenswrapper[4754]: W0126 17:00:23.004239 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4f2d7a6e_760a_4cd5_b9d0_7483d4f12933.slice/crio-fc5b1c8aea0416186f020c9948fb16a221dda8041329b18c9491b94c0ab5498f WatchSource:0}: Error finding container fc5b1c8aea0416186f020c9948fb16a221dda8041329b18c9491b94c0ab5498f: Status 404 returned error can't find the container with id fc5b1c8aea0416186f020c9948fb16a221dda8041329b18c9491b94c0ab5498f Jan 26 17:00:23 crc kubenswrapper[4754]: I0126 17:00:23.032646 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-9qhgb"] Jan 26 17:00:23 crc kubenswrapper[4754]: I0126 17:00:23.662080 4754 generic.go:334] "Generic (PLEG): container finished" podID="4f2d7a6e-760a-4cd5-b9d0-7483d4f12933" containerID="b6c77d399b6d69676f8c557d43ec9baff18e7851b914517948580ae3bb687f5a" exitCode=0 Jan 26 17:00:23 crc kubenswrapper[4754]: I0126 17:00:23.662162 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29490780-7wcg8" event={"ID":"4f2d7a6e-760a-4cd5-b9d0-7483d4f12933","Type":"ContainerDied","Data":"b6c77d399b6d69676f8c557d43ec9baff18e7851b914517948580ae3bb687f5a"} Jan 26 17:00:23 crc kubenswrapper[4754]: I0126 17:00:23.662451 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29490780-7wcg8" event={"ID":"4f2d7a6e-760a-4cd5-b9d0-7483d4f12933","Type":"ContainerStarted","Data":"fc5b1c8aea0416186f020c9948fb16a221dda8041329b18c9491b94c0ab5498f"} Jan 26 17:00:23 crc kubenswrapper[4754]: I0126 17:00:23.664701 4754 generic.go:334] "Generic (PLEG): container finished" podID="fa3e2ba5-d178-49a5-9d14-a0d4c94897ac" containerID="18fee32c82bd95f61a6d1b7d4ab9e93de5f45ace29da923a370e4b3487b20667" exitCode=0 Jan 26 17:00:23 crc kubenswrapper[4754]: I0126 17:00:23.664738 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9qhgb" event={"ID":"fa3e2ba5-d178-49a5-9d14-a0d4c94897ac","Type":"ContainerDied","Data":"18fee32c82bd95f61a6d1b7d4ab9e93de5f45ace29da923a370e4b3487b20667"} Jan 26 17:00:23 crc kubenswrapper[4754]: I0126 17:00:23.664754 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9qhgb" event={"ID":"fa3e2ba5-d178-49a5-9d14-a0d4c94897ac","Type":"ContainerStarted","Data":"ce79dd1874db819219f5b33738cb4c5afb8b2b430f34b017b572718977eb0baf"} Jan 26 17:00:24 crc kubenswrapper[4754]: I0126 17:00:24.683281 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9qhgb" event={"ID":"fa3e2ba5-d178-49a5-9d14-a0d4c94897ac","Type":"ContainerStarted","Data":"72a74560d825577e7d5ad81f301eb8d2f351333ca2bcb64ecad15825c5784edd"} Jan 26 17:00:24 crc kubenswrapper[4754]: I0126 17:00:24.931755 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29490780-7wcg8" Jan 26 17:00:25 crc kubenswrapper[4754]: I0126 17:00:25.037121 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4f2d7a6e-760a-4cd5-b9d0-7483d4f12933-config-volume\") pod \"4f2d7a6e-760a-4cd5-b9d0-7483d4f12933\" (UID: \"4f2d7a6e-760a-4cd5-b9d0-7483d4f12933\") " Jan 26 17:00:25 crc kubenswrapper[4754]: I0126 17:00:25.037231 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6dkbk\" (UniqueName: \"kubernetes.io/projected/4f2d7a6e-760a-4cd5-b9d0-7483d4f12933-kube-api-access-6dkbk\") pod \"4f2d7a6e-760a-4cd5-b9d0-7483d4f12933\" (UID: \"4f2d7a6e-760a-4cd5-b9d0-7483d4f12933\") " Jan 26 17:00:25 crc kubenswrapper[4754]: I0126 17:00:25.037321 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4f2d7a6e-760a-4cd5-b9d0-7483d4f12933-secret-volume\") pod \"4f2d7a6e-760a-4cd5-b9d0-7483d4f12933\" (UID: \"4f2d7a6e-760a-4cd5-b9d0-7483d4f12933\") " Jan 26 17:00:25 crc kubenswrapper[4754]: I0126 17:00:25.037901 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4f2d7a6e-760a-4cd5-b9d0-7483d4f12933-config-volume" (OuterVolumeSpecName: "config-volume") pod "4f2d7a6e-760a-4cd5-b9d0-7483d4f12933" (UID: "4f2d7a6e-760a-4cd5-b9d0-7483d4f12933"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:00:25 crc kubenswrapper[4754]: I0126 17:00:25.042618 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4f2d7a6e-760a-4cd5-b9d0-7483d4f12933-kube-api-access-6dkbk" (OuterVolumeSpecName: "kube-api-access-6dkbk") pod "4f2d7a6e-760a-4cd5-b9d0-7483d4f12933" (UID: "4f2d7a6e-760a-4cd5-b9d0-7483d4f12933"). InnerVolumeSpecName "kube-api-access-6dkbk". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:00:25 crc kubenswrapper[4754]: I0126 17:00:25.043689 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4f2d7a6e-760a-4cd5-b9d0-7483d4f12933-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "4f2d7a6e-760a-4cd5-b9d0-7483d4f12933" (UID: "4f2d7a6e-760a-4cd5-b9d0-7483d4f12933"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:00:25 crc kubenswrapper[4754]: I0126 17:00:25.138286 4754 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4f2d7a6e-760a-4cd5-b9d0-7483d4f12933-secret-volume\") on node \"crc\" DevicePath \"\"" Jan 26 17:00:25 crc kubenswrapper[4754]: I0126 17:00:25.138326 4754 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4f2d7a6e-760a-4cd5-b9d0-7483d4f12933-config-volume\") on node \"crc\" DevicePath \"\"" Jan 26 17:00:25 crc kubenswrapper[4754]: I0126 17:00:25.138338 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6dkbk\" (UniqueName: \"kubernetes.io/projected/4f2d7a6e-760a-4cd5-b9d0-7483d4f12933-kube-api-access-6dkbk\") on node \"crc\" DevicePath \"\"" Jan 26 17:00:25 crc kubenswrapper[4754]: I0126 17:00:25.690747 4754 generic.go:334] "Generic (PLEG): container finished" podID="fa3e2ba5-d178-49a5-9d14-a0d4c94897ac" containerID="72a74560d825577e7d5ad81f301eb8d2f351333ca2bcb64ecad15825c5784edd" exitCode=0 Jan 26 17:00:25 crc kubenswrapper[4754]: I0126 17:00:25.690796 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9qhgb" event={"ID":"fa3e2ba5-d178-49a5-9d14-a0d4c94897ac","Type":"ContainerDied","Data":"72a74560d825577e7d5ad81f301eb8d2f351333ca2bcb64ecad15825c5784edd"} Jan 26 17:00:25 crc kubenswrapper[4754]: I0126 17:00:25.693951 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29490780-7wcg8" event={"ID":"4f2d7a6e-760a-4cd5-b9d0-7483d4f12933","Type":"ContainerDied","Data":"fc5b1c8aea0416186f020c9948fb16a221dda8041329b18c9491b94c0ab5498f"} Jan 26 17:00:25 crc kubenswrapper[4754]: I0126 17:00:25.694016 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fc5b1c8aea0416186f020c9948fb16a221dda8041329b18c9491b94c0ab5498f" Jan 26 17:00:25 crc kubenswrapper[4754]: I0126 17:00:25.693977 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29490780-7wcg8" Jan 26 17:00:26 crc kubenswrapper[4754]: I0126 17:00:26.706004 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9qhgb" event={"ID":"fa3e2ba5-d178-49a5-9d14-a0d4c94897ac","Type":"ContainerStarted","Data":"01bd00463d28d749c230b1245f98a9ef85682c47aa7c795cdb4b6669f79c80f1"} Jan 26 17:00:26 crc kubenswrapper[4754]: I0126 17:00:26.721742 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-9qhgb" podStartSLOduration=20.218461397 podStartE2EDuration="22.721721117s" podCreationTimestamp="2026-01-26 17:00:04 +0000 UTC" firstStartedPulling="2026-01-26 17:00:23.667261661 +0000 UTC m=+790.191442105" lastFinishedPulling="2026-01-26 17:00:26.170521351 +0000 UTC m=+792.694701825" observedRunningTime="2026-01-26 17:00:26.721638305 +0000 UTC m=+793.245818769" watchObservedRunningTime="2026-01-26 17:00:26.721721117 +0000 UTC m=+793.245901571" Jan 26 17:00:30 crc kubenswrapper[4754]: I0126 17:00:30.327189 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-5tc6f" Jan 26 17:00:34 crc kubenswrapper[4754]: I0126 17:00:34.459203 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-9qhgb" Jan 26 17:00:34 crc kubenswrapper[4754]: I0126 17:00:34.460095 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-9qhgb" Jan 26 17:00:34 crc kubenswrapper[4754]: I0126 17:00:34.510439 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-9qhgb" Jan 26 17:00:34 crc kubenswrapper[4754]: I0126 17:00:34.806649 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-9qhgb" Jan 26 17:00:35 crc kubenswrapper[4754]: I0126 17:00:35.330368 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-9qhgb"] Jan 26 17:00:36 crc kubenswrapper[4754]: I0126 17:00:36.773042 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-9qhgb" podUID="fa3e2ba5-d178-49a5-9d14-a0d4c94897ac" containerName="registry-server" containerID="cri-o://01bd00463d28d749c230b1245f98a9ef85682c47aa7c795cdb4b6669f79c80f1" gracePeriod=2 Jan 26 17:00:37 crc kubenswrapper[4754]: I0126 17:00:37.129468 4754 patch_prober.go:28] interesting pod/machine-config-daemon-x65wv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 26 17:00:37 crc kubenswrapper[4754]: I0126 17:00:37.129991 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 26 17:00:37 crc kubenswrapper[4754]: I0126 17:00:37.130222 4754 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" Jan 26 17:00:37 crc kubenswrapper[4754]: I0126 17:00:37.131311 4754 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"2258d22dac1a41453a2ed61fe65a0c8edeaa3d706e7abb2b8003f3eba6c44d91"} pod="openshift-machine-config-operator/machine-config-daemon-x65wv" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 26 17:00:37 crc kubenswrapper[4754]: I0126 17:00:37.131805 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" containerName="machine-config-daemon" containerID="cri-o://2258d22dac1a41453a2ed61fe65a0c8edeaa3d706e7abb2b8003f3eba6c44d91" gracePeriod=600 Jan 26 17:00:37 crc kubenswrapper[4754]: I0126 17:00:37.665075 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9qhgb" Jan 26 17:00:37 crc kubenswrapper[4754]: I0126 17:00:37.702583 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fa3e2ba5-d178-49a5-9d14-a0d4c94897ac-catalog-content\") pod \"fa3e2ba5-d178-49a5-9d14-a0d4c94897ac\" (UID: \"fa3e2ba5-d178-49a5-9d14-a0d4c94897ac\") " Jan 26 17:00:37 crc kubenswrapper[4754]: I0126 17:00:37.702656 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xfgjp\" (UniqueName: \"kubernetes.io/projected/fa3e2ba5-d178-49a5-9d14-a0d4c94897ac-kube-api-access-xfgjp\") pod \"fa3e2ba5-d178-49a5-9d14-a0d4c94897ac\" (UID: \"fa3e2ba5-d178-49a5-9d14-a0d4c94897ac\") " Jan 26 17:00:37 crc kubenswrapper[4754]: I0126 17:00:37.702747 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fa3e2ba5-d178-49a5-9d14-a0d4c94897ac-utilities\") pod \"fa3e2ba5-d178-49a5-9d14-a0d4c94897ac\" (UID: \"fa3e2ba5-d178-49a5-9d14-a0d4c94897ac\") " Jan 26 17:00:37 crc kubenswrapper[4754]: I0126 17:00:37.703714 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fa3e2ba5-d178-49a5-9d14-a0d4c94897ac-utilities" (OuterVolumeSpecName: "utilities") pod "fa3e2ba5-d178-49a5-9d14-a0d4c94897ac" (UID: "fa3e2ba5-d178-49a5-9d14-a0d4c94897ac"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:00:37 crc kubenswrapper[4754]: I0126 17:00:37.709841 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fa3e2ba5-d178-49a5-9d14-a0d4c94897ac-kube-api-access-xfgjp" (OuterVolumeSpecName: "kube-api-access-xfgjp") pod "fa3e2ba5-d178-49a5-9d14-a0d4c94897ac" (UID: "fa3e2ba5-d178-49a5-9d14-a0d4c94897ac"). InnerVolumeSpecName "kube-api-access-xfgjp". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:00:37 crc kubenswrapper[4754]: I0126 17:00:37.751493 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fa3e2ba5-d178-49a5-9d14-a0d4c94897ac-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "fa3e2ba5-d178-49a5-9d14-a0d4c94897ac" (UID: "fa3e2ba5-d178-49a5-9d14-a0d4c94897ac"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:00:37 crc kubenswrapper[4754]: I0126 17:00:37.780715 4754 generic.go:334] "Generic (PLEG): container finished" podID="fa3e2ba5-d178-49a5-9d14-a0d4c94897ac" containerID="01bd00463d28d749c230b1245f98a9ef85682c47aa7c795cdb4b6669f79c80f1" exitCode=0 Jan 26 17:00:37 crc kubenswrapper[4754]: I0126 17:00:37.780791 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9qhgb" Jan 26 17:00:37 crc kubenswrapper[4754]: I0126 17:00:37.780790 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9qhgb" event={"ID":"fa3e2ba5-d178-49a5-9d14-a0d4c94897ac","Type":"ContainerDied","Data":"01bd00463d28d749c230b1245f98a9ef85682c47aa7c795cdb4b6669f79c80f1"} Jan 26 17:00:37 crc kubenswrapper[4754]: I0126 17:00:37.781627 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9qhgb" event={"ID":"fa3e2ba5-d178-49a5-9d14-a0d4c94897ac","Type":"ContainerDied","Data":"ce79dd1874db819219f5b33738cb4c5afb8b2b430f34b017b572718977eb0baf"} Jan 26 17:00:37 crc kubenswrapper[4754]: I0126 17:00:37.781655 4754 scope.go:117] "RemoveContainer" containerID="01bd00463d28d749c230b1245f98a9ef85682c47aa7c795cdb4b6669f79c80f1" Jan 26 17:00:37 crc kubenswrapper[4754]: I0126 17:00:37.785848 4754 generic.go:334] "Generic (PLEG): container finished" podID="8c3718a4-f354-4284-92e0-fdfb45a692bd" containerID="2258d22dac1a41453a2ed61fe65a0c8edeaa3d706e7abb2b8003f3eba6c44d91" exitCode=0 Jan 26 17:00:37 crc kubenswrapper[4754]: I0126 17:00:37.785897 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" event={"ID":"8c3718a4-f354-4284-92e0-fdfb45a692bd","Type":"ContainerDied","Data":"2258d22dac1a41453a2ed61fe65a0c8edeaa3d706e7abb2b8003f3eba6c44d91"} Jan 26 17:00:37 crc kubenswrapper[4754]: I0126 17:00:37.785930 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" event={"ID":"8c3718a4-f354-4284-92e0-fdfb45a692bd","Type":"ContainerStarted","Data":"2d157bed70bdc2504afc3419bb56bdffbceb22c7b7e8e1c5315d44974048523d"} Jan 26 17:00:37 crc kubenswrapper[4754]: I0126 17:00:37.801366 4754 scope.go:117] "RemoveContainer" containerID="72a74560d825577e7d5ad81f301eb8d2f351333ca2bcb64ecad15825c5784edd" Jan 26 17:00:37 crc kubenswrapper[4754]: I0126 17:00:37.803554 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xfgjp\" (UniqueName: \"kubernetes.io/projected/fa3e2ba5-d178-49a5-9d14-a0d4c94897ac-kube-api-access-xfgjp\") on node \"crc\" DevicePath \"\"" Jan 26 17:00:37 crc kubenswrapper[4754]: I0126 17:00:37.803604 4754 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fa3e2ba5-d178-49a5-9d14-a0d4c94897ac-utilities\") on node \"crc\" DevicePath \"\"" Jan 26 17:00:37 crc kubenswrapper[4754]: I0126 17:00:37.803660 4754 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fa3e2ba5-d178-49a5-9d14-a0d4c94897ac-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 26 17:00:37 crc kubenswrapper[4754]: I0126 17:00:37.814766 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-9qhgb"] Jan 26 17:00:37 crc kubenswrapper[4754]: I0126 17:00:37.817583 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-9qhgb"] Jan 26 17:00:37 crc kubenswrapper[4754]: I0126 17:00:37.821077 4754 scope.go:117] "RemoveContainer" containerID="18fee32c82bd95f61a6d1b7d4ab9e93de5f45ace29da923a370e4b3487b20667" Jan 26 17:00:37 crc kubenswrapper[4754]: I0126 17:00:37.833978 4754 scope.go:117] "RemoveContainer" containerID="01bd00463d28d749c230b1245f98a9ef85682c47aa7c795cdb4b6669f79c80f1" Jan 26 17:00:37 crc kubenswrapper[4754]: E0126 17:00:37.834574 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"01bd00463d28d749c230b1245f98a9ef85682c47aa7c795cdb4b6669f79c80f1\": container with ID starting with 01bd00463d28d749c230b1245f98a9ef85682c47aa7c795cdb4b6669f79c80f1 not found: ID does not exist" containerID="01bd00463d28d749c230b1245f98a9ef85682c47aa7c795cdb4b6669f79c80f1" Jan 26 17:00:37 crc kubenswrapper[4754]: I0126 17:00:37.834614 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"01bd00463d28d749c230b1245f98a9ef85682c47aa7c795cdb4b6669f79c80f1"} err="failed to get container status \"01bd00463d28d749c230b1245f98a9ef85682c47aa7c795cdb4b6669f79c80f1\": rpc error: code = NotFound desc = could not find container \"01bd00463d28d749c230b1245f98a9ef85682c47aa7c795cdb4b6669f79c80f1\": container with ID starting with 01bd00463d28d749c230b1245f98a9ef85682c47aa7c795cdb4b6669f79c80f1 not found: ID does not exist" Jan 26 17:00:37 crc kubenswrapper[4754]: I0126 17:00:37.834657 4754 scope.go:117] "RemoveContainer" containerID="72a74560d825577e7d5ad81f301eb8d2f351333ca2bcb64ecad15825c5784edd" Jan 26 17:00:37 crc kubenswrapper[4754]: E0126 17:00:37.835014 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"72a74560d825577e7d5ad81f301eb8d2f351333ca2bcb64ecad15825c5784edd\": container with ID starting with 72a74560d825577e7d5ad81f301eb8d2f351333ca2bcb64ecad15825c5784edd not found: ID does not exist" containerID="72a74560d825577e7d5ad81f301eb8d2f351333ca2bcb64ecad15825c5784edd" Jan 26 17:00:37 crc kubenswrapper[4754]: I0126 17:00:37.835052 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"72a74560d825577e7d5ad81f301eb8d2f351333ca2bcb64ecad15825c5784edd"} err="failed to get container status \"72a74560d825577e7d5ad81f301eb8d2f351333ca2bcb64ecad15825c5784edd\": rpc error: code = NotFound desc = could not find container \"72a74560d825577e7d5ad81f301eb8d2f351333ca2bcb64ecad15825c5784edd\": container with ID starting with 72a74560d825577e7d5ad81f301eb8d2f351333ca2bcb64ecad15825c5784edd not found: ID does not exist" Jan 26 17:00:37 crc kubenswrapper[4754]: I0126 17:00:37.835079 4754 scope.go:117] "RemoveContainer" containerID="18fee32c82bd95f61a6d1b7d4ab9e93de5f45ace29da923a370e4b3487b20667" Jan 26 17:00:37 crc kubenswrapper[4754]: E0126 17:00:37.835431 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"18fee32c82bd95f61a6d1b7d4ab9e93de5f45ace29da923a370e4b3487b20667\": container with ID starting with 18fee32c82bd95f61a6d1b7d4ab9e93de5f45ace29da923a370e4b3487b20667 not found: ID does not exist" containerID="18fee32c82bd95f61a6d1b7d4ab9e93de5f45ace29da923a370e4b3487b20667" Jan 26 17:00:37 crc kubenswrapper[4754]: I0126 17:00:37.835461 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"18fee32c82bd95f61a6d1b7d4ab9e93de5f45ace29da923a370e4b3487b20667"} err="failed to get container status \"18fee32c82bd95f61a6d1b7d4ab9e93de5f45ace29da923a370e4b3487b20667\": rpc error: code = NotFound desc = could not find container \"18fee32c82bd95f61a6d1b7d4ab9e93de5f45ace29da923a370e4b3487b20667\": container with ID starting with 18fee32c82bd95f61a6d1b7d4ab9e93de5f45ace29da923a370e4b3487b20667 not found: ID does not exist" Jan 26 17:00:37 crc kubenswrapper[4754]: I0126 17:00:37.835476 4754 scope.go:117] "RemoveContainer" containerID="7ba829a1fdd77f6bde538408ab51fc99c1f39952e719ae1ed5b880b92ca7240c" Jan 26 17:00:39 crc kubenswrapper[4754]: I0126 17:00:39.779769 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fa3e2ba5-d178-49a5-9d14-a0d4c94897ac" path="/var/lib/kubelet/pods/fa3e2ba5-d178-49a5-9d14-a0d4c94897ac/volumes" Jan 26 17:00:43 crc kubenswrapper[4754]: I0126 17:00:43.188245 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713v2nvl"] Jan 26 17:00:43 crc kubenswrapper[4754]: E0126 17:00:43.188783 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa3e2ba5-d178-49a5-9d14-a0d4c94897ac" containerName="extract-content" Jan 26 17:00:43 crc kubenswrapper[4754]: I0126 17:00:43.188798 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa3e2ba5-d178-49a5-9d14-a0d4c94897ac" containerName="extract-content" Jan 26 17:00:43 crc kubenswrapper[4754]: E0126 17:00:43.188819 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4f2d7a6e-760a-4cd5-b9d0-7483d4f12933" containerName="collect-profiles" Jan 26 17:00:43 crc kubenswrapper[4754]: I0126 17:00:43.188826 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="4f2d7a6e-760a-4cd5-b9d0-7483d4f12933" containerName="collect-profiles" Jan 26 17:00:43 crc kubenswrapper[4754]: E0126 17:00:43.188847 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa3e2ba5-d178-49a5-9d14-a0d4c94897ac" containerName="registry-server" Jan 26 17:00:43 crc kubenswrapper[4754]: I0126 17:00:43.188855 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa3e2ba5-d178-49a5-9d14-a0d4c94897ac" containerName="registry-server" Jan 26 17:00:43 crc kubenswrapper[4754]: E0126 17:00:43.188867 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa3e2ba5-d178-49a5-9d14-a0d4c94897ac" containerName="extract-utilities" Jan 26 17:00:43 crc kubenswrapper[4754]: I0126 17:00:43.188874 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa3e2ba5-d178-49a5-9d14-a0d4c94897ac" containerName="extract-utilities" Jan 26 17:00:43 crc kubenswrapper[4754]: I0126 17:00:43.188972 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="4f2d7a6e-760a-4cd5-b9d0-7483d4f12933" containerName="collect-profiles" Jan 26 17:00:43 crc kubenswrapper[4754]: I0126 17:00:43.188989 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="fa3e2ba5-d178-49a5-9d14-a0d4c94897ac" containerName="registry-server" Jan 26 17:00:43 crc kubenswrapper[4754]: I0126 17:00:43.189834 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713v2nvl" Jan 26 17:00:43 crc kubenswrapper[4754]: I0126 17:00:43.193820 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Jan 26 17:00:43 crc kubenswrapper[4754]: I0126 17:00:43.204976 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713v2nvl"] Jan 26 17:00:43 crc kubenswrapper[4754]: I0126 17:00:43.278407 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4212a345-63c7-4b01-b13d-12f9a3fc297a-bundle\") pod \"53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713v2nvl\" (UID: \"4212a345-63c7-4b01-b13d-12f9a3fc297a\") " pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713v2nvl" Jan 26 17:00:43 crc kubenswrapper[4754]: I0126 17:00:43.278896 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4212a345-63c7-4b01-b13d-12f9a3fc297a-util\") pod \"53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713v2nvl\" (UID: \"4212a345-63c7-4b01-b13d-12f9a3fc297a\") " pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713v2nvl" Jan 26 17:00:43 crc kubenswrapper[4754]: I0126 17:00:43.279041 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lr42t\" (UniqueName: \"kubernetes.io/projected/4212a345-63c7-4b01-b13d-12f9a3fc297a-kube-api-access-lr42t\") pod \"53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713v2nvl\" (UID: \"4212a345-63c7-4b01-b13d-12f9a3fc297a\") " pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713v2nvl" Jan 26 17:00:43 crc kubenswrapper[4754]: I0126 17:00:43.380311 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4212a345-63c7-4b01-b13d-12f9a3fc297a-util\") pod \"53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713v2nvl\" (UID: \"4212a345-63c7-4b01-b13d-12f9a3fc297a\") " pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713v2nvl" Jan 26 17:00:43 crc kubenswrapper[4754]: I0126 17:00:43.380384 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lr42t\" (UniqueName: \"kubernetes.io/projected/4212a345-63c7-4b01-b13d-12f9a3fc297a-kube-api-access-lr42t\") pod \"53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713v2nvl\" (UID: \"4212a345-63c7-4b01-b13d-12f9a3fc297a\") " pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713v2nvl" Jan 26 17:00:43 crc kubenswrapper[4754]: I0126 17:00:43.380698 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4212a345-63c7-4b01-b13d-12f9a3fc297a-bundle\") pod \"53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713v2nvl\" (UID: \"4212a345-63c7-4b01-b13d-12f9a3fc297a\") " pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713v2nvl" Jan 26 17:00:43 crc kubenswrapper[4754]: I0126 17:00:43.381165 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4212a345-63c7-4b01-b13d-12f9a3fc297a-util\") pod \"53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713v2nvl\" (UID: \"4212a345-63c7-4b01-b13d-12f9a3fc297a\") " pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713v2nvl" Jan 26 17:00:43 crc kubenswrapper[4754]: I0126 17:00:43.381538 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4212a345-63c7-4b01-b13d-12f9a3fc297a-bundle\") pod \"53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713v2nvl\" (UID: \"4212a345-63c7-4b01-b13d-12f9a3fc297a\") " pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713v2nvl" Jan 26 17:00:43 crc kubenswrapper[4754]: I0126 17:00:43.407657 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lr42t\" (UniqueName: \"kubernetes.io/projected/4212a345-63c7-4b01-b13d-12f9a3fc297a-kube-api-access-lr42t\") pod \"53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713v2nvl\" (UID: \"4212a345-63c7-4b01-b13d-12f9a3fc297a\") " pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713v2nvl" Jan 26 17:00:43 crc kubenswrapper[4754]: I0126 17:00:43.512057 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713v2nvl" Jan 26 17:00:43 crc kubenswrapper[4754]: I0126 17:00:43.976473 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713v2nvl"] Jan 26 17:00:43 crc kubenswrapper[4754]: W0126 17:00:43.987027 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4212a345_63c7_4b01_b13d_12f9a3fc297a.slice/crio-8d51c58dfeef4bc4e89c3e465625b813c8ae33e34854c11014d8d24c6c650124 WatchSource:0}: Error finding container 8d51c58dfeef4bc4e89c3e465625b813c8ae33e34854c11014d8d24c6c650124: Status 404 returned error can't find the container with id 8d51c58dfeef4bc4e89c3e465625b813c8ae33e34854c11014d8d24c6c650124 Jan 26 17:00:44 crc kubenswrapper[4754]: I0126 17:00:44.839211 4754 generic.go:334] "Generic (PLEG): container finished" podID="4212a345-63c7-4b01-b13d-12f9a3fc297a" containerID="fed4d682a58c3f43dc68d960d1a508ad67578c6214ec387a1c86da1a2cd915a8" exitCode=0 Jan 26 17:00:44 crc kubenswrapper[4754]: I0126 17:00:44.839271 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713v2nvl" event={"ID":"4212a345-63c7-4b01-b13d-12f9a3fc297a","Type":"ContainerDied","Data":"fed4d682a58c3f43dc68d960d1a508ad67578c6214ec387a1c86da1a2cd915a8"} Jan 26 17:00:44 crc kubenswrapper[4754]: I0126 17:00:44.839318 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713v2nvl" event={"ID":"4212a345-63c7-4b01-b13d-12f9a3fc297a","Type":"ContainerStarted","Data":"8d51c58dfeef4bc4e89c3e465625b813c8ae33e34854c11014d8d24c6c650124"} Jan 26 17:00:45 crc kubenswrapper[4754]: I0126 17:00:45.528342 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-f9ghz"] Jan 26 17:00:45 crc kubenswrapper[4754]: I0126 17:00:45.529317 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-f9ghz" Jan 26 17:00:45 crc kubenswrapper[4754]: I0126 17:00:45.543738 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-f9ghz"] Jan 26 17:00:45 crc kubenswrapper[4754]: I0126 17:00:45.609917 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b90dc7aa-8bf8-4770-9302-6e7186bbbd86-catalog-content\") pod \"redhat-operators-f9ghz\" (UID: \"b90dc7aa-8bf8-4770-9302-6e7186bbbd86\") " pod="openshift-marketplace/redhat-operators-f9ghz" Jan 26 17:00:45 crc kubenswrapper[4754]: I0126 17:00:45.610075 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b90dc7aa-8bf8-4770-9302-6e7186bbbd86-utilities\") pod \"redhat-operators-f9ghz\" (UID: \"b90dc7aa-8bf8-4770-9302-6e7186bbbd86\") " pod="openshift-marketplace/redhat-operators-f9ghz" Jan 26 17:00:45 crc kubenswrapper[4754]: I0126 17:00:45.610126 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tkd64\" (UniqueName: \"kubernetes.io/projected/b90dc7aa-8bf8-4770-9302-6e7186bbbd86-kube-api-access-tkd64\") pod \"redhat-operators-f9ghz\" (UID: \"b90dc7aa-8bf8-4770-9302-6e7186bbbd86\") " pod="openshift-marketplace/redhat-operators-f9ghz" Jan 26 17:00:45 crc kubenswrapper[4754]: I0126 17:00:45.711167 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b90dc7aa-8bf8-4770-9302-6e7186bbbd86-utilities\") pod \"redhat-operators-f9ghz\" (UID: \"b90dc7aa-8bf8-4770-9302-6e7186bbbd86\") " pod="openshift-marketplace/redhat-operators-f9ghz" Jan 26 17:00:45 crc kubenswrapper[4754]: I0126 17:00:45.711221 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tkd64\" (UniqueName: \"kubernetes.io/projected/b90dc7aa-8bf8-4770-9302-6e7186bbbd86-kube-api-access-tkd64\") pod \"redhat-operators-f9ghz\" (UID: \"b90dc7aa-8bf8-4770-9302-6e7186bbbd86\") " pod="openshift-marketplace/redhat-operators-f9ghz" Jan 26 17:00:45 crc kubenswrapper[4754]: I0126 17:00:45.711249 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b90dc7aa-8bf8-4770-9302-6e7186bbbd86-catalog-content\") pod \"redhat-operators-f9ghz\" (UID: \"b90dc7aa-8bf8-4770-9302-6e7186bbbd86\") " pod="openshift-marketplace/redhat-operators-f9ghz" Jan 26 17:00:45 crc kubenswrapper[4754]: I0126 17:00:45.711788 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b90dc7aa-8bf8-4770-9302-6e7186bbbd86-catalog-content\") pod \"redhat-operators-f9ghz\" (UID: \"b90dc7aa-8bf8-4770-9302-6e7186bbbd86\") " pod="openshift-marketplace/redhat-operators-f9ghz" Jan 26 17:00:45 crc kubenswrapper[4754]: I0126 17:00:45.711992 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b90dc7aa-8bf8-4770-9302-6e7186bbbd86-utilities\") pod \"redhat-operators-f9ghz\" (UID: \"b90dc7aa-8bf8-4770-9302-6e7186bbbd86\") " pod="openshift-marketplace/redhat-operators-f9ghz" Jan 26 17:00:45 crc kubenswrapper[4754]: I0126 17:00:45.728839 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tkd64\" (UniqueName: \"kubernetes.io/projected/b90dc7aa-8bf8-4770-9302-6e7186bbbd86-kube-api-access-tkd64\") pod \"redhat-operators-f9ghz\" (UID: \"b90dc7aa-8bf8-4770-9302-6e7186bbbd86\") " pod="openshift-marketplace/redhat-operators-f9ghz" Jan 26 17:00:45 crc kubenswrapper[4754]: I0126 17:00:45.842647 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-f9ghz" Jan 26 17:00:46 crc kubenswrapper[4754]: I0126 17:00:46.040845 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-f9ghz"] Jan 26 17:00:46 crc kubenswrapper[4754]: I0126 17:00:46.852192 4754 generic.go:334] "Generic (PLEG): container finished" podID="4212a345-63c7-4b01-b13d-12f9a3fc297a" containerID="06930cf953b61a1b9cf95c4ed66b3ead67261fa2d4a9c59dbf6f17564e759641" exitCode=0 Jan 26 17:00:46 crc kubenswrapper[4754]: I0126 17:00:46.852320 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713v2nvl" event={"ID":"4212a345-63c7-4b01-b13d-12f9a3fc297a","Type":"ContainerDied","Data":"06930cf953b61a1b9cf95c4ed66b3ead67261fa2d4a9c59dbf6f17564e759641"} Jan 26 17:00:46 crc kubenswrapper[4754]: I0126 17:00:46.855280 4754 generic.go:334] "Generic (PLEG): container finished" podID="b90dc7aa-8bf8-4770-9302-6e7186bbbd86" containerID="d38db17138431381a74f7c6df9152e77d0c18b7134aae7bed72eb1449bcbb606" exitCode=0 Jan 26 17:00:46 crc kubenswrapper[4754]: I0126 17:00:46.855337 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f9ghz" event={"ID":"b90dc7aa-8bf8-4770-9302-6e7186bbbd86","Type":"ContainerDied","Data":"d38db17138431381a74f7c6df9152e77d0c18b7134aae7bed72eb1449bcbb606"} Jan 26 17:00:46 crc kubenswrapper[4754]: I0126 17:00:46.855373 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f9ghz" event={"ID":"b90dc7aa-8bf8-4770-9302-6e7186bbbd86","Type":"ContainerStarted","Data":"431d1bee35c6f91690bad45754ee4613c95afacd3ba0533740873891fc1eff9c"} Jan 26 17:00:47 crc kubenswrapper[4754]: I0126 17:00:47.864207 4754 generic.go:334] "Generic (PLEG): container finished" podID="4212a345-63c7-4b01-b13d-12f9a3fc297a" containerID="8251ef8077cbd592b74c8f8818c21915ca71da2b663cef6400609adc781f3d89" exitCode=0 Jan 26 17:00:47 crc kubenswrapper[4754]: I0126 17:00:47.864322 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713v2nvl" event={"ID":"4212a345-63c7-4b01-b13d-12f9a3fc297a","Type":"ContainerDied","Data":"8251ef8077cbd592b74c8f8818c21915ca71da2b663cef6400609adc781f3d89"} Jan 26 17:00:47 crc kubenswrapper[4754]: I0126 17:00:47.866698 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f9ghz" event={"ID":"b90dc7aa-8bf8-4770-9302-6e7186bbbd86","Type":"ContainerStarted","Data":"6a78fceb2d0ae05ec20b59565e6cbd1147cf896d7a3500f7f2b6f02a6b1c38e6"} Jan 26 17:00:48 crc kubenswrapper[4754]: I0126 17:00:48.879477 4754 generic.go:334] "Generic (PLEG): container finished" podID="b90dc7aa-8bf8-4770-9302-6e7186bbbd86" containerID="6a78fceb2d0ae05ec20b59565e6cbd1147cf896d7a3500f7f2b6f02a6b1c38e6" exitCode=0 Jan 26 17:00:48 crc kubenswrapper[4754]: I0126 17:00:48.879579 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f9ghz" event={"ID":"b90dc7aa-8bf8-4770-9302-6e7186bbbd86","Type":"ContainerDied","Data":"6a78fceb2d0ae05ec20b59565e6cbd1147cf896d7a3500f7f2b6f02a6b1c38e6"} Jan 26 17:00:49 crc kubenswrapper[4754]: I0126 17:00:49.161694 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713v2nvl" Jan 26 17:00:49 crc kubenswrapper[4754]: I0126 17:00:49.281761 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4212a345-63c7-4b01-b13d-12f9a3fc297a-bundle\") pod \"4212a345-63c7-4b01-b13d-12f9a3fc297a\" (UID: \"4212a345-63c7-4b01-b13d-12f9a3fc297a\") " Jan 26 17:00:49 crc kubenswrapper[4754]: I0126 17:00:49.281802 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lr42t\" (UniqueName: \"kubernetes.io/projected/4212a345-63c7-4b01-b13d-12f9a3fc297a-kube-api-access-lr42t\") pod \"4212a345-63c7-4b01-b13d-12f9a3fc297a\" (UID: \"4212a345-63c7-4b01-b13d-12f9a3fc297a\") " Jan 26 17:00:49 crc kubenswrapper[4754]: I0126 17:00:49.281884 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4212a345-63c7-4b01-b13d-12f9a3fc297a-util\") pod \"4212a345-63c7-4b01-b13d-12f9a3fc297a\" (UID: \"4212a345-63c7-4b01-b13d-12f9a3fc297a\") " Jan 26 17:00:49 crc kubenswrapper[4754]: I0126 17:00:49.282418 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4212a345-63c7-4b01-b13d-12f9a3fc297a-bundle" (OuterVolumeSpecName: "bundle") pod "4212a345-63c7-4b01-b13d-12f9a3fc297a" (UID: "4212a345-63c7-4b01-b13d-12f9a3fc297a"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:00:49 crc kubenswrapper[4754]: I0126 17:00:49.282954 4754 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4212a345-63c7-4b01-b13d-12f9a3fc297a-bundle\") on node \"crc\" DevicePath \"\"" Jan 26 17:00:49 crc kubenswrapper[4754]: I0126 17:00:49.287401 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4212a345-63c7-4b01-b13d-12f9a3fc297a-kube-api-access-lr42t" (OuterVolumeSpecName: "kube-api-access-lr42t") pod "4212a345-63c7-4b01-b13d-12f9a3fc297a" (UID: "4212a345-63c7-4b01-b13d-12f9a3fc297a"). InnerVolumeSpecName "kube-api-access-lr42t". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:00:49 crc kubenswrapper[4754]: I0126 17:00:49.296386 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4212a345-63c7-4b01-b13d-12f9a3fc297a-util" (OuterVolumeSpecName: "util") pod "4212a345-63c7-4b01-b13d-12f9a3fc297a" (UID: "4212a345-63c7-4b01-b13d-12f9a3fc297a"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:00:49 crc kubenswrapper[4754]: I0126 17:00:49.383472 4754 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4212a345-63c7-4b01-b13d-12f9a3fc297a-util\") on node \"crc\" DevicePath \"\"" Jan 26 17:00:49 crc kubenswrapper[4754]: I0126 17:00:49.383512 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lr42t\" (UniqueName: \"kubernetes.io/projected/4212a345-63c7-4b01-b13d-12f9a3fc297a-kube-api-access-lr42t\") on node \"crc\" DevicePath \"\"" Jan 26 17:00:49 crc kubenswrapper[4754]: I0126 17:00:49.890312 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713v2nvl" event={"ID":"4212a345-63c7-4b01-b13d-12f9a3fc297a","Type":"ContainerDied","Data":"8d51c58dfeef4bc4e89c3e465625b813c8ae33e34854c11014d8d24c6c650124"} Jan 26 17:00:49 crc kubenswrapper[4754]: I0126 17:00:49.890634 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8d51c58dfeef4bc4e89c3e465625b813c8ae33e34854c11014d8d24c6c650124" Jan 26 17:00:49 crc kubenswrapper[4754]: I0126 17:00:49.890342 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713v2nvl" Jan 26 17:00:49 crc kubenswrapper[4754]: I0126 17:00:49.894019 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f9ghz" event={"ID":"b90dc7aa-8bf8-4770-9302-6e7186bbbd86","Type":"ContainerStarted","Data":"a0061f692d9c41eae4e5cbb88ab79b1cc8b77338db27b33a9852ac7353d8fbc9"} Jan 26 17:00:49 crc kubenswrapper[4754]: I0126 17:00:49.925882 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-f9ghz" podStartSLOduration=2.515861019 podStartE2EDuration="4.925854221s" podCreationTimestamp="2026-01-26 17:00:45 +0000 UTC" firstStartedPulling="2026-01-26 17:00:46.856986698 +0000 UTC m=+813.381167152" lastFinishedPulling="2026-01-26 17:00:49.2669799 +0000 UTC m=+815.791160354" observedRunningTime="2026-01-26 17:00:49.921149831 +0000 UTC m=+816.445330265" watchObservedRunningTime="2026-01-26 17:00:49.925854221 +0000 UTC m=+816.450034685" Jan 26 17:00:53 crc kubenswrapper[4754]: I0126 17:00:53.433529 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-646758c888-pwgfk"] Jan 26 17:00:53 crc kubenswrapper[4754]: E0126 17:00:53.434176 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4212a345-63c7-4b01-b13d-12f9a3fc297a" containerName="util" Jan 26 17:00:53 crc kubenswrapper[4754]: I0126 17:00:53.434194 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="4212a345-63c7-4b01-b13d-12f9a3fc297a" containerName="util" Jan 26 17:00:53 crc kubenswrapper[4754]: E0126 17:00:53.434207 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4212a345-63c7-4b01-b13d-12f9a3fc297a" containerName="pull" Jan 26 17:00:53 crc kubenswrapper[4754]: I0126 17:00:53.434215 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="4212a345-63c7-4b01-b13d-12f9a3fc297a" containerName="pull" Jan 26 17:00:53 crc kubenswrapper[4754]: E0126 17:00:53.434239 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4212a345-63c7-4b01-b13d-12f9a3fc297a" containerName="extract" Jan 26 17:00:53 crc kubenswrapper[4754]: I0126 17:00:53.434249 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="4212a345-63c7-4b01-b13d-12f9a3fc297a" containerName="extract" Jan 26 17:00:53 crc kubenswrapper[4754]: I0126 17:00:53.434365 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="4212a345-63c7-4b01-b13d-12f9a3fc297a" containerName="extract" Jan 26 17:00:53 crc kubenswrapper[4754]: I0126 17:00:53.434771 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-646758c888-pwgfk" Jan 26 17:00:53 crc kubenswrapper[4754]: I0126 17:00:53.437202 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-wqrjq" Jan 26 17:00:53 crc kubenswrapper[4754]: I0126 17:00:53.440033 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Jan 26 17:00:53 crc kubenswrapper[4754]: I0126 17:00:53.440037 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Jan 26 17:00:53 crc kubenswrapper[4754]: I0126 17:00:53.448304 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-646758c888-pwgfk"] Jan 26 17:00:53 crc kubenswrapper[4754]: I0126 17:00:53.539982 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6v79p\" (UniqueName: \"kubernetes.io/projected/9bbc5233-67ae-4cca-9a95-71da7e373005-kube-api-access-6v79p\") pod \"nmstate-operator-646758c888-pwgfk\" (UID: \"9bbc5233-67ae-4cca-9a95-71da7e373005\") " pod="openshift-nmstate/nmstate-operator-646758c888-pwgfk" Jan 26 17:00:53 crc kubenswrapper[4754]: I0126 17:00:53.641922 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6v79p\" (UniqueName: \"kubernetes.io/projected/9bbc5233-67ae-4cca-9a95-71da7e373005-kube-api-access-6v79p\") pod \"nmstate-operator-646758c888-pwgfk\" (UID: \"9bbc5233-67ae-4cca-9a95-71da7e373005\") " pod="openshift-nmstate/nmstate-operator-646758c888-pwgfk" Jan 26 17:00:53 crc kubenswrapper[4754]: I0126 17:00:53.666319 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6v79p\" (UniqueName: \"kubernetes.io/projected/9bbc5233-67ae-4cca-9a95-71da7e373005-kube-api-access-6v79p\") pod \"nmstate-operator-646758c888-pwgfk\" (UID: \"9bbc5233-67ae-4cca-9a95-71da7e373005\") " pod="openshift-nmstate/nmstate-operator-646758c888-pwgfk" Jan 26 17:00:53 crc kubenswrapper[4754]: I0126 17:00:53.755138 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-646758c888-pwgfk" Jan 26 17:00:53 crc kubenswrapper[4754]: I0126 17:00:53.957327 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-646758c888-pwgfk"] Jan 26 17:00:53 crc kubenswrapper[4754]: W0126 17:00:53.965958 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9bbc5233_67ae_4cca_9a95_71da7e373005.slice/crio-f526c69e6e894f4bb24f124252a9a5ebf00236ea67fb91c9c51cac8fed1e6cd7 WatchSource:0}: Error finding container f526c69e6e894f4bb24f124252a9a5ebf00236ea67fb91c9c51cac8fed1e6cd7: Status 404 returned error can't find the container with id f526c69e6e894f4bb24f124252a9a5ebf00236ea67fb91c9c51cac8fed1e6cd7 Jan 26 17:00:54 crc kubenswrapper[4754]: I0126 17:00:54.928893 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-646758c888-pwgfk" event={"ID":"9bbc5233-67ae-4cca-9a95-71da7e373005","Type":"ContainerStarted","Data":"f526c69e6e894f4bb24f124252a9a5ebf00236ea67fb91c9c51cac8fed1e6cd7"} Jan 26 17:00:55 crc kubenswrapper[4754]: I0126 17:00:55.843389 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-f9ghz" Jan 26 17:00:55 crc kubenswrapper[4754]: I0126 17:00:55.843715 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-f9ghz" Jan 26 17:00:56 crc kubenswrapper[4754]: I0126 17:00:56.891388 4754 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-f9ghz" podUID="b90dc7aa-8bf8-4770-9302-6e7186bbbd86" containerName="registry-server" probeResult="failure" output=< Jan 26 17:00:56 crc kubenswrapper[4754]: timeout: failed to connect service ":50051" within 1s Jan 26 17:00:56 crc kubenswrapper[4754]: > Jan 26 17:00:56 crc kubenswrapper[4754]: I0126 17:00:56.940557 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-646758c888-pwgfk" event={"ID":"9bbc5233-67ae-4cca-9a95-71da7e373005","Type":"ContainerStarted","Data":"11fc86ea378bd7893ab532bb0223624f3e781e9b430cb2e1cdf8650938f8c605"} Jan 26 17:00:56 crc kubenswrapper[4754]: I0126 17:00:56.956836 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-646758c888-pwgfk" podStartSLOduration=1.196465677 podStartE2EDuration="3.956818664s" podCreationTimestamp="2026-01-26 17:00:53 +0000 UTC" firstStartedPulling="2026-01-26 17:00:53.968661956 +0000 UTC m=+820.492842390" lastFinishedPulling="2026-01-26 17:00:56.729014933 +0000 UTC m=+823.253195377" observedRunningTime="2026-01-26 17:00:56.95626942 +0000 UTC m=+823.480449904" watchObservedRunningTime="2026-01-26 17:00:56.956818664 +0000 UTC m=+823.480999098" Jan 26 17:01:03 crc kubenswrapper[4754]: I0126 17:01:03.536380 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-8474b5b9d8-zgmpc"] Jan 26 17:01:03 crc kubenswrapper[4754]: I0126 17:01:03.537733 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-zgmpc" Jan 26 17:01:03 crc kubenswrapper[4754]: I0126 17:01:03.541454 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-8bp85" Jan 26 17:01:03 crc kubenswrapper[4754]: I0126 17:01:03.542063 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-54757c584b-vvxrw"] Jan 26 17:01:03 crc kubenswrapper[4754]: I0126 17:01:03.543454 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-54757c584b-vvxrw" Jan 26 17:01:03 crc kubenswrapper[4754]: I0126 17:01:03.546697 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Jan 26 17:01:03 crc kubenswrapper[4754]: I0126 17:01:03.553060 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-8474b5b9d8-zgmpc"] Jan 26 17:01:03 crc kubenswrapper[4754]: I0126 17:01:03.582549 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-54757c584b-vvxrw"] Jan 26 17:01:03 crc kubenswrapper[4754]: I0126 17:01:03.594381 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-qfmmc"] Jan 26 17:01:03 crc kubenswrapper[4754]: I0126 17:01:03.595461 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-qfmmc" Jan 26 17:01:03 crc kubenswrapper[4754]: I0126 17:01:03.665410 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7754f76f8b-dxcqw"] Jan 26 17:01:03 crc kubenswrapper[4754]: I0126 17:01:03.666496 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-dxcqw" Jan 26 17:01:03 crc kubenswrapper[4754]: I0126 17:01:03.668824 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Jan 26 17:01:03 crc kubenswrapper[4754]: I0126 17:01:03.669116 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-7x7jh" Jan 26 17:01:03 crc kubenswrapper[4754]: I0126 17:01:03.669968 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Jan 26 17:01:03 crc kubenswrapper[4754]: I0126 17:01:03.677374 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7754f76f8b-dxcqw"] Jan 26 17:01:03 crc kubenswrapper[4754]: I0126 17:01:03.686901 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/fa03d76c-cc42-4ba9-ad6f-671b4b63dbab-tls-key-pair\") pod \"nmstate-webhook-8474b5b9d8-zgmpc\" (UID: \"fa03d76c-cc42-4ba9-ad6f-671b4b63dbab\") " pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-zgmpc" Jan 26 17:01:03 crc kubenswrapper[4754]: I0126 17:01:03.686976 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n2tpx\" (UniqueName: \"kubernetes.io/projected/fa03d76c-cc42-4ba9-ad6f-671b4b63dbab-kube-api-access-n2tpx\") pod \"nmstate-webhook-8474b5b9d8-zgmpc\" (UID: \"fa03d76c-cc42-4ba9-ad6f-671b4b63dbab\") " pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-zgmpc" Jan 26 17:01:03 crc kubenswrapper[4754]: I0126 17:01:03.687010 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rqd46\" (UniqueName: \"kubernetes.io/projected/6506a03e-2445-4e3c-9814-f5bb1cfbaa06-kube-api-access-rqd46\") pod \"nmstate-metrics-54757c584b-vvxrw\" (UID: \"6506a03e-2445-4e3c-9814-f5bb1cfbaa06\") " pod="openshift-nmstate/nmstate-metrics-54757c584b-vvxrw" Jan 26 17:01:03 crc kubenswrapper[4754]: I0126 17:01:03.788430 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/73c39dd3-4d2e-4ee8-a218-f88eb59f36b7-ovs-socket\") pod \"nmstate-handler-qfmmc\" (UID: \"73c39dd3-4d2e-4ee8-a218-f88eb59f36b7\") " pod="openshift-nmstate/nmstate-handler-qfmmc" Jan 26 17:01:03 crc kubenswrapper[4754]: I0126 17:01:03.788480 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n2tpx\" (UniqueName: \"kubernetes.io/projected/fa03d76c-cc42-4ba9-ad6f-671b4b63dbab-kube-api-access-n2tpx\") pod \"nmstate-webhook-8474b5b9d8-zgmpc\" (UID: \"fa03d76c-cc42-4ba9-ad6f-671b4b63dbab\") " pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-zgmpc" Jan 26 17:01:03 crc kubenswrapper[4754]: I0126 17:01:03.788503 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rqd46\" (UniqueName: \"kubernetes.io/projected/6506a03e-2445-4e3c-9814-f5bb1cfbaa06-kube-api-access-rqd46\") pod \"nmstate-metrics-54757c584b-vvxrw\" (UID: \"6506a03e-2445-4e3c-9814-f5bb1cfbaa06\") " pod="openshift-nmstate/nmstate-metrics-54757c584b-vvxrw" Jan 26 17:01:03 crc kubenswrapper[4754]: I0126 17:01:03.788526 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/73c39dd3-4d2e-4ee8-a218-f88eb59f36b7-dbus-socket\") pod \"nmstate-handler-qfmmc\" (UID: \"73c39dd3-4d2e-4ee8-a218-f88eb59f36b7\") " pod="openshift-nmstate/nmstate-handler-qfmmc" Jan 26 17:01:03 crc kubenswrapper[4754]: I0126 17:01:03.788629 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xp85k\" (UniqueName: \"kubernetes.io/projected/97017e08-dd1d-494a-b4bf-0800824588a2-kube-api-access-xp85k\") pod \"nmstate-console-plugin-7754f76f8b-dxcqw\" (UID: \"97017e08-dd1d-494a-b4bf-0800824588a2\") " pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-dxcqw" Jan 26 17:01:03 crc kubenswrapper[4754]: I0126 17:01:03.788731 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/97017e08-dd1d-494a-b4bf-0800824588a2-plugin-serving-cert\") pod \"nmstate-console-plugin-7754f76f8b-dxcqw\" (UID: \"97017e08-dd1d-494a-b4bf-0800824588a2\") " pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-dxcqw" Jan 26 17:01:03 crc kubenswrapper[4754]: I0126 17:01:03.788834 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/97017e08-dd1d-494a-b4bf-0800824588a2-nginx-conf\") pod \"nmstate-console-plugin-7754f76f8b-dxcqw\" (UID: \"97017e08-dd1d-494a-b4bf-0800824588a2\") " pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-dxcqw" Jan 26 17:01:03 crc kubenswrapper[4754]: I0126 17:01:03.788881 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/73c39dd3-4d2e-4ee8-a218-f88eb59f36b7-nmstate-lock\") pod \"nmstate-handler-qfmmc\" (UID: \"73c39dd3-4d2e-4ee8-a218-f88eb59f36b7\") " pod="openshift-nmstate/nmstate-handler-qfmmc" Jan 26 17:01:03 crc kubenswrapper[4754]: I0126 17:01:03.789295 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/fa03d76c-cc42-4ba9-ad6f-671b4b63dbab-tls-key-pair\") pod \"nmstate-webhook-8474b5b9d8-zgmpc\" (UID: \"fa03d76c-cc42-4ba9-ad6f-671b4b63dbab\") " pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-zgmpc" Jan 26 17:01:03 crc kubenswrapper[4754]: I0126 17:01:03.789719 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6fwss\" (UniqueName: \"kubernetes.io/projected/73c39dd3-4d2e-4ee8-a218-f88eb59f36b7-kube-api-access-6fwss\") pod \"nmstate-handler-qfmmc\" (UID: \"73c39dd3-4d2e-4ee8-a218-f88eb59f36b7\") " pod="openshift-nmstate/nmstate-handler-qfmmc" Jan 26 17:01:03 crc kubenswrapper[4754]: I0126 17:01:03.797312 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/fa03d76c-cc42-4ba9-ad6f-671b4b63dbab-tls-key-pair\") pod \"nmstate-webhook-8474b5b9d8-zgmpc\" (UID: \"fa03d76c-cc42-4ba9-ad6f-671b4b63dbab\") " pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-zgmpc" Jan 26 17:01:03 crc kubenswrapper[4754]: I0126 17:01:03.816293 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n2tpx\" (UniqueName: \"kubernetes.io/projected/fa03d76c-cc42-4ba9-ad6f-671b4b63dbab-kube-api-access-n2tpx\") pod \"nmstate-webhook-8474b5b9d8-zgmpc\" (UID: \"fa03d76c-cc42-4ba9-ad6f-671b4b63dbab\") " pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-zgmpc" Jan 26 17:01:03 crc kubenswrapper[4754]: I0126 17:01:03.817927 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rqd46\" (UniqueName: \"kubernetes.io/projected/6506a03e-2445-4e3c-9814-f5bb1cfbaa06-kube-api-access-rqd46\") pod \"nmstate-metrics-54757c584b-vvxrw\" (UID: \"6506a03e-2445-4e3c-9814-f5bb1cfbaa06\") " pod="openshift-nmstate/nmstate-metrics-54757c584b-vvxrw" Jan 26 17:01:03 crc kubenswrapper[4754]: I0126 17:01:03.841269 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-75765cbdc-w5c9p"] Jan 26 17:01:03 crc kubenswrapper[4754]: I0126 17:01:03.842579 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-75765cbdc-w5c9p" Jan 26 17:01:03 crc kubenswrapper[4754]: I0126 17:01:03.852093 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-75765cbdc-w5c9p"] Jan 26 17:01:03 crc kubenswrapper[4754]: I0126 17:01:03.872742 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-zgmpc" Jan 26 17:01:03 crc kubenswrapper[4754]: I0126 17:01:03.886810 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-54757c584b-vvxrw" Jan 26 17:01:03 crc kubenswrapper[4754]: I0126 17:01:03.891556 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/73c39dd3-4d2e-4ee8-a218-f88eb59f36b7-ovs-socket\") pod \"nmstate-handler-qfmmc\" (UID: \"73c39dd3-4d2e-4ee8-a218-f88eb59f36b7\") " pod="openshift-nmstate/nmstate-handler-qfmmc" Jan 26 17:01:03 crc kubenswrapper[4754]: I0126 17:01:03.891767 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/73c39dd3-4d2e-4ee8-a218-f88eb59f36b7-dbus-socket\") pod \"nmstate-handler-qfmmc\" (UID: \"73c39dd3-4d2e-4ee8-a218-f88eb59f36b7\") " pod="openshift-nmstate/nmstate-handler-qfmmc" Jan 26 17:01:03 crc kubenswrapper[4754]: I0126 17:01:03.891866 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xp85k\" (UniqueName: \"kubernetes.io/projected/97017e08-dd1d-494a-b4bf-0800824588a2-kube-api-access-xp85k\") pod \"nmstate-console-plugin-7754f76f8b-dxcqw\" (UID: \"97017e08-dd1d-494a-b4bf-0800824588a2\") " pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-dxcqw" Jan 26 17:01:03 crc kubenswrapper[4754]: I0126 17:01:03.891942 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/97017e08-dd1d-494a-b4bf-0800824588a2-plugin-serving-cert\") pod \"nmstate-console-plugin-7754f76f8b-dxcqw\" (UID: \"97017e08-dd1d-494a-b4bf-0800824588a2\") " pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-dxcqw" Jan 26 17:01:03 crc kubenswrapper[4754]: I0126 17:01:03.892018 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/97017e08-dd1d-494a-b4bf-0800824588a2-nginx-conf\") pod \"nmstate-console-plugin-7754f76f8b-dxcqw\" (UID: \"97017e08-dd1d-494a-b4bf-0800824588a2\") " pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-dxcqw" Jan 26 17:01:03 crc kubenswrapper[4754]: I0126 17:01:03.892084 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/73c39dd3-4d2e-4ee8-a218-f88eb59f36b7-nmstate-lock\") pod \"nmstate-handler-qfmmc\" (UID: \"73c39dd3-4d2e-4ee8-a218-f88eb59f36b7\") " pod="openshift-nmstate/nmstate-handler-qfmmc" Jan 26 17:01:03 crc kubenswrapper[4754]: I0126 17:01:03.892162 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6fwss\" (UniqueName: \"kubernetes.io/projected/73c39dd3-4d2e-4ee8-a218-f88eb59f36b7-kube-api-access-6fwss\") pod \"nmstate-handler-qfmmc\" (UID: \"73c39dd3-4d2e-4ee8-a218-f88eb59f36b7\") " pod="openshift-nmstate/nmstate-handler-qfmmc" Jan 26 17:01:03 crc kubenswrapper[4754]: I0126 17:01:03.891727 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/73c39dd3-4d2e-4ee8-a218-f88eb59f36b7-ovs-socket\") pod \"nmstate-handler-qfmmc\" (UID: \"73c39dd3-4d2e-4ee8-a218-f88eb59f36b7\") " pod="openshift-nmstate/nmstate-handler-qfmmc" Jan 26 17:01:03 crc kubenswrapper[4754]: I0126 17:01:03.892719 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/73c39dd3-4d2e-4ee8-a218-f88eb59f36b7-dbus-socket\") pod \"nmstate-handler-qfmmc\" (UID: \"73c39dd3-4d2e-4ee8-a218-f88eb59f36b7\") " pod="openshift-nmstate/nmstate-handler-qfmmc" Jan 26 17:01:03 crc kubenswrapper[4754]: I0126 17:01:03.892759 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/73c39dd3-4d2e-4ee8-a218-f88eb59f36b7-nmstate-lock\") pod \"nmstate-handler-qfmmc\" (UID: \"73c39dd3-4d2e-4ee8-a218-f88eb59f36b7\") " pod="openshift-nmstate/nmstate-handler-qfmmc" Jan 26 17:01:03 crc kubenswrapper[4754]: I0126 17:01:03.893640 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/97017e08-dd1d-494a-b4bf-0800824588a2-nginx-conf\") pod \"nmstate-console-plugin-7754f76f8b-dxcqw\" (UID: \"97017e08-dd1d-494a-b4bf-0800824588a2\") " pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-dxcqw" Jan 26 17:01:03 crc kubenswrapper[4754]: I0126 17:01:03.895606 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/97017e08-dd1d-494a-b4bf-0800824588a2-plugin-serving-cert\") pod \"nmstate-console-plugin-7754f76f8b-dxcqw\" (UID: \"97017e08-dd1d-494a-b4bf-0800824588a2\") " pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-dxcqw" Jan 26 17:01:03 crc kubenswrapper[4754]: I0126 17:01:03.907561 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xp85k\" (UniqueName: \"kubernetes.io/projected/97017e08-dd1d-494a-b4bf-0800824588a2-kube-api-access-xp85k\") pod \"nmstate-console-plugin-7754f76f8b-dxcqw\" (UID: \"97017e08-dd1d-494a-b4bf-0800824588a2\") " pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-dxcqw" Jan 26 17:01:03 crc kubenswrapper[4754]: I0126 17:01:03.908572 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6fwss\" (UniqueName: \"kubernetes.io/projected/73c39dd3-4d2e-4ee8-a218-f88eb59f36b7-kube-api-access-6fwss\") pod \"nmstate-handler-qfmmc\" (UID: \"73c39dd3-4d2e-4ee8-a218-f88eb59f36b7\") " pod="openshift-nmstate/nmstate-handler-qfmmc" Jan 26 17:01:03 crc kubenswrapper[4754]: I0126 17:01:03.910547 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-qfmmc" Jan 26 17:01:03 crc kubenswrapper[4754]: I0126 17:01:03.989500 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-qfmmc" event={"ID":"73c39dd3-4d2e-4ee8-a218-f88eb59f36b7","Type":"ContainerStarted","Data":"520a7de62b216c5032a2d51478969ae58003ef4c46981e1cf64670188dc70019"} Jan 26 17:01:03 crc kubenswrapper[4754]: I0126 17:01:03.992129 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-dxcqw" Jan 26 17:01:03 crc kubenswrapper[4754]: I0126 17:01:03.993206 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/99e872d5-3e8c-41c5-9c12-18f70ab31205-service-ca\") pod \"console-75765cbdc-w5c9p\" (UID: \"99e872d5-3e8c-41c5-9c12-18f70ab31205\") " pod="openshift-console/console-75765cbdc-w5c9p" Jan 26 17:01:03 crc kubenswrapper[4754]: I0126 17:01:03.993239 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/99e872d5-3e8c-41c5-9c12-18f70ab31205-console-oauth-config\") pod \"console-75765cbdc-w5c9p\" (UID: \"99e872d5-3e8c-41c5-9c12-18f70ab31205\") " pod="openshift-console/console-75765cbdc-w5c9p" Jan 26 17:01:03 crc kubenswrapper[4754]: I0126 17:01:03.993284 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/99e872d5-3e8c-41c5-9c12-18f70ab31205-console-serving-cert\") pod \"console-75765cbdc-w5c9p\" (UID: \"99e872d5-3e8c-41c5-9c12-18f70ab31205\") " pod="openshift-console/console-75765cbdc-w5c9p" Jan 26 17:01:03 crc kubenswrapper[4754]: I0126 17:01:03.993304 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/99e872d5-3e8c-41c5-9c12-18f70ab31205-console-config\") pod \"console-75765cbdc-w5c9p\" (UID: \"99e872d5-3e8c-41c5-9c12-18f70ab31205\") " pod="openshift-console/console-75765cbdc-w5c9p" Jan 26 17:01:03 crc kubenswrapper[4754]: I0126 17:01:03.993332 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t8cf7\" (UniqueName: \"kubernetes.io/projected/99e872d5-3e8c-41c5-9c12-18f70ab31205-kube-api-access-t8cf7\") pod \"console-75765cbdc-w5c9p\" (UID: \"99e872d5-3e8c-41c5-9c12-18f70ab31205\") " pod="openshift-console/console-75765cbdc-w5c9p" Jan 26 17:01:03 crc kubenswrapper[4754]: I0126 17:01:03.993352 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/99e872d5-3e8c-41c5-9c12-18f70ab31205-trusted-ca-bundle\") pod \"console-75765cbdc-w5c9p\" (UID: \"99e872d5-3e8c-41c5-9c12-18f70ab31205\") " pod="openshift-console/console-75765cbdc-w5c9p" Jan 26 17:01:03 crc kubenswrapper[4754]: I0126 17:01:03.993420 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/99e872d5-3e8c-41c5-9c12-18f70ab31205-oauth-serving-cert\") pod \"console-75765cbdc-w5c9p\" (UID: \"99e872d5-3e8c-41c5-9c12-18f70ab31205\") " pod="openshift-console/console-75765cbdc-w5c9p" Jan 26 17:01:04 crc kubenswrapper[4754]: I0126 17:01:04.042830 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-8474b5b9d8-zgmpc"] Jan 26 17:01:04 crc kubenswrapper[4754]: W0126 17:01:04.052503 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfa03d76c_cc42_4ba9_ad6f_671b4b63dbab.slice/crio-66071f605e00479142880113f06e9730a2dce81e0968ba9680c02fd73900de67 WatchSource:0}: Error finding container 66071f605e00479142880113f06e9730a2dce81e0968ba9680c02fd73900de67: Status 404 returned error can't find the container with id 66071f605e00479142880113f06e9730a2dce81e0968ba9680c02fd73900de67 Jan 26 17:01:04 crc kubenswrapper[4754]: I0126 17:01:04.099112 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/99e872d5-3e8c-41c5-9c12-18f70ab31205-service-ca\") pod \"console-75765cbdc-w5c9p\" (UID: \"99e872d5-3e8c-41c5-9c12-18f70ab31205\") " pod="openshift-console/console-75765cbdc-w5c9p" Jan 26 17:01:04 crc kubenswrapper[4754]: I0126 17:01:04.099162 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/99e872d5-3e8c-41c5-9c12-18f70ab31205-console-oauth-config\") pod \"console-75765cbdc-w5c9p\" (UID: \"99e872d5-3e8c-41c5-9c12-18f70ab31205\") " pod="openshift-console/console-75765cbdc-w5c9p" Jan 26 17:01:04 crc kubenswrapper[4754]: I0126 17:01:04.099223 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/99e872d5-3e8c-41c5-9c12-18f70ab31205-console-serving-cert\") pod \"console-75765cbdc-w5c9p\" (UID: \"99e872d5-3e8c-41c5-9c12-18f70ab31205\") " pod="openshift-console/console-75765cbdc-w5c9p" Jan 26 17:01:04 crc kubenswrapper[4754]: I0126 17:01:04.099245 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/99e872d5-3e8c-41c5-9c12-18f70ab31205-console-config\") pod \"console-75765cbdc-w5c9p\" (UID: \"99e872d5-3e8c-41c5-9c12-18f70ab31205\") " pod="openshift-console/console-75765cbdc-w5c9p" Jan 26 17:01:04 crc kubenswrapper[4754]: I0126 17:01:04.099282 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t8cf7\" (UniqueName: \"kubernetes.io/projected/99e872d5-3e8c-41c5-9c12-18f70ab31205-kube-api-access-t8cf7\") pod \"console-75765cbdc-w5c9p\" (UID: \"99e872d5-3e8c-41c5-9c12-18f70ab31205\") " pod="openshift-console/console-75765cbdc-w5c9p" Jan 26 17:01:04 crc kubenswrapper[4754]: I0126 17:01:04.099311 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/99e872d5-3e8c-41c5-9c12-18f70ab31205-trusted-ca-bundle\") pod \"console-75765cbdc-w5c9p\" (UID: \"99e872d5-3e8c-41c5-9c12-18f70ab31205\") " pod="openshift-console/console-75765cbdc-w5c9p" Jan 26 17:01:04 crc kubenswrapper[4754]: I0126 17:01:04.099353 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/99e872d5-3e8c-41c5-9c12-18f70ab31205-oauth-serving-cert\") pod \"console-75765cbdc-w5c9p\" (UID: \"99e872d5-3e8c-41c5-9c12-18f70ab31205\") " pod="openshift-console/console-75765cbdc-w5c9p" Jan 26 17:01:04 crc kubenswrapper[4754]: I0126 17:01:04.100116 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-54757c584b-vvxrw"] Jan 26 17:01:04 crc kubenswrapper[4754]: I0126 17:01:04.100980 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/99e872d5-3e8c-41c5-9c12-18f70ab31205-oauth-serving-cert\") pod \"console-75765cbdc-w5c9p\" (UID: \"99e872d5-3e8c-41c5-9c12-18f70ab31205\") " pod="openshift-console/console-75765cbdc-w5c9p" Jan 26 17:01:04 crc kubenswrapper[4754]: I0126 17:01:04.101252 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/99e872d5-3e8c-41c5-9c12-18f70ab31205-service-ca\") pod \"console-75765cbdc-w5c9p\" (UID: \"99e872d5-3e8c-41c5-9c12-18f70ab31205\") " pod="openshift-console/console-75765cbdc-w5c9p" Jan 26 17:01:04 crc kubenswrapper[4754]: I0126 17:01:04.102271 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/99e872d5-3e8c-41c5-9c12-18f70ab31205-trusted-ca-bundle\") pod \"console-75765cbdc-w5c9p\" (UID: \"99e872d5-3e8c-41c5-9c12-18f70ab31205\") " pod="openshift-console/console-75765cbdc-w5c9p" Jan 26 17:01:04 crc kubenswrapper[4754]: I0126 17:01:04.101320 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/99e872d5-3e8c-41c5-9c12-18f70ab31205-console-config\") pod \"console-75765cbdc-w5c9p\" (UID: \"99e872d5-3e8c-41c5-9c12-18f70ab31205\") " pod="openshift-console/console-75765cbdc-w5c9p" Jan 26 17:01:04 crc kubenswrapper[4754]: I0126 17:01:04.105340 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/99e872d5-3e8c-41c5-9c12-18f70ab31205-console-oauth-config\") pod \"console-75765cbdc-w5c9p\" (UID: \"99e872d5-3e8c-41c5-9c12-18f70ab31205\") " pod="openshift-console/console-75765cbdc-w5c9p" Jan 26 17:01:04 crc kubenswrapper[4754]: W0126 17:01:04.106843 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6506a03e_2445_4e3c_9814_f5bb1cfbaa06.slice/crio-82ce2b1c3c0cf15a5bcc0dfe37d25b7b523c63d446d1e8b076cda2495b346ecc WatchSource:0}: Error finding container 82ce2b1c3c0cf15a5bcc0dfe37d25b7b523c63d446d1e8b076cda2495b346ecc: Status 404 returned error can't find the container with id 82ce2b1c3c0cf15a5bcc0dfe37d25b7b523c63d446d1e8b076cda2495b346ecc Jan 26 17:01:04 crc kubenswrapper[4754]: I0126 17:01:04.108008 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/99e872d5-3e8c-41c5-9c12-18f70ab31205-console-serving-cert\") pod \"console-75765cbdc-w5c9p\" (UID: \"99e872d5-3e8c-41c5-9c12-18f70ab31205\") " pod="openshift-console/console-75765cbdc-w5c9p" Jan 26 17:01:04 crc kubenswrapper[4754]: I0126 17:01:04.116648 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t8cf7\" (UniqueName: \"kubernetes.io/projected/99e872d5-3e8c-41c5-9c12-18f70ab31205-kube-api-access-t8cf7\") pod \"console-75765cbdc-w5c9p\" (UID: \"99e872d5-3e8c-41c5-9c12-18f70ab31205\") " pod="openshift-console/console-75765cbdc-w5c9p" Jan 26 17:01:04 crc kubenswrapper[4754]: I0126 17:01:04.164713 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-75765cbdc-w5c9p" Jan 26 17:01:04 crc kubenswrapper[4754]: I0126 17:01:04.340947 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-75765cbdc-w5c9p"] Jan 26 17:01:04 crc kubenswrapper[4754]: W0126 17:01:04.344424 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod99e872d5_3e8c_41c5_9c12_18f70ab31205.slice/crio-b493643b0808e252866be84fca091448be3186ee9ad2882daeed5ef5c0641f4f WatchSource:0}: Error finding container b493643b0808e252866be84fca091448be3186ee9ad2882daeed5ef5c0641f4f: Status 404 returned error can't find the container with id b493643b0808e252866be84fca091448be3186ee9ad2882daeed5ef5c0641f4f Jan 26 17:01:04 crc kubenswrapper[4754]: I0126 17:01:04.367653 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7754f76f8b-dxcqw"] Jan 26 17:01:04 crc kubenswrapper[4754]: W0126 17:01:04.371077 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod97017e08_dd1d_494a_b4bf_0800824588a2.slice/crio-467bd1d7e01a2423fce642f1700f460e93a37f482ac8d47553989a0dd86fc58a WatchSource:0}: Error finding container 467bd1d7e01a2423fce642f1700f460e93a37f482ac8d47553989a0dd86fc58a: Status 404 returned error can't find the container with id 467bd1d7e01a2423fce642f1700f460e93a37f482ac8d47553989a0dd86fc58a Jan 26 17:01:04 crc kubenswrapper[4754]: I0126 17:01:04.997928 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-75765cbdc-w5c9p" event={"ID":"99e872d5-3e8c-41c5-9c12-18f70ab31205","Type":"ContainerStarted","Data":"b493643b0808e252866be84fca091448be3186ee9ad2882daeed5ef5c0641f4f"} Jan 26 17:01:05 crc kubenswrapper[4754]: I0126 17:01:05.000356 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-zgmpc" event={"ID":"fa03d76c-cc42-4ba9-ad6f-671b4b63dbab","Type":"ContainerStarted","Data":"66071f605e00479142880113f06e9730a2dce81e0968ba9680c02fd73900de67"} Jan 26 17:01:05 crc kubenswrapper[4754]: I0126 17:01:05.002113 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-54757c584b-vvxrw" event={"ID":"6506a03e-2445-4e3c-9814-f5bb1cfbaa06","Type":"ContainerStarted","Data":"82ce2b1c3c0cf15a5bcc0dfe37d25b7b523c63d446d1e8b076cda2495b346ecc"} Jan 26 17:01:05 crc kubenswrapper[4754]: I0126 17:01:05.004245 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-dxcqw" event={"ID":"97017e08-dd1d-494a-b4bf-0800824588a2","Type":"ContainerStarted","Data":"467bd1d7e01a2423fce642f1700f460e93a37f482ac8d47553989a0dd86fc58a"} Jan 26 17:01:05 crc kubenswrapper[4754]: I0126 17:01:05.889843 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-f9ghz" Jan 26 17:01:05 crc kubenswrapper[4754]: I0126 17:01:05.930775 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-f9ghz" Jan 26 17:01:06 crc kubenswrapper[4754]: I0126 17:01:06.015641 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-75765cbdc-w5c9p" event={"ID":"99e872d5-3e8c-41c5-9c12-18f70ab31205","Type":"ContainerStarted","Data":"2616ef24b102d717755c2573d4001afcba41eb22638860640a6b1f009059ac63"} Jan 26 17:01:06 crc kubenswrapper[4754]: I0126 17:01:06.033127 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-75765cbdc-w5c9p" podStartSLOduration=3.033113691 podStartE2EDuration="3.033113691s" podCreationTimestamp="2026-01-26 17:01:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 17:01:06.032172866 +0000 UTC m=+832.556353310" watchObservedRunningTime="2026-01-26 17:01:06.033113691 +0000 UTC m=+832.557294115" Jan 26 17:01:06 crc kubenswrapper[4754]: I0126 17:01:06.120469 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-f9ghz"] Jan 26 17:01:07 crc kubenswrapper[4754]: I0126 17:01:07.021885 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-f9ghz" podUID="b90dc7aa-8bf8-4770-9302-6e7186bbbd86" containerName="registry-server" containerID="cri-o://a0061f692d9c41eae4e5cbb88ab79b1cc8b77338db27b33a9852ac7353d8fbc9" gracePeriod=2 Jan 26 17:01:07 crc kubenswrapper[4754]: I0126 17:01:07.553336 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-f9ghz" Jan 26 17:01:07 crc kubenswrapper[4754]: I0126 17:01:07.750348 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b90dc7aa-8bf8-4770-9302-6e7186bbbd86-utilities\") pod \"b90dc7aa-8bf8-4770-9302-6e7186bbbd86\" (UID: \"b90dc7aa-8bf8-4770-9302-6e7186bbbd86\") " Jan 26 17:01:07 crc kubenswrapper[4754]: I0126 17:01:07.750495 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b90dc7aa-8bf8-4770-9302-6e7186bbbd86-catalog-content\") pod \"b90dc7aa-8bf8-4770-9302-6e7186bbbd86\" (UID: \"b90dc7aa-8bf8-4770-9302-6e7186bbbd86\") " Jan 26 17:01:07 crc kubenswrapper[4754]: I0126 17:01:07.750536 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tkd64\" (UniqueName: \"kubernetes.io/projected/b90dc7aa-8bf8-4770-9302-6e7186bbbd86-kube-api-access-tkd64\") pod \"b90dc7aa-8bf8-4770-9302-6e7186bbbd86\" (UID: \"b90dc7aa-8bf8-4770-9302-6e7186bbbd86\") " Jan 26 17:01:07 crc kubenswrapper[4754]: I0126 17:01:07.751211 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b90dc7aa-8bf8-4770-9302-6e7186bbbd86-utilities" (OuterVolumeSpecName: "utilities") pod "b90dc7aa-8bf8-4770-9302-6e7186bbbd86" (UID: "b90dc7aa-8bf8-4770-9302-6e7186bbbd86"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:01:07 crc kubenswrapper[4754]: I0126 17:01:07.755107 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b90dc7aa-8bf8-4770-9302-6e7186bbbd86-kube-api-access-tkd64" (OuterVolumeSpecName: "kube-api-access-tkd64") pod "b90dc7aa-8bf8-4770-9302-6e7186bbbd86" (UID: "b90dc7aa-8bf8-4770-9302-6e7186bbbd86"). InnerVolumeSpecName "kube-api-access-tkd64". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:01:07 crc kubenswrapper[4754]: I0126 17:01:07.852352 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tkd64\" (UniqueName: \"kubernetes.io/projected/b90dc7aa-8bf8-4770-9302-6e7186bbbd86-kube-api-access-tkd64\") on node \"crc\" DevicePath \"\"" Jan 26 17:01:07 crc kubenswrapper[4754]: I0126 17:01:07.852413 4754 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b90dc7aa-8bf8-4770-9302-6e7186bbbd86-utilities\") on node \"crc\" DevicePath \"\"" Jan 26 17:01:07 crc kubenswrapper[4754]: I0126 17:01:07.910943 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b90dc7aa-8bf8-4770-9302-6e7186bbbd86-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b90dc7aa-8bf8-4770-9302-6e7186bbbd86" (UID: "b90dc7aa-8bf8-4770-9302-6e7186bbbd86"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:01:07 crc kubenswrapper[4754]: I0126 17:01:07.953653 4754 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b90dc7aa-8bf8-4770-9302-6e7186bbbd86-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 26 17:01:08 crc kubenswrapper[4754]: I0126 17:01:08.030296 4754 generic.go:334] "Generic (PLEG): container finished" podID="b90dc7aa-8bf8-4770-9302-6e7186bbbd86" containerID="a0061f692d9c41eae4e5cbb88ab79b1cc8b77338db27b33a9852ac7353d8fbc9" exitCode=0 Jan 26 17:01:08 crc kubenswrapper[4754]: I0126 17:01:08.030381 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f9ghz" event={"ID":"b90dc7aa-8bf8-4770-9302-6e7186bbbd86","Type":"ContainerDied","Data":"a0061f692d9c41eae4e5cbb88ab79b1cc8b77338db27b33a9852ac7353d8fbc9"} Jan 26 17:01:08 crc kubenswrapper[4754]: I0126 17:01:08.030418 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-f9ghz" Jan 26 17:01:08 crc kubenswrapper[4754]: I0126 17:01:08.030504 4754 scope.go:117] "RemoveContainer" containerID="a0061f692d9c41eae4e5cbb88ab79b1cc8b77338db27b33a9852ac7353d8fbc9" Jan 26 17:01:08 crc kubenswrapper[4754]: I0126 17:01:08.030483 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f9ghz" event={"ID":"b90dc7aa-8bf8-4770-9302-6e7186bbbd86","Type":"ContainerDied","Data":"431d1bee35c6f91690bad45754ee4613c95afacd3ba0533740873891fc1eff9c"} Jan 26 17:01:08 crc kubenswrapper[4754]: I0126 17:01:08.033186 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-dxcqw" event={"ID":"97017e08-dd1d-494a-b4bf-0800824588a2","Type":"ContainerStarted","Data":"2396ea0a306f73f79e3e86060e4b64243a23d1ad54d2b71cdc8e0e634a0b9967"} Jan 26 17:01:08 crc kubenswrapper[4754]: I0126 17:01:08.036811 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-zgmpc" event={"ID":"fa03d76c-cc42-4ba9-ad6f-671b4b63dbab","Type":"ContainerStarted","Data":"653ec4bccd431aa1239586bab65a2f72f88db99e3bb6c36e266e57c143751433"} Jan 26 17:01:08 crc kubenswrapper[4754]: I0126 17:01:08.036983 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-zgmpc" Jan 26 17:01:08 crc kubenswrapper[4754]: I0126 17:01:08.042772 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-qfmmc" event={"ID":"73c39dd3-4d2e-4ee8-a218-f88eb59f36b7","Type":"ContainerStarted","Data":"a7d494ba213262f2f1e611a49757f7d80140dcec0f5962bac9773d9593a484ba"} Jan 26 17:01:08 crc kubenswrapper[4754]: I0126 17:01:08.043567 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-qfmmc" Jan 26 17:01:08 crc kubenswrapper[4754]: I0126 17:01:08.046337 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-54757c584b-vvxrw" event={"ID":"6506a03e-2445-4e3c-9814-f5bb1cfbaa06","Type":"ContainerStarted","Data":"0a23f0d277a58950e9ff2d8f2ab91c72eda1458771b73de8f15e7ad6fe952464"} Jan 26 17:01:08 crc kubenswrapper[4754]: I0126 17:01:08.051642 4754 scope.go:117] "RemoveContainer" containerID="6a78fceb2d0ae05ec20b59565e6cbd1147cf896d7a3500f7f2b6f02a6b1c38e6" Jan 26 17:01:08 crc kubenswrapper[4754]: I0126 17:01:08.057054 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-dxcqw" podStartSLOduration=2.049184491 podStartE2EDuration="5.057036834s" podCreationTimestamp="2026-01-26 17:01:03 +0000 UTC" firstStartedPulling="2026-01-26 17:01:04.373425447 +0000 UTC m=+830.897605891" lastFinishedPulling="2026-01-26 17:01:07.3812778 +0000 UTC m=+833.905458234" observedRunningTime="2026-01-26 17:01:08.052771416 +0000 UTC m=+834.576951860" watchObservedRunningTime="2026-01-26 17:01:08.057036834 +0000 UTC m=+834.581217268" Jan 26 17:01:08 crc kubenswrapper[4754]: I0126 17:01:08.077245 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-f9ghz"] Jan 26 17:01:08 crc kubenswrapper[4754]: I0126 17:01:08.080142 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-f9ghz"] Jan 26 17:01:08 crc kubenswrapper[4754]: I0126 17:01:08.093120 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-zgmpc" podStartSLOduration=1.769248784 podStartE2EDuration="5.093099377s" podCreationTimestamp="2026-01-26 17:01:03 +0000 UTC" firstStartedPulling="2026-01-26 17:01:04.055225866 +0000 UTC m=+830.579406300" lastFinishedPulling="2026-01-26 17:01:07.379076449 +0000 UTC m=+833.903256893" observedRunningTime="2026-01-26 17:01:08.087591606 +0000 UTC m=+834.611772040" watchObservedRunningTime="2026-01-26 17:01:08.093099377 +0000 UTC m=+834.617279821" Jan 26 17:01:08 crc kubenswrapper[4754]: I0126 17:01:08.093375 4754 scope.go:117] "RemoveContainer" containerID="d38db17138431381a74f7c6df9152e77d0c18b7134aae7bed72eb1449bcbb606" Jan 26 17:01:08 crc kubenswrapper[4754]: I0126 17:01:08.112881 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-qfmmc" podStartSLOduration=1.656154482 podStartE2EDuration="5.112844351s" podCreationTimestamp="2026-01-26 17:01:03 +0000 UTC" firstStartedPulling="2026-01-26 17:01:03.939435959 +0000 UTC m=+830.463616393" lastFinishedPulling="2026-01-26 17:01:07.396125828 +0000 UTC m=+833.920306262" observedRunningTime="2026-01-26 17:01:08.106831045 +0000 UTC m=+834.631011479" watchObservedRunningTime="2026-01-26 17:01:08.112844351 +0000 UTC m=+834.637024795" Jan 26 17:01:08 crc kubenswrapper[4754]: I0126 17:01:08.118850 4754 scope.go:117] "RemoveContainer" containerID="a0061f692d9c41eae4e5cbb88ab79b1cc8b77338db27b33a9852ac7353d8fbc9" Jan 26 17:01:08 crc kubenswrapper[4754]: E0126 17:01:08.119700 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a0061f692d9c41eae4e5cbb88ab79b1cc8b77338db27b33a9852ac7353d8fbc9\": container with ID starting with a0061f692d9c41eae4e5cbb88ab79b1cc8b77338db27b33a9852ac7353d8fbc9 not found: ID does not exist" containerID="a0061f692d9c41eae4e5cbb88ab79b1cc8b77338db27b33a9852ac7353d8fbc9" Jan 26 17:01:08 crc kubenswrapper[4754]: I0126 17:01:08.119797 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a0061f692d9c41eae4e5cbb88ab79b1cc8b77338db27b33a9852ac7353d8fbc9"} err="failed to get container status \"a0061f692d9c41eae4e5cbb88ab79b1cc8b77338db27b33a9852ac7353d8fbc9\": rpc error: code = NotFound desc = could not find container \"a0061f692d9c41eae4e5cbb88ab79b1cc8b77338db27b33a9852ac7353d8fbc9\": container with ID starting with a0061f692d9c41eae4e5cbb88ab79b1cc8b77338db27b33a9852ac7353d8fbc9 not found: ID does not exist" Jan 26 17:01:08 crc kubenswrapper[4754]: I0126 17:01:08.119843 4754 scope.go:117] "RemoveContainer" containerID="6a78fceb2d0ae05ec20b59565e6cbd1147cf896d7a3500f7f2b6f02a6b1c38e6" Jan 26 17:01:08 crc kubenswrapper[4754]: E0126 17:01:08.120351 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6a78fceb2d0ae05ec20b59565e6cbd1147cf896d7a3500f7f2b6f02a6b1c38e6\": container with ID starting with 6a78fceb2d0ae05ec20b59565e6cbd1147cf896d7a3500f7f2b6f02a6b1c38e6 not found: ID does not exist" containerID="6a78fceb2d0ae05ec20b59565e6cbd1147cf896d7a3500f7f2b6f02a6b1c38e6" Jan 26 17:01:08 crc kubenswrapper[4754]: I0126 17:01:08.120416 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6a78fceb2d0ae05ec20b59565e6cbd1147cf896d7a3500f7f2b6f02a6b1c38e6"} err="failed to get container status \"6a78fceb2d0ae05ec20b59565e6cbd1147cf896d7a3500f7f2b6f02a6b1c38e6\": rpc error: code = NotFound desc = could not find container \"6a78fceb2d0ae05ec20b59565e6cbd1147cf896d7a3500f7f2b6f02a6b1c38e6\": container with ID starting with 6a78fceb2d0ae05ec20b59565e6cbd1147cf896d7a3500f7f2b6f02a6b1c38e6 not found: ID does not exist" Jan 26 17:01:08 crc kubenswrapper[4754]: I0126 17:01:08.120430 4754 scope.go:117] "RemoveContainer" containerID="d38db17138431381a74f7c6df9152e77d0c18b7134aae7bed72eb1449bcbb606" Jan 26 17:01:08 crc kubenswrapper[4754]: E0126 17:01:08.120745 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d38db17138431381a74f7c6df9152e77d0c18b7134aae7bed72eb1449bcbb606\": container with ID starting with d38db17138431381a74f7c6df9152e77d0c18b7134aae7bed72eb1449bcbb606 not found: ID does not exist" containerID="d38db17138431381a74f7c6df9152e77d0c18b7134aae7bed72eb1449bcbb606" Jan 26 17:01:08 crc kubenswrapper[4754]: I0126 17:01:08.120767 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d38db17138431381a74f7c6df9152e77d0c18b7134aae7bed72eb1449bcbb606"} err="failed to get container status \"d38db17138431381a74f7c6df9152e77d0c18b7134aae7bed72eb1449bcbb606\": rpc error: code = NotFound desc = could not find container \"d38db17138431381a74f7c6df9152e77d0c18b7134aae7bed72eb1449bcbb606\": container with ID starting with d38db17138431381a74f7c6df9152e77d0c18b7134aae7bed72eb1449bcbb606 not found: ID does not exist" Jan 26 17:01:09 crc kubenswrapper[4754]: I0126 17:01:09.775722 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b90dc7aa-8bf8-4770-9302-6e7186bbbd86" path="/var/lib/kubelet/pods/b90dc7aa-8bf8-4770-9302-6e7186bbbd86/volumes" Jan 26 17:01:11 crc kubenswrapper[4754]: I0126 17:01:11.074459 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-54757c584b-vvxrw" event={"ID":"6506a03e-2445-4e3c-9814-f5bb1cfbaa06","Type":"ContainerStarted","Data":"dcade1479fd61165b72b26ad55b55c3921ccdd9b3aa2a4cc61ddb6eb0968a524"} Jan 26 17:01:11 crc kubenswrapper[4754]: I0126 17:01:11.108662 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-54757c584b-vvxrw" podStartSLOduration=2.168077665 podStartE2EDuration="8.10862835s" podCreationTimestamp="2026-01-26 17:01:03 +0000 UTC" firstStartedPulling="2026-01-26 17:01:04.116336469 +0000 UTC m=+830.640516903" lastFinishedPulling="2026-01-26 17:01:10.056887154 +0000 UTC m=+836.581067588" observedRunningTime="2026-01-26 17:01:11.099802927 +0000 UTC m=+837.623983441" watchObservedRunningTime="2026-01-26 17:01:11.10862835 +0000 UTC m=+837.632808814" Jan 26 17:01:13 crc kubenswrapper[4754]: I0126 17:01:13.937305 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-qfmmc" Jan 26 17:01:14 crc kubenswrapper[4754]: I0126 17:01:14.164879 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-75765cbdc-w5c9p" Jan 26 17:01:14 crc kubenswrapper[4754]: I0126 17:01:14.165224 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-75765cbdc-w5c9p" Jan 26 17:01:14 crc kubenswrapper[4754]: I0126 17:01:14.171057 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-75765cbdc-w5c9p" Jan 26 17:01:15 crc kubenswrapper[4754]: I0126 17:01:15.108984 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-75765cbdc-w5c9p" Jan 26 17:01:15 crc kubenswrapper[4754]: I0126 17:01:15.162953 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-5p6w2"] Jan 26 17:01:23 crc kubenswrapper[4754]: I0126 17:01:23.881422 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-zgmpc" Jan 26 17:01:39 crc kubenswrapper[4754]: I0126 17:01:39.014952 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dckzdjj"] Jan 26 17:01:39 crc kubenswrapper[4754]: E0126 17:01:39.015546 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b90dc7aa-8bf8-4770-9302-6e7186bbbd86" containerName="extract-content" Jan 26 17:01:39 crc kubenswrapper[4754]: I0126 17:01:39.015558 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="b90dc7aa-8bf8-4770-9302-6e7186bbbd86" containerName="extract-content" Jan 26 17:01:39 crc kubenswrapper[4754]: E0126 17:01:39.015575 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b90dc7aa-8bf8-4770-9302-6e7186bbbd86" containerName="registry-server" Jan 26 17:01:39 crc kubenswrapper[4754]: I0126 17:01:39.015580 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="b90dc7aa-8bf8-4770-9302-6e7186bbbd86" containerName="registry-server" Jan 26 17:01:39 crc kubenswrapper[4754]: E0126 17:01:39.015591 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b90dc7aa-8bf8-4770-9302-6e7186bbbd86" containerName="extract-utilities" Jan 26 17:01:39 crc kubenswrapper[4754]: I0126 17:01:39.015598 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="b90dc7aa-8bf8-4770-9302-6e7186bbbd86" containerName="extract-utilities" Jan 26 17:01:39 crc kubenswrapper[4754]: I0126 17:01:39.015723 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="b90dc7aa-8bf8-4770-9302-6e7186bbbd86" containerName="registry-server" Jan 26 17:01:39 crc kubenswrapper[4754]: I0126 17:01:39.016418 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dckzdjj" Jan 26 17:01:39 crc kubenswrapper[4754]: I0126 17:01:39.018175 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Jan 26 17:01:39 crc kubenswrapper[4754]: I0126 17:01:39.022338 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dckzdjj"] Jan 26 17:01:39 crc kubenswrapper[4754]: I0126 17:01:39.090742 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/997bf4f3-9268-4058-9ee7-80c581651bcd-bundle\") pod \"270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dckzdjj\" (UID: \"997bf4f3-9268-4058-9ee7-80c581651bcd\") " pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dckzdjj" Jan 26 17:01:39 crc kubenswrapper[4754]: I0126 17:01:39.091074 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2l85s\" (UniqueName: \"kubernetes.io/projected/997bf4f3-9268-4058-9ee7-80c581651bcd-kube-api-access-2l85s\") pod \"270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dckzdjj\" (UID: \"997bf4f3-9268-4058-9ee7-80c581651bcd\") " pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dckzdjj" Jan 26 17:01:39 crc kubenswrapper[4754]: I0126 17:01:39.091204 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/997bf4f3-9268-4058-9ee7-80c581651bcd-util\") pod \"270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dckzdjj\" (UID: \"997bf4f3-9268-4058-9ee7-80c581651bcd\") " pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dckzdjj" Jan 26 17:01:39 crc kubenswrapper[4754]: I0126 17:01:39.191825 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/997bf4f3-9268-4058-9ee7-80c581651bcd-util\") pod \"270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dckzdjj\" (UID: \"997bf4f3-9268-4058-9ee7-80c581651bcd\") " pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dckzdjj" Jan 26 17:01:39 crc kubenswrapper[4754]: I0126 17:01:39.191888 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/997bf4f3-9268-4058-9ee7-80c581651bcd-bundle\") pod \"270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dckzdjj\" (UID: \"997bf4f3-9268-4058-9ee7-80c581651bcd\") " pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dckzdjj" Jan 26 17:01:39 crc kubenswrapper[4754]: I0126 17:01:39.191948 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2l85s\" (UniqueName: \"kubernetes.io/projected/997bf4f3-9268-4058-9ee7-80c581651bcd-kube-api-access-2l85s\") pod \"270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dckzdjj\" (UID: \"997bf4f3-9268-4058-9ee7-80c581651bcd\") " pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dckzdjj" Jan 26 17:01:39 crc kubenswrapper[4754]: I0126 17:01:39.192534 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/997bf4f3-9268-4058-9ee7-80c581651bcd-bundle\") pod \"270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dckzdjj\" (UID: \"997bf4f3-9268-4058-9ee7-80c581651bcd\") " pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dckzdjj" Jan 26 17:01:39 crc kubenswrapper[4754]: I0126 17:01:39.192806 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/997bf4f3-9268-4058-9ee7-80c581651bcd-util\") pod \"270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dckzdjj\" (UID: \"997bf4f3-9268-4058-9ee7-80c581651bcd\") " pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dckzdjj" Jan 26 17:01:39 crc kubenswrapper[4754]: I0126 17:01:39.217102 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2l85s\" (UniqueName: \"kubernetes.io/projected/997bf4f3-9268-4058-9ee7-80c581651bcd-kube-api-access-2l85s\") pod \"270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dckzdjj\" (UID: \"997bf4f3-9268-4058-9ee7-80c581651bcd\") " pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dckzdjj" Jan 26 17:01:39 crc kubenswrapper[4754]: I0126 17:01:39.332321 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dckzdjj" Jan 26 17:01:39 crc kubenswrapper[4754]: I0126 17:01:39.517891 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dckzdjj"] Jan 26 17:01:40 crc kubenswrapper[4754]: I0126 17:01:40.210261 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-5p6w2" podUID="528dc2c4-71cc-46cb-b847-426bb8638188" containerName="console" containerID="cri-o://283c7af4a93330b2596db693edb773b68c3ccab0a3c72eaf457b334081943e8f" gracePeriod=15 Jan 26 17:01:40 crc kubenswrapper[4754]: I0126 17:01:40.257192 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dckzdjj" event={"ID":"997bf4f3-9268-4058-9ee7-80c581651bcd","Type":"ContainerStarted","Data":"49a6742e91d3f0cd9436c00d7ee241a84fae0245b2ef03975ca96b943f26141c"} Jan 26 17:01:41 crc kubenswrapper[4754]: I0126 17:01:41.128435 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-5p6w2_528dc2c4-71cc-46cb-b847-426bb8638188/console/0.log" Jan 26 17:01:41 crc kubenswrapper[4754]: I0126 17:01:41.128492 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-5p6w2" Jan 26 17:01:41 crc kubenswrapper[4754]: I0126 17:01:41.219331 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/528dc2c4-71cc-46cb-b847-426bb8638188-console-oauth-config\") pod \"528dc2c4-71cc-46cb-b847-426bb8638188\" (UID: \"528dc2c4-71cc-46cb-b847-426bb8638188\") " Jan 26 17:01:41 crc kubenswrapper[4754]: I0126 17:01:41.219386 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/528dc2c4-71cc-46cb-b847-426bb8638188-console-serving-cert\") pod \"528dc2c4-71cc-46cb-b847-426bb8638188\" (UID: \"528dc2c4-71cc-46cb-b847-426bb8638188\") " Jan 26 17:01:41 crc kubenswrapper[4754]: I0126 17:01:41.219410 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/528dc2c4-71cc-46cb-b847-426bb8638188-console-config\") pod \"528dc2c4-71cc-46cb-b847-426bb8638188\" (UID: \"528dc2c4-71cc-46cb-b847-426bb8638188\") " Jan 26 17:01:41 crc kubenswrapper[4754]: I0126 17:01:41.219429 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/528dc2c4-71cc-46cb-b847-426bb8638188-oauth-serving-cert\") pod \"528dc2c4-71cc-46cb-b847-426bb8638188\" (UID: \"528dc2c4-71cc-46cb-b847-426bb8638188\") " Jan 26 17:01:41 crc kubenswrapper[4754]: I0126 17:01:41.219462 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5jq47\" (UniqueName: \"kubernetes.io/projected/528dc2c4-71cc-46cb-b847-426bb8638188-kube-api-access-5jq47\") pod \"528dc2c4-71cc-46cb-b847-426bb8638188\" (UID: \"528dc2c4-71cc-46cb-b847-426bb8638188\") " Jan 26 17:01:41 crc kubenswrapper[4754]: I0126 17:01:41.219481 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/528dc2c4-71cc-46cb-b847-426bb8638188-service-ca\") pod \"528dc2c4-71cc-46cb-b847-426bb8638188\" (UID: \"528dc2c4-71cc-46cb-b847-426bb8638188\") " Jan 26 17:01:41 crc kubenswrapper[4754]: I0126 17:01:41.219503 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/528dc2c4-71cc-46cb-b847-426bb8638188-trusted-ca-bundle\") pod \"528dc2c4-71cc-46cb-b847-426bb8638188\" (UID: \"528dc2c4-71cc-46cb-b847-426bb8638188\") " Jan 26 17:01:41 crc kubenswrapper[4754]: I0126 17:01:41.220468 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/528dc2c4-71cc-46cb-b847-426bb8638188-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "528dc2c4-71cc-46cb-b847-426bb8638188" (UID: "528dc2c4-71cc-46cb-b847-426bb8638188"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:01:41 crc kubenswrapper[4754]: I0126 17:01:41.220496 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/528dc2c4-71cc-46cb-b847-426bb8638188-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "528dc2c4-71cc-46cb-b847-426bb8638188" (UID: "528dc2c4-71cc-46cb-b847-426bb8638188"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:01:41 crc kubenswrapper[4754]: I0126 17:01:41.220513 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/528dc2c4-71cc-46cb-b847-426bb8638188-service-ca" (OuterVolumeSpecName: "service-ca") pod "528dc2c4-71cc-46cb-b847-426bb8638188" (UID: "528dc2c4-71cc-46cb-b847-426bb8638188"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:01:41 crc kubenswrapper[4754]: I0126 17:01:41.220587 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/528dc2c4-71cc-46cb-b847-426bb8638188-console-config" (OuterVolumeSpecName: "console-config") pod "528dc2c4-71cc-46cb-b847-426bb8638188" (UID: "528dc2c4-71cc-46cb-b847-426bb8638188"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:01:41 crc kubenswrapper[4754]: I0126 17:01:41.225403 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/528dc2c4-71cc-46cb-b847-426bb8638188-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "528dc2c4-71cc-46cb-b847-426bb8638188" (UID: "528dc2c4-71cc-46cb-b847-426bb8638188"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:01:41 crc kubenswrapper[4754]: I0126 17:01:41.226470 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/528dc2c4-71cc-46cb-b847-426bb8638188-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "528dc2c4-71cc-46cb-b847-426bb8638188" (UID: "528dc2c4-71cc-46cb-b847-426bb8638188"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:01:41 crc kubenswrapper[4754]: I0126 17:01:41.226790 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/528dc2c4-71cc-46cb-b847-426bb8638188-kube-api-access-5jq47" (OuterVolumeSpecName: "kube-api-access-5jq47") pod "528dc2c4-71cc-46cb-b847-426bb8638188" (UID: "528dc2c4-71cc-46cb-b847-426bb8638188"). InnerVolumeSpecName "kube-api-access-5jq47". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:01:41 crc kubenswrapper[4754]: I0126 17:01:41.266098 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-5p6w2_528dc2c4-71cc-46cb-b847-426bb8638188/console/0.log" Jan 26 17:01:41 crc kubenswrapper[4754]: I0126 17:01:41.266154 4754 generic.go:334] "Generic (PLEG): container finished" podID="528dc2c4-71cc-46cb-b847-426bb8638188" containerID="283c7af4a93330b2596db693edb773b68c3ccab0a3c72eaf457b334081943e8f" exitCode=2 Jan 26 17:01:41 crc kubenswrapper[4754]: I0126 17:01:41.266216 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-5p6w2" event={"ID":"528dc2c4-71cc-46cb-b847-426bb8638188","Type":"ContainerDied","Data":"283c7af4a93330b2596db693edb773b68c3ccab0a3c72eaf457b334081943e8f"} Jan 26 17:01:41 crc kubenswrapper[4754]: I0126 17:01:41.266245 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-5p6w2" event={"ID":"528dc2c4-71cc-46cb-b847-426bb8638188","Type":"ContainerDied","Data":"3f08ff855a9c086991567039090cb37f1bc644f627cc0ba3983e9ab1d9ca0f38"} Jan 26 17:01:41 crc kubenswrapper[4754]: I0126 17:01:41.266263 4754 scope.go:117] "RemoveContainer" containerID="283c7af4a93330b2596db693edb773b68c3ccab0a3c72eaf457b334081943e8f" Jan 26 17:01:41 crc kubenswrapper[4754]: I0126 17:01:41.266370 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-5p6w2" Jan 26 17:01:41 crc kubenswrapper[4754]: I0126 17:01:41.268189 4754 generic.go:334] "Generic (PLEG): container finished" podID="997bf4f3-9268-4058-9ee7-80c581651bcd" containerID="bfa91d3362719415344b4e0d01ac7820b8a8953c83e25110cc06e8bc7afaec8a" exitCode=0 Jan 26 17:01:41 crc kubenswrapper[4754]: I0126 17:01:41.268218 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dckzdjj" event={"ID":"997bf4f3-9268-4058-9ee7-80c581651bcd","Type":"ContainerDied","Data":"bfa91d3362719415344b4e0d01ac7820b8a8953c83e25110cc06e8bc7afaec8a"} Jan 26 17:01:41 crc kubenswrapper[4754]: I0126 17:01:41.283827 4754 scope.go:117] "RemoveContainer" containerID="283c7af4a93330b2596db693edb773b68c3ccab0a3c72eaf457b334081943e8f" Jan 26 17:01:41 crc kubenswrapper[4754]: E0126 17:01:41.284592 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"283c7af4a93330b2596db693edb773b68c3ccab0a3c72eaf457b334081943e8f\": container with ID starting with 283c7af4a93330b2596db693edb773b68c3ccab0a3c72eaf457b334081943e8f not found: ID does not exist" containerID="283c7af4a93330b2596db693edb773b68c3ccab0a3c72eaf457b334081943e8f" Jan 26 17:01:41 crc kubenswrapper[4754]: I0126 17:01:41.284863 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"283c7af4a93330b2596db693edb773b68c3ccab0a3c72eaf457b334081943e8f"} err="failed to get container status \"283c7af4a93330b2596db693edb773b68c3ccab0a3c72eaf457b334081943e8f\": rpc error: code = NotFound desc = could not find container \"283c7af4a93330b2596db693edb773b68c3ccab0a3c72eaf457b334081943e8f\": container with ID starting with 283c7af4a93330b2596db693edb773b68c3ccab0a3c72eaf457b334081943e8f not found: ID does not exist" Jan 26 17:01:41 crc kubenswrapper[4754]: I0126 17:01:41.302340 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-5p6w2"] Jan 26 17:01:41 crc kubenswrapper[4754]: I0126 17:01:41.306148 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-5p6w2"] Jan 26 17:01:41 crc kubenswrapper[4754]: I0126 17:01:41.320422 4754 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/528dc2c4-71cc-46cb-b847-426bb8638188-console-oauth-config\") on node \"crc\" DevicePath \"\"" Jan 26 17:01:41 crc kubenswrapper[4754]: I0126 17:01:41.320455 4754 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/528dc2c4-71cc-46cb-b847-426bb8638188-console-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 26 17:01:41 crc kubenswrapper[4754]: I0126 17:01:41.320463 4754 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/528dc2c4-71cc-46cb-b847-426bb8638188-console-config\") on node \"crc\" DevicePath \"\"" Jan 26 17:01:41 crc kubenswrapper[4754]: I0126 17:01:41.320473 4754 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/528dc2c4-71cc-46cb-b847-426bb8638188-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 26 17:01:41 crc kubenswrapper[4754]: I0126 17:01:41.320482 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5jq47\" (UniqueName: \"kubernetes.io/projected/528dc2c4-71cc-46cb-b847-426bb8638188-kube-api-access-5jq47\") on node \"crc\" DevicePath \"\"" Jan 26 17:01:41 crc kubenswrapper[4754]: I0126 17:01:41.320491 4754 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/528dc2c4-71cc-46cb-b847-426bb8638188-service-ca\") on node \"crc\" DevicePath \"\"" Jan 26 17:01:41 crc kubenswrapper[4754]: I0126 17:01:41.320501 4754 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/528dc2c4-71cc-46cb-b847-426bb8638188-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 26 17:01:41 crc kubenswrapper[4754]: I0126 17:01:41.780350 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="528dc2c4-71cc-46cb-b847-426bb8638188" path="/var/lib/kubelet/pods/528dc2c4-71cc-46cb-b847-426bb8638188/volumes" Jan 26 17:01:48 crc kubenswrapper[4754]: I0126 17:01:48.322424 4754 generic.go:334] "Generic (PLEG): container finished" podID="997bf4f3-9268-4058-9ee7-80c581651bcd" containerID="ead238776e73483ef681de7097a962c33aa9581116f4829b8e98ba68099f2b51" exitCode=0 Jan 26 17:01:48 crc kubenswrapper[4754]: I0126 17:01:48.322485 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dckzdjj" event={"ID":"997bf4f3-9268-4058-9ee7-80c581651bcd","Type":"ContainerDied","Data":"ead238776e73483ef681de7097a962c33aa9581116f4829b8e98ba68099f2b51"} Jan 26 17:01:49 crc kubenswrapper[4754]: I0126 17:01:49.331791 4754 generic.go:334] "Generic (PLEG): container finished" podID="997bf4f3-9268-4058-9ee7-80c581651bcd" containerID="941c2242484f807446d3ebee2cb9eeca9839ec0a161243135e46ad6922a08b7b" exitCode=0 Jan 26 17:01:49 crc kubenswrapper[4754]: I0126 17:01:49.331890 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dckzdjj" event={"ID":"997bf4f3-9268-4058-9ee7-80c581651bcd","Type":"ContainerDied","Data":"941c2242484f807446d3ebee2cb9eeca9839ec0a161243135e46ad6922a08b7b"} Jan 26 17:01:50 crc kubenswrapper[4754]: I0126 17:01:50.598052 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dckzdjj" Jan 26 17:01:50 crc kubenswrapper[4754]: I0126 17:01:50.643717 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2l85s\" (UniqueName: \"kubernetes.io/projected/997bf4f3-9268-4058-9ee7-80c581651bcd-kube-api-access-2l85s\") pod \"997bf4f3-9268-4058-9ee7-80c581651bcd\" (UID: \"997bf4f3-9268-4058-9ee7-80c581651bcd\") " Jan 26 17:01:50 crc kubenswrapper[4754]: I0126 17:01:50.643768 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/997bf4f3-9268-4058-9ee7-80c581651bcd-bundle\") pod \"997bf4f3-9268-4058-9ee7-80c581651bcd\" (UID: \"997bf4f3-9268-4058-9ee7-80c581651bcd\") " Jan 26 17:01:50 crc kubenswrapper[4754]: I0126 17:01:50.643857 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/997bf4f3-9268-4058-9ee7-80c581651bcd-util\") pod \"997bf4f3-9268-4058-9ee7-80c581651bcd\" (UID: \"997bf4f3-9268-4058-9ee7-80c581651bcd\") " Jan 26 17:01:50 crc kubenswrapper[4754]: I0126 17:01:50.646259 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/997bf4f3-9268-4058-9ee7-80c581651bcd-bundle" (OuterVolumeSpecName: "bundle") pod "997bf4f3-9268-4058-9ee7-80c581651bcd" (UID: "997bf4f3-9268-4058-9ee7-80c581651bcd"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:01:50 crc kubenswrapper[4754]: I0126 17:01:50.650632 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/997bf4f3-9268-4058-9ee7-80c581651bcd-kube-api-access-2l85s" (OuterVolumeSpecName: "kube-api-access-2l85s") pod "997bf4f3-9268-4058-9ee7-80c581651bcd" (UID: "997bf4f3-9268-4058-9ee7-80c581651bcd"). InnerVolumeSpecName "kube-api-access-2l85s". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:01:50 crc kubenswrapper[4754]: I0126 17:01:50.655281 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/997bf4f3-9268-4058-9ee7-80c581651bcd-util" (OuterVolumeSpecName: "util") pod "997bf4f3-9268-4058-9ee7-80c581651bcd" (UID: "997bf4f3-9268-4058-9ee7-80c581651bcd"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:01:50 crc kubenswrapper[4754]: I0126 17:01:50.745489 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2l85s\" (UniqueName: \"kubernetes.io/projected/997bf4f3-9268-4058-9ee7-80c581651bcd-kube-api-access-2l85s\") on node \"crc\" DevicePath \"\"" Jan 26 17:01:50 crc kubenswrapper[4754]: I0126 17:01:50.745548 4754 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/997bf4f3-9268-4058-9ee7-80c581651bcd-bundle\") on node \"crc\" DevicePath \"\"" Jan 26 17:01:50 crc kubenswrapper[4754]: I0126 17:01:50.745567 4754 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/997bf4f3-9268-4058-9ee7-80c581651bcd-util\") on node \"crc\" DevicePath \"\"" Jan 26 17:01:51 crc kubenswrapper[4754]: I0126 17:01:51.352211 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dckzdjj" event={"ID":"997bf4f3-9268-4058-9ee7-80c581651bcd","Type":"ContainerDied","Data":"49a6742e91d3f0cd9436c00d7ee241a84fae0245b2ef03975ca96b943f26141c"} Jan 26 17:01:51 crc kubenswrapper[4754]: I0126 17:01:51.352328 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="49a6742e91d3f0cd9436c00d7ee241a84fae0245b2ef03975ca96b943f26141c" Jan 26 17:01:51 crc kubenswrapper[4754]: I0126 17:01:51.352447 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dckzdjj" Jan 26 17:02:02 crc kubenswrapper[4754]: I0126 17:02:02.216185 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-755bc5d786-g2p48"] Jan 26 17:02:02 crc kubenswrapper[4754]: E0126 17:02:02.221575 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="528dc2c4-71cc-46cb-b847-426bb8638188" containerName="console" Jan 26 17:02:02 crc kubenswrapper[4754]: I0126 17:02:02.221640 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="528dc2c4-71cc-46cb-b847-426bb8638188" containerName="console" Jan 26 17:02:02 crc kubenswrapper[4754]: E0126 17:02:02.221712 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="997bf4f3-9268-4058-9ee7-80c581651bcd" containerName="extract" Jan 26 17:02:02 crc kubenswrapper[4754]: I0126 17:02:02.221722 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="997bf4f3-9268-4058-9ee7-80c581651bcd" containerName="extract" Jan 26 17:02:02 crc kubenswrapper[4754]: E0126 17:02:02.221748 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="997bf4f3-9268-4058-9ee7-80c581651bcd" containerName="pull" Jan 26 17:02:02 crc kubenswrapper[4754]: I0126 17:02:02.221768 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="997bf4f3-9268-4058-9ee7-80c581651bcd" containerName="pull" Jan 26 17:02:02 crc kubenswrapper[4754]: E0126 17:02:02.221780 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="997bf4f3-9268-4058-9ee7-80c581651bcd" containerName="util" Jan 26 17:02:02 crc kubenswrapper[4754]: I0126 17:02:02.221791 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="997bf4f3-9268-4058-9ee7-80c581651bcd" containerName="util" Jan 26 17:02:02 crc kubenswrapper[4754]: I0126 17:02:02.222280 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="528dc2c4-71cc-46cb-b847-426bb8638188" containerName="console" Jan 26 17:02:02 crc kubenswrapper[4754]: I0126 17:02:02.222319 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="997bf4f3-9268-4058-9ee7-80c581651bcd" containerName="extract" Jan 26 17:02:02 crc kubenswrapper[4754]: I0126 17:02:02.222948 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-755bc5d786-g2p48" Jan 26 17:02:02 crc kubenswrapper[4754]: I0126 17:02:02.227268 4754 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-62pw8" Jan 26 17:02:02 crc kubenswrapper[4754]: I0126 17:02:02.227793 4754 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Jan 26 17:02:02 crc kubenswrapper[4754]: I0126 17:02:02.227896 4754 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Jan 26 17:02:02 crc kubenswrapper[4754]: I0126 17:02:02.227924 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Jan 26 17:02:02 crc kubenswrapper[4754]: I0126 17:02:02.228037 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Jan 26 17:02:02 crc kubenswrapper[4754]: I0126 17:02:02.230886 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-755bc5d786-g2p48"] Jan 26 17:02:02 crc kubenswrapper[4754]: I0126 17:02:02.300930 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/83c99759-10ef-42eb-a58f-23c4e6c7d089-apiservice-cert\") pod \"metallb-operator-controller-manager-755bc5d786-g2p48\" (UID: \"83c99759-10ef-42eb-a58f-23c4e6c7d089\") " pod="metallb-system/metallb-operator-controller-manager-755bc5d786-g2p48" Jan 26 17:02:02 crc kubenswrapper[4754]: I0126 17:02:02.301020 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/83c99759-10ef-42eb-a58f-23c4e6c7d089-webhook-cert\") pod \"metallb-operator-controller-manager-755bc5d786-g2p48\" (UID: \"83c99759-10ef-42eb-a58f-23c4e6c7d089\") " pod="metallb-system/metallb-operator-controller-manager-755bc5d786-g2p48" Jan 26 17:02:02 crc kubenswrapper[4754]: I0126 17:02:02.301079 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-djgs9\" (UniqueName: \"kubernetes.io/projected/83c99759-10ef-42eb-a58f-23c4e6c7d089-kube-api-access-djgs9\") pod \"metallb-operator-controller-manager-755bc5d786-g2p48\" (UID: \"83c99759-10ef-42eb-a58f-23c4e6c7d089\") " pod="metallb-system/metallb-operator-controller-manager-755bc5d786-g2p48" Jan 26 17:02:02 crc kubenswrapper[4754]: I0126 17:02:02.402617 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/83c99759-10ef-42eb-a58f-23c4e6c7d089-webhook-cert\") pod \"metallb-operator-controller-manager-755bc5d786-g2p48\" (UID: \"83c99759-10ef-42eb-a58f-23c4e6c7d089\") " pod="metallb-system/metallb-operator-controller-manager-755bc5d786-g2p48" Jan 26 17:02:02 crc kubenswrapper[4754]: I0126 17:02:02.402768 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-djgs9\" (UniqueName: \"kubernetes.io/projected/83c99759-10ef-42eb-a58f-23c4e6c7d089-kube-api-access-djgs9\") pod \"metallb-operator-controller-manager-755bc5d786-g2p48\" (UID: \"83c99759-10ef-42eb-a58f-23c4e6c7d089\") " pod="metallb-system/metallb-operator-controller-manager-755bc5d786-g2p48" Jan 26 17:02:02 crc kubenswrapper[4754]: I0126 17:02:02.402802 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/83c99759-10ef-42eb-a58f-23c4e6c7d089-apiservice-cert\") pod \"metallb-operator-controller-manager-755bc5d786-g2p48\" (UID: \"83c99759-10ef-42eb-a58f-23c4e6c7d089\") " pod="metallb-system/metallb-operator-controller-manager-755bc5d786-g2p48" Jan 26 17:02:02 crc kubenswrapper[4754]: I0126 17:02:02.408498 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/83c99759-10ef-42eb-a58f-23c4e6c7d089-webhook-cert\") pod \"metallb-operator-controller-manager-755bc5d786-g2p48\" (UID: \"83c99759-10ef-42eb-a58f-23c4e6c7d089\") " pod="metallb-system/metallb-operator-controller-manager-755bc5d786-g2p48" Jan 26 17:02:02 crc kubenswrapper[4754]: I0126 17:02:02.411089 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/83c99759-10ef-42eb-a58f-23c4e6c7d089-apiservice-cert\") pod \"metallb-operator-controller-manager-755bc5d786-g2p48\" (UID: \"83c99759-10ef-42eb-a58f-23c4e6c7d089\") " pod="metallb-system/metallb-operator-controller-manager-755bc5d786-g2p48" Jan 26 17:02:02 crc kubenswrapper[4754]: I0126 17:02:02.433347 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-djgs9\" (UniqueName: \"kubernetes.io/projected/83c99759-10ef-42eb-a58f-23c4e6c7d089-kube-api-access-djgs9\") pod \"metallb-operator-controller-manager-755bc5d786-g2p48\" (UID: \"83c99759-10ef-42eb-a58f-23c4e6c7d089\") " pod="metallb-system/metallb-operator-controller-manager-755bc5d786-g2p48" Jan 26 17:02:02 crc kubenswrapper[4754]: I0126 17:02:02.532049 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-755c486f67-b6mn4"] Jan 26 17:02:02 crc kubenswrapper[4754]: I0126 17:02:02.532728 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-755c486f67-b6mn4" Jan 26 17:02:02 crc kubenswrapper[4754]: I0126 17:02:02.536744 4754 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Jan 26 17:02:02 crc kubenswrapper[4754]: I0126 17:02:02.537999 4754 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-8xfkn" Jan 26 17:02:02 crc kubenswrapper[4754]: I0126 17:02:02.547203 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-755bc5d786-g2p48" Jan 26 17:02:02 crc kubenswrapper[4754]: I0126 17:02:02.549646 4754 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Jan 26 17:02:02 crc kubenswrapper[4754]: I0126 17:02:02.600622 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-755c486f67-b6mn4"] Jan 26 17:02:02 crc kubenswrapper[4754]: I0126 17:02:02.604182 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/ee80d5f4-a13e-434d-b250-1e005d84dc59-apiservice-cert\") pod \"metallb-operator-webhook-server-755c486f67-b6mn4\" (UID: \"ee80d5f4-a13e-434d-b250-1e005d84dc59\") " pod="metallb-system/metallb-operator-webhook-server-755c486f67-b6mn4" Jan 26 17:02:02 crc kubenswrapper[4754]: I0126 17:02:02.604232 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nmltz\" (UniqueName: \"kubernetes.io/projected/ee80d5f4-a13e-434d-b250-1e005d84dc59-kube-api-access-nmltz\") pod \"metallb-operator-webhook-server-755c486f67-b6mn4\" (UID: \"ee80d5f4-a13e-434d-b250-1e005d84dc59\") " pod="metallb-system/metallb-operator-webhook-server-755c486f67-b6mn4" Jan 26 17:02:02 crc kubenswrapper[4754]: I0126 17:02:02.604284 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ee80d5f4-a13e-434d-b250-1e005d84dc59-webhook-cert\") pod \"metallb-operator-webhook-server-755c486f67-b6mn4\" (UID: \"ee80d5f4-a13e-434d-b250-1e005d84dc59\") " pod="metallb-system/metallb-operator-webhook-server-755c486f67-b6mn4" Jan 26 17:02:02 crc kubenswrapper[4754]: I0126 17:02:02.705725 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/ee80d5f4-a13e-434d-b250-1e005d84dc59-apiservice-cert\") pod \"metallb-operator-webhook-server-755c486f67-b6mn4\" (UID: \"ee80d5f4-a13e-434d-b250-1e005d84dc59\") " pod="metallb-system/metallb-operator-webhook-server-755c486f67-b6mn4" Jan 26 17:02:02 crc kubenswrapper[4754]: I0126 17:02:02.705803 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nmltz\" (UniqueName: \"kubernetes.io/projected/ee80d5f4-a13e-434d-b250-1e005d84dc59-kube-api-access-nmltz\") pod \"metallb-operator-webhook-server-755c486f67-b6mn4\" (UID: \"ee80d5f4-a13e-434d-b250-1e005d84dc59\") " pod="metallb-system/metallb-operator-webhook-server-755c486f67-b6mn4" Jan 26 17:02:02 crc kubenswrapper[4754]: I0126 17:02:02.705873 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ee80d5f4-a13e-434d-b250-1e005d84dc59-webhook-cert\") pod \"metallb-operator-webhook-server-755c486f67-b6mn4\" (UID: \"ee80d5f4-a13e-434d-b250-1e005d84dc59\") " pod="metallb-system/metallb-operator-webhook-server-755c486f67-b6mn4" Jan 26 17:02:02 crc kubenswrapper[4754]: I0126 17:02:02.711571 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ee80d5f4-a13e-434d-b250-1e005d84dc59-webhook-cert\") pod \"metallb-operator-webhook-server-755c486f67-b6mn4\" (UID: \"ee80d5f4-a13e-434d-b250-1e005d84dc59\") " pod="metallb-system/metallb-operator-webhook-server-755c486f67-b6mn4" Jan 26 17:02:02 crc kubenswrapper[4754]: I0126 17:02:02.721634 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/ee80d5f4-a13e-434d-b250-1e005d84dc59-apiservice-cert\") pod \"metallb-operator-webhook-server-755c486f67-b6mn4\" (UID: \"ee80d5f4-a13e-434d-b250-1e005d84dc59\") " pod="metallb-system/metallb-operator-webhook-server-755c486f67-b6mn4" Jan 26 17:02:02 crc kubenswrapper[4754]: I0126 17:02:02.725331 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nmltz\" (UniqueName: \"kubernetes.io/projected/ee80d5f4-a13e-434d-b250-1e005d84dc59-kube-api-access-nmltz\") pod \"metallb-operator-webhook-server-755c486f67-b6mn4\" (UID: \"ee80d5f4-a13e-434d-b250-1e005d84dc59\") " pod="metallb-system/metallb-operator-webhook-server-755c486f67-b6mn4" Jan 26 17:02:02 crc kubenswrapper[4754]: I0126 17:02:02.852360 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-755c486f67-b6mn4" Jan 26 17:02:03 crc kubenswrapper[4754]: I0126 17:02:03.007289 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-755bc5d786-g2p48"] Jan 26 17:02:03 crc kubenswrapper[4754]: W0126 17:02:03.009193 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod83c99759_10ef_42eb_a58f_23c4e6c7d089.slice/crio-a2d21b4f8f09eac8e6ab5b83c86f713eb7a5f060cf7ba184b512fac4dd124d0b WatchSource:0}: Error finding container a2d21b4f8f09eac8e6ab5b83c86f713eb7a5f060cf7ba184b512fac4dd124d0b: Status 404 returned error can't find the container with id a2d21b4f8f09eac8e6ab5b83c86f713eb7a5f060cf7ba184b512fac4dd124d0b Jan 26 17:02:03 crc kubenswrapper[4754]: I0126 17:02:03.061887 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-755c486f67-b6mn4"] Jan 26 17:02:03 crc kubenswrapper[4754]: W0126 17:02:03.063950 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podee80d5f4_a13e_434d_b250_1e005d84dc59.slice/crio-11130c834354f4b77c5f06709f5fc12675a15db08002a8f635e2fd5da79beba3 WatchSource:0}: Error finding container 11130c834354f4b77c5f06709f5fc12675a15db08002a8f635e2fd5da79beba3: Status 404 returned error can't find the container with id 11130c834354f4b77c5f06709f5fc12675a15db08002a8f635e2fd5da79beba3 Jan 26 17:02:03 crc kubenswrapper[4754]: I0126 17:02:03.432146 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-755bc5d786-g2p48" event={"ID":"83c99759-10ef-42eb-a58f-23c4e6c7d089","Type":"ContainerStarted","Data":"a2d21b4f8f09eac8e6ab5b83c86f713eb7a5f060cf7ba184b512fac4dd124d0b"} Jan 26 17:02:03 crc kubenswrapper[4754]: I0126 17:02:03.433394 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-755c486f67-b6mn4" event={"ID":"ee80d5f4-a13e-434d-b250-1e005d84dc59","Type":"ContainerStarted","Data":"11130c834354f4b77c5f06709f5fc12675a15db08002a8f635e2fd5da79beba3"} Jan 26 17:02:07 crc kubenswrapper[4754]: I0126 17:02:07.461692 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-755bc5d786-g2p48" event={"ID":"83c99759-10ef-42eb-a58f-23c4e6c7d089","Type":"ContainerStarted","Data":"d3ee5a32e487b30a390c261466642ed4a1db135af0b4cc9b3c66c361a6029589"} Jan 26 17:02:07 crc kubenswrapper[4754]: I0126 17:02:07.462326 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-755bc5d786-g2p48" Jan 26 17:02:07 crc kubenswrapper[4754]: I0126 17:02:07.466147 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-755c486f67-b6mn4" event={"ID":"ee80d5f4-a13e-434d-b250-1e005d84dc59","Type":"ContainerStarted","Data":"d98844321aef32240c66e05ebe45d6b1a37e61c409e94f1fe7b4ec67c38dbea5"} Jan 26 17:02:07 crc kubenswrapper[4754]: I0126 17:02:07.466301 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-755c486f67-b6mn4" Jan 26 17:02:07 crc kubenswrapper[4754]: I0126 17:02:07.499113 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-755bc5d786-g2p48" podStartSLOduration=1.239189003 podStartE2EDuration="5.49909332s" podCreationTimestamp="2026-01-26 17:02:02 +0000 UTC" firstStartedPulling="2026-01-26 17:02:03.01188956 +0000 UTC m=+889.536069994" lastFinishedPulling="2026-01-26 17:02:07.271793877 +0000 UTC m=+893.795974311" observedRunningTime="2026-01-26 17:02:07.490235211 +0000 UTC m=+894.014415675" watchObservedRunningTime="2026-01-26 17:02:07.49909332 +0000 UTC m=+894.023273754" Jan 26 17:02:07 crc kubenswrapper[4754]: I0126 17:02:07.510318 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-755c486f67-b6mn4" podStartSLOduration=1.291692877 podStartE2EDuration="5.510299885s" podCreationTimestamp="2026-01-26 17:02:02 +0000 UTC" firstStartedPulling="2026-01-26 17:02:03.06683192 +0000 UTC m=+889.591012354" lastFinishedPulling="2026-01-26 17:02:07.285438928 +0000 UTC m=+893.809619362" observedRunningTime="2026-01-26 17:02:07.505788938 +0000 UTC m=+894.029969392" watchObservedRunningTime="2026-01-26 17:02:07.510299885 +0000 UTC m=+894.034480329" Jan 26 17:02:22 crc kubenswrapper[4754]: I0126 17:02:22.858057 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-755c486f67-b6mn4" Jan 26 17:02:37 crc kubenswrapper[4754]: I0126 17:02:37.130009 4754 patch_prober.go:28] interesting pod/machine-config-daemon-x65wv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 26 17:02:37 crc kubenswrapper[4754]: I0126 17:02:37.130848 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 26 17:02:42 crc kubenswrapper[4754]: I0126 17:02:42.551601 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-755bc5d786-g2p48" Jan 26 17:02:43 crc kubenswrapper[4754]: I0126 17:02:43.398389 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-5h7fb"] Jan 26 17:02:43 crc kubenswrapper[4754]: I0126 17:02:43.400616 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-5h7fb" Jan 26 17:02:43 crc kubenswrapper[4754]: I0126 17:02:43.406253 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-7df86c4f6c-zbwg9"] Jan 26 17:02:43 crc kubenswrapper[4754]: I0126 17:02:43.406911 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-zbwg9" Jan 26 17:02:43 crc kubenswrapper[4754]: I0126 17:02:43.408085 4754 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Jan 26 17:02:43 crc kubenswrapper[4754]: I0126 17:02:43.408175 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Jan 26 17:02:43 crc kubenswrapper[4754]: I0126 17:02:43.408973 4754 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Jan 26 17:02:43 crc kubenswrapper[4754]: I0126 17:02:43.409179 4754 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-2nhhn" Jan 26 17:02:43 crc kubenswrapper[4754]: I0126 17:02:43.420398 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7df86c4f6c-zbwg9"] Jan 26 17:02:43 crc kubenswrapper[4754]: I0126 17:02:43.480604 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-drvw7"] Jan 26 17:02:43 crc kubenswrapper[4754]: I0126 17:02:43.481712 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-drvw7" Jan 26 17:02:43 crc kubenswrapper[4754]: I0126 17:02:43.486917 4754 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Jan 26 17:02:43 crc kubenswrapper[4754]: I0126 17:02:43.487552 4754 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-9zzl7" Jan 26 17:02:43 crc kubenswrapper[4754]: I0126 17:02:43.487642 4754 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Jan 26 17:02:43 crc kubenswrapper[4754]: I0126 17:02:43.488221 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Jan 26 17:02:43 crc kubenswrapper[4754]: I0126 17:02:43.494708 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-6968d8fdc4-8458x"] Jan 26 17:02:43 crc kubenswrapper[4754]: I0126 17:02:43.495514 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-6968d8fdc4-8458x" Jan 26 17:02:43 crc kubenswrapper[4754]: I0126 17:02:43.497046 4754 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Jan 26 17:02:43 crc kubenswrapper[4754]: I0126 17:02:43.507776 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/9f0c3994-3669-4cb5-89b3-b9e9dbc316de-frr-sockets\") pod \"frr-k8s-5h7fb\" (UID: \"9f0c3994-3669-4cb5-89b3-b9e9dbc316de\") " pod="metallb-system/frr-k8s-5h7fb" Jan 26 17:02:43 crc kubenswrapper[4754]: I0126 17:02:43.507812 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xbmcm\" (UniqueName: \"kubernetes.io/projected/9f0c3994-3669-4cb5-89b3-b9e9dbc316de-kube-api-access-xbmcm\") pod \"frr-k8s-5h7fb\" (UID: \"9f0c3994-3669-4cb5-89b3-b9e9dbc316de\") " pod="metallb-system/frr-k8s-5h7fb" Jan 26 17:02:43 crc kubenswrapper[4754]: I0126 17:02:43.507841 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/9f0c3994-3669-4cb5-89b3-b9e9dbc316de-frr-startup\") pod \"frr-k8s-5h7fb\" (UID: \"9f0c3994-3669-4cb5-89b3-b9e9dbc316de\") " pod="metallb-system/frr-k8s-5h7fb" Jan 26 17:02:43 crc kubenswrapper[4754]: I0126 17:02:43.507861 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/9f0c3994-3669-4cb5-89b3-b9e9dbc316de-frr-conf\") pod \"frr-k8s-5h7fb\" (UID: \"9f0c3994-3669-4cb5-89b3-b9e9dbc316de\") " pod="metallb-system/frr-k8s-5h7fb" Jan 26 17:02:43 crc kubenswrapper[4754]: I0126 17:02:43.507886 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/9f0c3994-3669-4cb5-89b3-b9e9dbc316de-reloader\") pod \"frr-k8s-5h7fb\" (UID: \"9f0c3994-3669-4cb5-89b3-b9e9dbc316de\") " pod="metallb-system/frr-k8s-5h7fb" Jan 26 17:02:43 crc kubenswrapper[4754]: I0126 17:02:43.507929 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qpxwh\" (UniqueName: \"kubernetes.io/projected/941e5b29-8892-451b-9ba6-afdad3c1c77b-kube-api-access-qpxwh\") pod \"frr-k8s-webhook-server-7df86c4f6c-zbwg9\" (UID: \"941e5b29-8892-451b-9ba6-afdad3c1c77b\") " pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-zbwg9" Jan 26 17:02:43 crc kubenswrapper[4754]: I0126 17:02:43.507953 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/941e5b29-8892-451b-9ba6-afdad3c1c77b-cert\") pod \"frr-k8s-webhook-server-7df86c4f6c-zbwg9\" (UID: \"941e5b29-8892-451b-9ba6-afdad3c1c77b\") " pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-zbwg9" Jan 26 17:02:43 crc kubenswrapper[4754]: I0126 17:02:43.507980 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9f0c3994-3669-4cb5-89b3-b9e9dbc316de-metrics-certs\") pod \"frr-k8s-5h7fb\" (UID: \"9f0c3994-3669-4cb5-89b3-b9e9dbc316de\") " pod="metallb-system/frr-k8s-5h7fb" Jan 26 17:02:43 crc kubenswrapper[4754]: I0126 17:02:43.508003 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/9f0c3994-3669-4cb5-89b3-b9e9dbc316de-metrics\") pod \"frr-k8s-5h7fb\" (UID: \"9f0c3994-3669-4cb5-89b3-b9e9dbc316de\") " pod="metallb-system/frr-k8s-5h7fb" Jan 26 17:02:43 crc kubenswrapper[4754]: I0126 17:02:43.515497 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-6968d8fdc4-8458x"] Jan 26 17:02:43 crc kubenswrapper[4754]: I0126 17:02:43.609527 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/9f0c3994-3669-4cb5-89b3-b9e9dbc316de-frr-conf\") pod \"frr-k8s-5h7fb\" (UID: \"9f0c3994-3669-4cb5-89b3-b9e9dbc316de\") " pod="metallb-system/frr-k8s-5h7fb" Jan 26 17:02:43 crc kubenswrapper[4754]: I0126 17:02:43.609583 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/9f0c3994-3669-4cb5-89b3-b9e9dbc316de-reloader\") pod \"frr-k8s-5h7fb\" (UID: \"9f0c3994-3669-4cb5-89b3-b9e9dbc316de\") " pod="metallb-system/frr-k8s-5h7fb" Jan 26 17:02:43 crc kubenswrapper[4754]: I0126 17:02:43.609713 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/b3234c0c-afa9-41b4-88e3-70f44df6c2ce-memberlist\") pod \"speaker-drvw7\" (UID: \"b3234c0c-afa9-41b4-88e3-70f44df6c2ce\") " pod="metallb-system/speaker-drvw7" Jan 26 17:02:43 crc kubenswrapper[4754]: I0126 17:02:43.609752 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qpxwh\" (UniqueName: \"kubernetes.io/projected/941e5b29-8892-451b-9ba6-afdad3c1c77b-kube-api-access-qpxwh\") pod \"frr-k8s-webhook-server-7df86c4f6c-zbwg9\" (UID: \"941e5b29-8892-451b-9ba6-afdad3c1c77b\") " pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-zbwg9" Jan 26 17:02:43 crc kubenswrapper[4754]: I0126 17:02:43.609798 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/941e5b29-8892-451b-9ba6-afdad3c1c77b-cert\") pod \"frr-k8s-webhook-server-7df86c4f6c-zbwg9\" (UID: \"941e5b29-8892-451b-9ba6-afdad3c1c77b\") " pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-zbwg9" Jan 26 17:02:43 crc kubenswrapper[4754]: I0126 17:02:43.609823 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/dadf6410-9227-422a-8954-a2488091b15e-metrics-certs\") pod \"controller-6968d8fdc4-8458x\" (UID: \"dadf6410-9227-422a-8954-a2488091b15e\") " pod="metallb-system/controller-6968d8fdc4-8458x" Jan 26 17:02:43 crc kubenswrapper[4754]: I0126 17:02:43.610016 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9f0c3994-3669-4cb5-89b3-b9e9dbc316de-metrics-certs\") pod \"frr-k8s-5h7fb\" (UID: \"9f0c3994-3669-4cb5-89b3-b9e9dbc316de\") " pod="metallb-system/frr-k8s-5h7fb" Jan 26 17:02:43 crc kubenswrapper[4754]: I0126 17:02:43.610096 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/b3234c0c-afa9-41b4-88e3-70f44df6c2ce-metallb-excludel2\") pod \"speaker-drvw7\" (UID: \"b3234c0c-afa9-41b4-88e3-70f44df6c2ce\") " pod="metallb-system/speaker-drvw7" Jan 26 17:02:43 crc kubenswrapper[4754]: I0126 17:02:43.610140 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/9f0c3994-3669-4cb5-89b3-b9e9dbc316de-reloader\") pod \"frr-k8s-5h7fb\" (UID: \"9f0c3994-3669-4cb5-89b3-b9e9dbc316de\") " pod="metallb-system/frr-k8s-5h7fb" Jan 26 17:02:43 crc kubenswrapper[4754]: I0126 17:02:43.610163 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/9f0c3994-3669-4cb5-89b3-b9e9dbc316de-metrics\") pod \"frr-k8s-5h7fb\" (UID: \"9f0c3994-3669-4cb5-89b3-b9e9dbc316de\") " pod="metallb-system/frr-k8s-5h7fb" Jan 26 17:02:43 crc kubenswrapper[4754]: I0126 17:02:43.610179 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/9f0c3994-3669-4cb5-89b3-b9e9dbc316de-frr-conf\") pod \"frr-k8s-5h7fb\" (UID: \"9f0c3994-3669-4cb5-89b3-b9e9dbc316de\") " pod="metallb-system/frr-k8s-5h7fb" Jan 26 17:02:43 crc kubenswrapper[4754]: I0126 17:02:43.610261 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/9f0c3994-3669-4cb5-89b3-b9e9dbc316de-frr-sockets\") pod \"frr-k8s-5h7fb\" (UID: \"9f0c3994-3669-4cb5-89b3-b9e9dbc316de\") " pod="metallb-system/frr-k8s-5h7fb" Jan 26 17:02:43 crc kubenswrapper[4754]: I0126 17:02:43.610292 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xbmcm\" (UniqueName: \"kubernetes.io/projected/9f0c3994-3669-4cb5-89b3-b9e9dbc316de-kube-api-access-xbmcm\") pod \"frr-k8s-5h7fb\" (UID: \"9f0c3994-3669-4cb5-89b3-b9e9dbc316de\") " pod="metallb-system/frr-k8s-5h7fb" Jan 26 17:02:43 crc kubenswrapper[4754]: I0126 17:02:43.610326 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b3234c0c-afa9-41b4-88e3-70f44df6c2ce-metrics-certs\") pod \"speaker-drvw7\" (UID: \"b3234c0c-afa9-41b4-88e3-70f44df6c2ce\") " pod="metallb-system/speaker-drvw7" Jan 26 17:02:43 crc kubenswrapper[4754]: I0126 17:02:43.610362 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k5gk2\" (UniqueName: \"kubernetes.io/projected/dadf6410-9227-422a-8954-a2488091b15e-kube-api-access-k5gk2\") pod \"controller-6968d8fdc4-8458x\" (UID: \"dadf6410-9227-422a-8954-a2488091b15e\") " pod="metallb-system/controller-6968d8fdc4-8458x" Jan 26 17:02:43 crc kubenswrapper[4754]: I0126 17:02:43.610387 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/9f0c3994-3669-4cb5-89b3-b9e9dbc316de-metrics\") pod \"frr-k8s-5h7fb\" (UID: \"9f0c3994-3669-4cb5-89b3-b9e9dbc316de\") " pod="metallb-system/frr-k8s-5h7fb" Jan 26 17:02:43 crc kubenswrapper[4754]: I0126 17:02:43.610389 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/dadf6410-9227-422a-8954-a2488091b15e-cert\") pod \"controller-6968d8fdc4-8458x\" (UID: \"dadf6410-9227-422a-8954-a2488091b15e\") " pod="metallb-system/controller-6968d8fdc4-8458x" Jan 26 17:02:43 crc kubenswrapper[4754]: I0126 17:02:43.610434 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ccgwr\" (UniqueName: \"kubernetes.io/projected/b3234c0c-afa9-41b4-88e3-70f44df6c2ce-kube-api-access-ccgwr\") pod \"speaker-drvw7\" (UID: \"b3234c0c-afa9-41b4-88e3-70f44df6c2ce\") " pod="metallb-system/speaker-drvw7" Jan 26 17:02:43 crc kubenswrapper[4754]: I0126 17:02:43.610461 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/9f0c3994-3669-4cb5-89b3-b9e9dbc316de-frr-startup\") pod \"frr-k8s-5h7fb\" (UID: \"9f0c3994-3669-4cb5-89b3-b9e9dbc316de\") " pod="metallb-system/frr-k8s-5h7fb" Jan 26 17:02:43 crc kubenswrapper[4754]: I0126 17:02:43.610473 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/9f0c3994-3669-4cb5-89b3-b9e9dbc316de-frr-sockets\") pod \"frr-k8s-5h7fb\" (UID: \"9f0c3994-3669-4cb5-89b3-b9e9dbc316de\") " pod="metallb-system/frr-k8s-5h7fb" Jan 26 17:02:43 crc kubenswrapper[4754]: I0126 17:02:43.611336 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/9f0c3994-3669-4cb5-89b3-b9e9dbc316de-frr-startup\") pod \"frr-k8s-5h7fb\" (UID: \"9f0c3994-3669-4cb5-89b3-b9e9dbc316de\") " pod="metallb-system/frr-k8s-5h7fb" Jan 26 17:02:43 crc kubenswrapper[4754]: I0126 17:02:43.620326 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9f0c3994-3669-4cb5-89b3-b9e9dbc316de-metrics-certs\") pod \"frr-k8s-5h7fb\" (UID: \"9f0c3994-3669-4cb5-89b3-b9e9dbc316de\") " pod="metallb-system/frr-k8s-5h7fb" Jan 26 17:02:43 crc kubenswrapper[4754]: I0126 17:02:43.640226 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/941e5b29-8892-451b-9ba6-afdad3c1c77b-cert\") pod \"frr-k8s-webhook-server-7df86c4f6c-zbwg9\" (UID: \"941e5b29-8892-451b-9ba6-afdad3c1c77b\") " pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-zbwg9" Jan 26 17:02:43 crc kubenswrapper[4754]: I0126 17:02:43.662319 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xbmcm\" (UniqueName: \"kubernetes.io/projected/9f0c3994-3669-4cb5-89b3-b9e9dbc316de-kube-api-access-xbmcm\") pod \"frr-k8s-5h7fb\" (UID: \"9f0c3994-3669-4cb5-89b3-b9e9dbc316de\") " pod="metallb-system/frr-k8s-5h7fb" Jan 26 17:02:43 crc kubenswrapper[4754]: I0126 17:02:43.662690 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qpxwh\" (UniqueName: \"kubernetes.io/projected/941e5b29-8892-451b-9ba6-afdad3c1c77b-kube-api-access-qpxwh\") pod \"frr-k8s-webhook-server-7df86c4f6c-zbwg9\" (UID: \"941e5b29-8892-451b-9ba6-afdad3c1c77b\") " pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-zbwg9" Jan 26 17:02:43 crc kubenswrapper[4754]: I0126 17:02:43.711373 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/dadf6410-9227-422a-8954-a2488091b15e-cert\") pod \"controller-6968d8fdc4-8458x\" (UID: \"dadf6410-9227-422a-8954-a2488091b15e\") " pod="metallb-system/controller-6968d8fdc4-8458x" Jan 26 17:02:43 crc kubenswrapper[4754]: I0126 17:02:43.711435 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ccgwr\" (UniqueName: \"kubernetes.io/projected/b3234c0c-afa9-41b4-88e3-70f44df6c2ce-kube-api-access-ccgwr\") pod \"speaker-drvw7\" (UID: \"b3234c0c-afa9-41b4-88e3-70f44df6c2ce\") " pod="metallb-system/speaker-drvw7" Jan 26 17:02:43 crc kubenswrapper[4754]: I0126 17:02:43.711484 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/b3234c0c-afa9-41b4-88e3-70f44df6c2ce-memberlist\") pod \"speaker-drvw7\" (UID: \"b3234c0c-afa9-41b4-88e3-70f44df6c2ce\") " pod="metallb-system/speaker-drvw7" Jan 26 17:02:43 crc kubenswrapper[4754]: I0126 17:02:43.711532 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/dadf6410-9227-422a-8954-a2488091b15e-metrics-certs\") pod \"controller-6968d8fdc4-8458x\" (UID: \"dadf6410-9227-422a-8954-a2488091b15e\") " pod="metallb-system/controller-6968d8fdc4-8458x" Jan 26 17:02:43 crc kubenswrapper[4754]: I0126 17:02:43.711559 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/b3234c0c-afa9-41b4-88e3-70f44df6c2ce-metallb-excludel2\") pod \"speaker-drvw7\" (UID: \"b3234c0c-afa9-41b4-88e3-70f44df6c2ce\") " pod="metallb-system/speaker-drvw7" Jan 26 17:02:43 crc kubenswrapper[4754]: I0126 17:02:43.711601 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b3234c0c-afa9-41b4-88e3-70f44df6c2ce-metrics-certs\") pod \"speaker-drvw7\" (UID: \"b3234c0c-afa9-41b4-88e3-70f44df6c2ce\") " pod="metallb-system/speaker-drvw7" Jan 26 17:02:43 crc kubenswrapper[4754]: I0126 17:02:43.711625 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k5gk2\" (UniqueName: \"kubernetes.io/projected/dadf6410-9227-422a-8954-a2488091b15e-kube-api-access-k5gk2\") pod \"controller-6968d8fdc4-8458x\" (UID: \"dadf6410-9227-422a-8954-a2488091b15e\") " pod="metallb-system/controller-6968d8fdc4-8458x" Jan 26 17:02:43 crc kubenswrapper[4754]: E0126 17:02:43.711766 4754 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Jan 26 17:02:43 crc kubenswrapper[4754]: E0126 17:02:43.711859 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b3234c0c-afa9-41b4-88e3-70f44df6c2ce-memberlist podName:b3234c0c-afa9-41b4-88e3-70f44df6c2ce nodeName:}" failed. No retries permitted until 2026-01-26 17:02:44.211836333 +0000 UTC m=+930.736016767 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/b3234c0c-afa9-41b4-88e3-70f44df6c2ce-memberlist") pod "speaker-drvw7" (UID: "b3234c0c-afa9-41b4-88e3-70f44df6c2ce") : secret "metallb-memberlist" not found Jan 26 17:02:43 crc kubenswrapper[4754]: I0126 17:02:43.712646 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/b3234c0c-afa9-41b4-88e3-70f44df6c2ce-metallb-excludel2\") pod \"speaker-drvw7\" (UID: \"b3234c0c-afa9-41b4-88e3-70f44df6c2ce\") " pod="metallb-system/speaker-drvw7" Jan 26 17:02:43 crc kubenswrapper[4754]: I0126 17:02:43.713335 4754 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Jan 26 17:02:43 crc kubenswrapper[4754]: I0126 17:02:43.715232 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b3234c0c-afa9-41b4-88e3-70f44df6c2ce-metrics-certs\") pod \"speaker-drvw7\" (UID: \"b3234c0c-afa9-41b4-88e3-70f44df6c2ce\") " pod="metallb-system/speaker-drvw7" Jan 26 17:02:43 crc kubenswrapper[4754]: I0126 17:02:43.715697 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/dadf6410-9227-422a-8954-a2488091b15e-metrics-certs\") pod \"controller-6968d8fdc4-8458x\" (UID: \"dadf6410-9227-422a-8954-a2488091b15e\") " pod="metallb-system/controller-6968d8fdc4-8458x" Jan 26 17:02:43 crc kubenswrapper[4754]: I0126 17:02:43.724415 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-5h7fb" Jan 26 17:02:43 crc kubenswrapper[4754]: I0126 17:02:43.727617 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/dadf6410-9227-422a-8954-a2488091b15e-cert\") pod \"controller-6968d8fdc4-8458x\" (UID: \"dadf6410-9227-422a-8954-a2488091b15e\") " pod="metallb-system/controller-6968d8fdc4-8458x" Jan 26 17:02:43 crc kubenswrapper[4754]: I0126 17:02:43.728877 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ccgwr\" (UniqueName: \"kubernetes.io/projected/b3234c0c-afa9-41b4-88e3-70f44df6c2ce-kube-api-access-ccgwr\") pod \"speaker-drvw7\" (UID: \"b3234c0c-afa9-41b4-88e3-70f44df6c2ce\") " pod="metallb-system/speaker-drvw7" Jan 26 17:02:43 crc kubenswrapper[4754]: I0126 17:02:43.729025 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k5gk2\" (UniqueName: \"kubernetes.io/projected/dadf6410-9227-422a-8954-a2488091b15e-kube-api-access-k5gk2\") pod \"controller-6968d8fdc4-8458x\" (UID: \"dadf6410-9227-422a-8954-a2488091b15e\") " pod="metallb-system/controller-6968d8fdc4-8458x" Jan 26 17:02:43 crc kubenswrapper[4754]: I0126 17:02:43.731836 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-zbwg9" Jan 26 17:02:43 crc kubenswrapper[4754]: I0126 17:02:43.809660 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-6968d8fdc4-8458x" Jan 26 17:02:44 crc kubenswrapper[4754]: I0126 17:02:44.007829 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-6968d8fdc4-8458x"] Jan 26 17:02:44 crc kubenswrapper[4754]: W0126 17:02:44.010216 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddadf6410_9227_422a_8954_a2488091b15e.slice/crio-6091e8bf62c26c647c7670055b560b7c807d07aaf9ea7facc0af414d71e7e67f WatchSource:0}: Error finding container 6091e8bf62c26c647c7670055b560b7c807d07aaf9ea7facc0af414d71e7e67f: Status 404 returned error can't find the container with id 6091e8bf62c26c647c7670055b560b7c807d07aaf9ea7facc0af414d71e7e67f Jan 26 17:02:44 crc kubenswrapper[4754]: I0126 17:02:44.133484 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7df86c4f6c-zbwg9"] Jan 26 17:02:44 crc kubenswrapper[4754]: W0126 17:02:44.139880 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod941e5b29_8892_451b_9ba6_afdad3c1c77b.slice/crio-d8e232c544ddd34909d019f1b2b7aeb416559858e0ecf0220e5ea65d1bfc1a12 WatchSource:0}: Error finding container d8e232c544ddd34909d019f1b2b7aeb416559858e0ecf0220e5ea65d1bfc1a12: Status 404 returned error can't find the container with id d8e232c544ddd34909d019f1b2b7aeb416559858e0ecf0220e5ea65d1bfc1a12 Jan 26 17:02:44 crc kubenswrapper[4754]: I0126 17:02:44.217336 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/b3234c0c-afa9-41b4-88e3-70f44df6c2ce-memberlist\") pod \"speaker-drvw7\" (UID: \"b3234c0c-afa9-41b4-88e3-70f44df6c2ce\") " pod="metallb-system/speaker-drvw7" Jan 26 17:02:44 crc kubenswrapper[4754]: E0126 17:02:44.217520 4754 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Jan 26 17:02:44 crc kubenswrapper[4754]: E0126 17:02:44.217744 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b3234c0c-afa9-41b4-88e3-70f44df6c2ce-memberlist podName:b3234c0c-afa9-41b4-88e3-70f44df6c2ce nodeName:}" failed. No retries permitted until 2026-01-26 17:02:45.217725049 +0000 UTC m=+931.741905493 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/b3234c0c-afa9-41b4-88e3-70f44df6c2ce-memberlist") pod "speaker-drvw7" (UID: "b3234c0c-afa9-41b4-88e3-70f44df6c2ce") : secret "metallb-memberlist" not found Jan 26 17:02:44 crc kubenswrapper[4754]: I0126 17:02:44.850416 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-6968d8fdc4-8458x" event={"ID":"dadf6410-9227-422a-8954-a2488091b15e","Type":"ContainerStarted","Data":"e5cb04fcdabc29a3d8e43a11e5c37a92031011767e856b91962f85a8b77c07b6"} Jan 26 17:02:44 crc kubenswrapper[4754]: I0126 17:02:44.850488 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-6968d8fdc4-8458x" event={"ID":"dadf6410-9227-422a-8954-a2488091b15e","Type":"ContainerStarted","Data":"6091e8bf62c26c647c7670055b560b7c807d07aaf9ea7facc0af414d71e7e67f"} Jan 26 17:02:44 crc kubenswrapper[4754]: I0126 17:02:44.851209 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-zbwg9" event={"ID":"941e5b29-8892-451b-9ba6-afdad3c1c77b","Type":"ContainerStarted","Data":"d8e232c544ddd34909d019f1b2b7aeb416559858e0ecf0220e5ea65d1bfc1a12"} Jan 26 17:02:44 crc kubenswrapper[4754]: I0126 17:02:44.852555 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-5h7fb" event={"ID":"9f0c3994-3669-4cb5-89b3-b9e9dbc316de","Type":"ContainerStarted","Data":"76dcfaa6ec23715f1bcce0d752cf060509e76bcded71dfe56b0d2d545bf523e7"} Jan 26 17:02:45 crc kubenswrapper[4754]: I0126 17:02:45.228916 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/b3234c0c-afa9-41b4-88e3-70f44df6c2ce-memberlist\") pod \"speaker-drvw7\" (UID: \"b3234c0c-afa9-41b4-88e3-70f44df6c2ce\") " pod="metallb-system/speaker-drvw7" Jan 26 17:02:45 crc kubenswrapper[4754]: I0126 17:02:45.234444 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/b3234c0c-afa9-41b4-88e3-70f44df6c2ce-memberlist\") pod \"speaker-drvw7\" (UID: \"b3234c0c-afa9-41b4-88e3-70f44df6c2ce\") " pod="metallb-system/speaker-drvw7" Jan 26 17:02:45 crc kubenswrapper[4754]: I0126 17:02:45.295629 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-drvw7" Jan 26 17:02:45 crc kubenswrapper[4754]: W0126 17:02:45.318341 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb3234c0c_afa9_41b4_88e3_70f44df6c2ce.slice/crio-02a973d166e2831c46cfa34adcff38d05ecb57632227d9a847c4458e7139c5bd WatchSource:0}: Error finding container 02a973d166e2831c46cfa34adcff38d05ecb57632227d9a847c4458e7139c5bd: Status 404 returned error can't find the container with id 02a973d166e2831c46cfa34adcff38d05ecb57632227d9a847c4458e7139c5bd Jan 26 17:02:45 crc kubenswrapper[4754]: I0126 17:02:45.870530 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-6968d8fdc4-8458x" event={"ID":"dadf6410-9227-422a-8954-a2488091b15e","Type":"ContainerStarted","Data":"7e10e056e348e532652525c7bb42cefab29a1ba1e498d0041540deeec5f1cc66"} Jan 26 17:02:45 crc kubenswrapper[4754]: I0126 17:02:45.871157 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-6968d8fdc4-8458x" Jan 26 17:02:45 crc kubenswrapper[4754]: I0126 17:02:45.877738 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-drvw7" event={"ID":"b3234c0c-afa9-41b4-88e3-70f44df6c2ce","Type":"ContainerStarted","Data":"dacf4699515921175fa507d93e05d90f0329ae90156151bee613c6ba8ec80adb"} Jan 26 17:02:45 crc kubenswrapper[4754]: I0126 17:02:45.877795 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-drvw7" event={"ID":"b3234c0c-afa9-41b4-88e3-70f44df6c2ce","Type":"ContainerStarted","Data":"acad9ca0eddbdd8b892d2106116849fa694b8c7d04c01f7b5174404ac1f0ae11"} Jan 26 17:02:45 crc kubenswrapper[4754]: I0126 17:02:45.877817 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-drvw7" event={"ID":"b3234c0c-afa9-41b4-88e3-70f44df6c2ce","Type":"ContainerStarted","Data":"02a973d166e2831c46cfa34adcff38d05ecb57632227d9a847c4458e7139c5bd"} Jan 26 17:02:45 crc kubenswrapper[4754]: I0126 17:02:45.878776 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-drvw7" Jan 26 17:02:45 crc kubenswrapper[4754]: I0126 17:02:45.897971 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-6968d8fdc4-8458x" podStartSLOduration=2.897952046 podStartE2EDuration="2.897952046s" podCreationTimestamp="2026-01-26 17:02:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 17:02:45.895056995 +0000 UTC m=+932.419237449" watchObservedRunningTime="2026-01-26 17:02:45.897952046 +0000 UTC m=+932.422132500" Jan 26 17:02:45 crc kubenswrapper[4754]: I0126 17:02:45.915054 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-drvw7" podStartSLOduration=2.915034415 podStartE2EDuration="2.915034415s" podCreationTimestamp="2026-01-26 17:02:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 17:02:45.912812993 +0000 UTC m=+932.436993447" watchObservedRunningTime="2026-01-26 17:02:45.915034415 +0000 UTC m=+932.439214859" Jan 26 17:02:51 crc kubenswrapper[4754]: I0126 17:02:51.537792 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-qh8sw"] Jan 26 17:02:51 crc kubenswrapper[4754]: I0126 17:02:51.542746 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qh8sw" Jan 26 17:02:51 crc kubenswrapper[4754]: I0126 17:02:51.555528 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-qh8sw"] Jan 26 17:02:51 crc kubenswrapper[4754]: I0126 17:02:51.644647 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6666899f-4615-424e-a0c9-80bc260eef85-utilities\") pod \"redhat-marketplace-qh8sw\" (UID: \"6666899f-4615-424e-a0c9-80bc260eef85\") " pod="openshift-marketplace/redhat-marketplace-qh8sw" Jan 26 17:02:51 crc kubenswrapper[4754]: I0126 17:02:51.644730 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6666899f-4615-424e-a0c9-80bc260eef85-catalog-content\") pod \"redhat-marketplace-qh8sw\" (UID: \"6666899f-4615-424e-a0c9-80bc260eef85\") " pod="openshift-marketplace/redhat-marketplace-qh8sw" Jan 26 17:02:51 crc kubenswrapper[4754]: I0126 17:02:51.644795 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sgmxz\" (UniqueName: \"kubernetes.io/projected/6666899f-4615-424e-a0c9-80bc260eef85-kube-api-access-sgmxz\") pod \"redhat-marketplace-qh8sw\" (UID: \"6666899f-4615-424e-a0c9-80bc260eef85\") " pod="openshift-marketplace/redhat-marketplace-qh8sw" Jan 26 17:02:51 crc kubenswrapper[4754]: I0126 17:02:51.745863 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6666899f-4615-424e-a0c9-80bc260eef85-catalog-content\") pod \"redhat-marketplace-qh8sw\" (UID: \"6666899f-4615-424e-a0c9-80bc260eef85\") " pod="openshift-marketplace/redhat-marketplace-qh8sw" Jan 26 17:02:51 crc kubenswrapper[4754]: I0126 17:02:51.745975 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sgmxz\" (UniqueName: \"kubernetes.io/projected/6666899f-4615-424e-a0c9-80bc260eef85-kube-api-access-sgmxz\") pod \"redhat-marketplace-qh8sw\" (UID: \"6666899f-4615-424e-a0c9-80bc260eef85\") " pod="openshift-marketplace/redhat-marketplace-qh8sw" Jan 26 17:02:51 crc kubenswrapper[4754]: I0126 17:02:51.746042 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6666899f-4615-424e-a0c9-80bc260eef85-utilities\") pod \"redhat-marketplace-qh8sw\" (UID: \"6666899f-4615-424e-a0c9-80bc260eef85\") " pod="openshift-marketplace/redhat-marketplace-qh8sw" Jan 26 17:02:51 crc kubenswrapper[4754]: I0126 17:02:51.746510 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6666899f-4615-424e-a0c9-80bc260eef85-catalog-content\") pod \"redhat-marketplace-qh8sw\" (UID: \"6666899f-4615-424e-a0c9-80bc260eef85\") " pod="openshift-marketplace/redhat-marketplace-qh8sw" Jan 26 17:02:51 crc kubenswrapper[4754]: I0126 17:02:51.746546 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6666899f-4615-424e-a0c9-80bc260eef85-utilities\") pod \"redhat-marketplace-qh8sw\" (UID: \"6666899f-4615-424e-a0c9-80bc260eef85\") " pod="openshift-marketplace/redhat-marketplace-qh8sw" Jan 26 17:02:51 crc kubenswrapper[4754]: I0126 17:02:51.768446 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sgmxz\" (UniqueName: \"kubernetes.io/projected/6666899f-4615-424e-a0c9-80bc260eef85-kube-api-access-sgmxz\") pod \"redhat-marketplace-qh8sw\" (UID: \"6666899f-4615-424e-a0c9-80bc260eef85\") " pod="openshift-marketplace/redhat-marketplace-qh8sw" Jan 26 17:02:51 crc kubenswrapper[4754]: I0126 17:02:51.863905 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qh8sw" Jan 26 17:02:51 crc kubenswrapper[4754]: I0126 17:02:51.927254 4754 generic.go:334] "Generic (PLEG): container finished" podID="9f0c3994-3669-4cb5-89b3-b9e9dbc316de" containerID="ef0c69c38d9d09e5c8bc55603717bfe17ac79d4fd32fd9dee9e6127a886df8a0" exitCode=0 Jan 26 17:02:51 crc kubenswrapper[4754]: I0126 17:02:51.927536 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-5h7fb" event={"ID":"9f0c3994-3669-4cb5-89b3-b9e9dbc316de","Type":"ContainerDied","Data":"ef0c69c38d9d09e5c8bc55603717bfe17ac79d4fd32fd9dee9e6127a886df8a0"} Jan 26 17:02:51 crc kubenswrapper[4754]: I0126 17:02:51.939585 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-zbwg9" event={"ID":"941e5b29-8892-451b-9ba6-afdad3c1c77b","Type":"ContainerStarted","Data":"61652250aba00d46e4f82b91697fa98dcd7db17b99a23f3316f5e0fd86094e5c"} Jan 26 17:02:51 crc kubenswrapper[4754]: I0126 17:02:51.940143 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-zbwg9" Jan 26 17:02:51 crc kubenswrapper[4754]: I0126 17:02:51.998560 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-zbwg9" podStartSLOduration=1.4638110690000001 podStartE2EDuration="8.998539987s" podCreationTimestamp="2026-01-26 17:02:43 +0000 UTC" firstStartedPulling="2026-01-26 17:02:44.142010086 +0000 UTC m=+930.666190540" lastFinishedPulling="2026-01-26 17:02:51.676739024 +0000 UTC m=+938.200919458" observedRunningTime="2026-01-26 17:02:51.99506172 +0000 UTC m=+938.519242164" watchObservedRunningTime="2026-01-26 17:02:51.998539987 +0000 UTC m=+938.522720421" Jan 26 17:02:52 crc kubenswrapper[4754]: I0126 17:02:52.339948 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-qh8sw"] Jan 26 17:02:52 crc kubenswrapper[4754]: W0126 17:02:52.347209 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6666899f_4615_424e_a0c9_80bc260eef85.slice/crio-f2c7a7c3e877b50aa6a3342098b754bb82fbcae455a432c1500a9ba539a0abf4 WatchSource:0}: Error finding container f2c7a7c3e877b50aa6a3342098b754bb82fbcae455a432c1500a9ba539a0abf4: Status 404 returned error can't find the container with id f2c7a7c3e877b50aa6a3342098b754bb82fbcae455a432c1500a9ba539a0abf4 Jan 26 17:02:52 crc kubenswrapper[4754]: I0126 17:02:52.949262 4754 generic.go:334] "Generic (PLEG): container finished" podID="6666899f-4615-424e-a0c9-80bc260eef85" containerID="6ce40029bc2f482d3f8847d8deec9eb922afb7272ab37ed70c016b805d748912" exitCode=0 Jan 26 17:02:52 crc kubenswrapper[4754]: I0126 17:02:52.949423 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qh8sw" event={"ID":"6666899f-4615-424e-a0c9-80bc260eef85","Type":"ContainerDied","Data":"6ce40029bc2f482d3f8847d8deec9eb922afb7272ab37ed70c016b805d748912"} Jan 26 17:02:52 crc kubenswrapper[4754]: I0126 17:02:52.949471 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qh8sw" event={"ID":"6666899f-4615-424e-a0c9-80bc260eef85","Type":"ContainerStarted","Data":"f2c7a7c3e877b50aa6a3342098b754bb82fbcae455a432c1500a9ba539a0abf4"} Jan 26 17:02:52 crc kubenswrapper[4754]: I0126 17:02:52.952313 4754 generic.go:334] "Generic (PLEG): container finished" podID="9f0c3994-3669-4cb5-89b3-b9e9dbc316de" containerID="138d28855f4d9607b3153701683d8f540148b0818a99a90a0691ce621abaf303" exitCode=0 Jan 26 17:02:52 crc kubenswrapper[4754]: I0126 17:02:52.952404 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-5h7fb" event={"ID":"9f0c3994-3669-4cb5-89b3-b9e9dbc316de","Type":"ContainerDied","Data":"138d28855f4d9607b3153701683d8f540148b0818a99a90a0691ce621abaf303"} Jan 26 17:02:53 crc kubenswrapper[4754]: I0126 17:02:53.960187 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-5h7fb" event={"ID":"9f0c3994-3669-4cb5-89b3-b9e9dbc316de","Type":"ContainerDied","Data":"cac3b4f874b3d103e28054ff7d95633d7e44f8b59d9a11357ccdf70bcba4f493"} Jan 26 17:02:53 crc kubenswrapper[4754]: I0126 17:02:53.960125 4754 generic.go:334] "Generic (PLEG): container finished" podID="9f0c3994-3669-4cb5-89b3-b9e9dbc316de" containerID="cac3b4f874b3d103e28054ff7d95633d7e44f8b59d9a11357ccdf70bcba4f493" exitCode=0 Jan 26 17:02:54 crc kubenswrapper[4754]: I0126 17:02:54.969398 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-5h7fb" event={"ID":"9f0c3994-3669-4cb5-89b3-b9e9dbc316de","Type":"ContainerStarted","Data":"410b17a7b5bd13f27757e67cf91b77c1a994e77ae2ab9d8ffd7f38dfe74235c1"} Jan 26 17:02:54 crc kubenswrapper[4754]: I0126 17:02:54.969705 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-5h7fb" event={"ID":"9f0c3994-3669-4cb5-89b3-b9e9dbc316de","Type":"ContainerStarted","Data":"f6a54dc3eb35d6f3ac28c70eeef65e172fb4c398dc928fa15be418cd7ed0bed6"} Jan 26 17:02:54 crc kubenswrapper[4754]: I0126 17:02:54.969715 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-5h7fb" event={"ID":"9f0c3994-3669-4cb5-89b3-b9e9dbc316de","Type":"ContainerStarted","Data":"604bae8d9cfefa8875512bce2dc76a0ca54a33dddb7ab543fee929e5af0eedf5"} Jan 26 17:02:54 crc kubenswrapper[4754]: I0126 17:02:54.969723 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-5h7fb" event={"ID":"9f0c3994-3669-4cb5-89b3-b9e9dbc316de","Type":"ContainerStarted","Data":"221665c89910658c1d49c6830412449d3d2d0c1bc84a02c1611083abdbefe91b"} Jan 26 17:02:54 crc kubenswrapper[4754]: I0126 17:02:54.969732 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-5h7fb" event={"ID":"9f0c3994-3669-4cb5-89b3-b9e9dbc316de","Type":"ContainerStarted","Data":"ffbca3f9108b80cf66b619985fac125cd233297b5f0b568fbc8e378592870482"} Jan 26 17:02:55 crc kubenswrapper[4754]: I0126 17:02:55.301639 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-drvw7" Jan 26 17:02:55 crc kubenswrapper[4754]: I0126 17:02:55.978957 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-5h7fb" event={"ID":"9f0c3994-3669-4cb5-89b3-b9e9dbc316de","Type":"ContainerStarted","Data":"de3df21bad4ee675537730d314c19ea19272efa4f4b7b613551bd3309cba945a"} Jan 26 17:02:55 crc kubenswrapper[4754]: I0126 17:02:55.979486 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-5h7fb" Jan 26 17:02:56 crc kubenswrapper[4754]: I0126 17:02:56.006421 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-5h7fb" podStartSLOduration=6.049761968 podStartE2EDuration="13.006404005s" podCreationTimestamp="2026-01-26 17:02:43 +0000 UTC" firstStartedPulling="2026-01-26 17:02:44.732558046 +0000 UTC m=+931.256738500" lastFinishedPulling="2026-01-26 17:02:51.689200103 +0000 UTC m=+938.213380537" observedRunningTime="2026-01-26 17:02:56.002525777 +0000 UTC m=+942.526706211" watchObservedRunningTime="2026-01-26 17:02:56.006404005 +0000 UTC m=+942.530584439" Jan 26 17:02:58 crc kubenswrapper[4754]: I0126 17:02:58.006985 4754 generic.go:334] "Generic (PLEG): container finished" podID="6666899f-4615-424e-a0c9-80bc260eef85" containerID="9c69138b331561fc2ce032914e28cffc9f28c66a8dcadd05ada91c5c0a8553e4" exitCode=0 Jan 26 17:02:58 crc kubenswrapper[4754]: I0126 17:02:58.007037 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qh8sw" event={"ID":"6666899f-4615-424e-a0c9-80bc260eef85","Type":"ContainerDied","Data":"9c69138b331561fc2ce032914e28cffc9f28c66a8dcadd05ada91c5c0a8553e4"} Jan 26 17:02:58 crc kubenswrapper[4754]: I0126 17:02:58.588887 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-k7g8c"] Jan 26 17:02:58 crc kubenswrapper[4754]: I0126 17:02:58.595808 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-k7g8c" Jan 26 17:02:58 crc kubenswrapper[4754]: I0126 17:02:58.598274 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Jan 26 17:02:58 crc kubenswrapper[4754]: I0126 17:02:58.598625 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Jan 26 17:02:58 crc kubenswrapper[4754]: I0126 17:02:58.598833 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-hrqpx" Jan 26 17:02:58 crc kubenswrapper[4754]: I0126 17:02:58.617116 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-k7g8c"] Jan 26 17:02:58 crc kubenswrapper[4754]: I0126 17:02:58.639193 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sxf26\" (UniqueName: \"kubernetes.io/projected/31e2139e-589c-4d52-8a92-5835f038d3b1-kube-api-access-sxf26\") pod \"openstack-operator-index-k7g8c\" (UID: \"31e2139e-589c-4d52-8a92-5835f038d3b1\") " pod="openstack-operators/openstack-operator-index-k7g8c" Jan 26 17:02:58 crc kubenswrapper[4754]: I0126 17:02:58.726240 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-5h7fb" Jan 26 17:02:58 crc kubenswrapper[4754]: I0126 17:02:58.740518 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sxf26\" (UniqueName: \"kubernetes.io/projected/31e2139e-589c-4d52-8a92-5835f038d3b1-kube-api-access-sxf26\") pod \"openstack-operator-index-k7g8c\" (UID: \"31e2139e-589c-4d52-8a92-5835f038d3b1\") " pod="openstack-operators/openstack-operator-index-k7g8c" Jan 26 17:02:58 crc kubenswrapper[4754]: I0126 17:02:58.758812 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sxf26\" (UniqueName: \"kubernetes.io/projected/31e2139e-589c-4d52-8a92-5835f038d3b1-kube-api-access-sxf26\") pod \"openstack-operator-index-k7g8c\" (UID: \"31e2139e-589c-4d52-8a92-5835f038d3b1\") " pod="openstack-operators/openstack-operator-index-k7g8c" Jan 26 17:02:58 crc kubenswrapper[4754]: I0126 17:02:58.765329 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-5h7fb" Jan 26 17:02:58 crc kubenswrapper[4754]: I0126 17:02:58.917078 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-k7g8c" Jan 26 17:02:59 crc kubenswrapper[4754]: I0126 17:02:59.019341 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qh8sw" event={"ID":"6666899f-4615-424e-a0c9-80bc260eef85","Type":"ContainerStarted","Data":"3dcbe30c19091114d3255a9235e0681ef1a75c3b62f4d34cf92146af875b2b60"} Jan 26 17:02:59 crc kubenswrapper[4754]: I0126 17:02:59.037138 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-qh8sw" podStartSLOduration=2.575943692 podStartE2EDuration="8.037124832s" podCreationTimestamp="2026-01-26 17:02:51 +0000 UTC" firstStartedPulling="2026-01-26 17:02:52.951722767 +0000 UTC m=+939.475903211" lastFinishedPulling="2026-01-26 17:02:58.412903917 +0000 UTC m=+944.937084351" observedRunningTime="2026-01-26 17:02:59.035832986 +0000 UTC m=+945.560013480" watchObservedRunningTime="2026-01-26 17:02:59.037124832 +0000 UTC m=+945.561305266" Jan 26 17:02:59 crc kubenswrapper[4754]: I0126 17:02:59.398969 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-k7g8c"] Jan 26 17:03:00 crc kubenswrapper[4754]: I0126 17:03:00.025499 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-k7g8c" event={"ID":"31e2139e-589c-4d52-8a92-5835f038d3b1","Type":"ContainerStarted","Data":"6e735f24efeaeeaa5cbace76d752ad3e01c7cea2e4db94e65da6449dcd41ff09"} Jan 26 17:03:01 crc kubenswrapper[4754]: I0126 17:03:01.865282 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-qh8sw" Jan 26 17:03:01 crc kubenswrapper[4754]: I0126 17:03:01.865635 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-qh8sw" Jan 26 17:03:01 crc kubenswrapper[4754]: I0126 17:03:01.923409 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-qh8sw" Jan 26 17:03:02 crc kubenswrapper[4754]: I0126 17:03:02.640019 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-k7g8c"] Jan 26 17:03:03 crc kubenswrapper[4754]: I0126 17:03:03.450554 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-zv866"] Jan 26 17:03:03 crc kubenswrapper[4754]: I0126 17:03:03.452444 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-zv866" Jan 26 17:03:03 crc kubenswrapper[4754]: I0126 17:03:03.461950 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-zv866"] Jan 26 17:03:03 crc kubenswrapper[4754]: I0126 17:03:03.501771 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4vwk9\" (UniqueName: \"kubernetes.io/projected/39f92bf3-b95c-4950-ad79-05dedec74b02-kube-api-access-4vwk9\") pod \"openstack-operator-index-zv866\" (UID: \"39f92bf3-b95c-4950-ad79-05dedec74b02\") " pod="openstack-operators/openstack-operator-index-zv866" Jan 26 17:03:03 crc kubenswrapper[4754]: I0126 17:03:03.603262 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4vwk9\" (UniqueName: \"kubernetes.io/projected/39f92bf3-b95c-4950-ad79-05dedec74b02-kube-api-access-4vwk9\") pod \"openstack-operator-index-zv866\" (UID: \"39f92bf3-b95c-4950-ad79-05dedec74b02\") " pod="openstack-operators/openstack-operator-index-zv866" Jan 26 17:03:03 crc kubenswrapper[4754]: I0126 17:03:03.629068 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4vwk9\" (UniqueName: \"kubernetes.io/projected/39f92bf3-b95c-4950-ad79-05dedec74b02-kube-api-access-4vwk9\") pod \"openstack-operator-index-zv866\" (UID: \"39f92bf3-b95c-4950-ad79-05dedec74b02\") " pod="openstack-operators/openstack-operator-index-zv866" Jan 26 17:03:03 crc kubenswrapper[4754]: I0126 17:03:03.741185 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-zbwg9" Jan 26 17:03:03 crc kubenswrapper[4754]: I0126 17:03:03.773643 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-zv866" Jan 26 17:03:03 crc kubenswrapper[4754]: I0126 17:03:03.821117 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-6968d8fdc4-8458x" Jan 26 17:03:04 crc kubenswrapper[4754]: I0126 17:03:04.053936 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-k7g8c" event={"ID":"31e2139e-589c-4d52-8a92-5835f038d3b1","Type":"ContainerStarted","Data":"52c1e7f68887ea695d19b4f82f1b23b7e0ec428bf7f42c1b107b010826bb3068"} Jan 26 17:03:04 crc kubenswrapper[4754]: I0126 17:03:04.054048 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-k7g8c" podUID="31e2139e-589c-4d52-8a92-5835f038d3b1" containerName="registry-server" containerID="cri-o://52c1e7f68887ea695d19b4f82f1b23b7e0ec428bf7f42c1b107b010826bb3068" gracePeriod=2 Jan 26 17:03:04 crc kubenswrapper[4754]: I0126 17:03:04.072502 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-k7g8c" podStartSLOduration=2.186308366 podStartE2EDuration="6.072461481s" podCreationTimestamp="2026-01-26 17:02:58 +0000 UTC" firstStartedPulling="2026-01-26 17:02:59.406554761 +0000 UTC m=+945.930735195" lastFinishedPulling="2026-01-26 17:03:03.292707876 +0000 UTC m=+949.816888310" observedRunningTime="2026-01-26 17:03:04.071755832 +0000 UTC m=+950.595936296" watchObservedRunningTime="2026-01-26 17:03:04.072461481 +0000 UTC m=+950.596641915" Jan 26 17:03:04 crc kubenswrapper[4754]: I0126 17:03:04.217623 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-zv866"] Jan 26 17:03:04 crc kubenswrapper[4754]: W0126 17:03:04.218853 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod39f92bf3_b95c_4950_ad79_05dedec74b02.slice/crio-52941d3e90ddce48448787736cbdee230a440ca3a09119b5fccf61531e4383e9 WatchSource:0}: Error finding container 52941d3e90ddce48448787736cbdee230a440ca3a09119b5fccf61531e4383e9: Status 404 returned error can't find the container with id 52941d3e90ddce48448787736cbdee230a440ca3a09119b5fccf61531e4383e9 Jan 26 17:03:05 crc kubenswrapper[4754]: I0126 17:03:05.068880 4754 generic.go:334] "Generic (PLEG): container finished" podID="31e2139e-589c-4d52-8a92-5835f038d3b1" containerID="52c1e7f68887ea695d19b4f82f1b23b7e0ec428bf7f42c1b107b010826bb3068" exitCode=0 Jan 26 17:03:05 crc kubenswrapper[4754]: I0126 17:03:05.068985 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-k7g8c" event={"ID":"31e2139e-589c-4d52-8a92-5835f038d3b1","Type":"ContainerDied","Data":"52c1e7f68887ea695d19b4f82f1b23b7e0ec428bf7f42c1b107b010826bb3068"} Jan 26 17:03:05 crc kubenswrapper[4754]: I0126 17:03:05.070384 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-zv866" event={"ID":"39f92bf3-b95c-4950-ad79-05dedec74b02","Type":"ContainerStarted","Data":"52941d3e90ddce48448787736cbdee230a440ca3a09119b5fccf61531e4383e9"} Jan 26 17:03:05 crc kubenswrapper[4754]: I0126 17:03:05.509354 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-k7g8c" Jan 26 17:03:05 crc kubenswrapper[4754]: I0126 17:03:05.530876 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sxf26\" (UniqueName: \"kubernetes.io/projected/31e2139e-589c-4d52-8a92-5835f038d3b1-kube-api-access-sxf26\") pod \"31e2139e-589c-4d52-8a92-5835f038d3b1\" (UID: \"31e2139e-589c-4d52-8a92-5835f038d3b1\") " Jan 26 17:03:05 crc kubenswrapper[4754]: I0126 17:03:05.537193 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31e2139e-589c-4d52-8a92-5835f038d3b1-kube-api-access-sxf26" (OuterVolumeSpecName: "kube-api-access-sxf26") pod "31e2139e-589c-4d52-8a92-5835f038d3b1" (UID: "31e2139e-589c-4d52-8a92-5835f038d3b1"). InnerVolumeSpecName "kube-api-access-sxf26". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:03:05 crc kubenswrapper[4754]: I0126 17:03:05.633560 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sxf26\" (UniqueName: \"kubernetes.io/projected/31e2139e-589c-4d52-8a92-5835f038d3b1-kube-api-access-sxf26\") on node \"crc\" DevicePath \"\"" Jan 26 17:03:06 crc kubenswrapper[4754]: I0126 17:03:06.079712 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-k7g8c" Jan 26 17:03:06 crc kubenswrapper[4754]: I0126 17:03:06.079706 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-k7g8c" event={"ID":"31e2139e-589c-4d52-8a92-5835f038d3b1","Type":"ContainerDied","Data":"6e735f24efeaeeaa5cbace76d752ad3e01c7cea2e4db94e65da6449dcd41ff09"} Jan 26 17:03:06 crc kubenswrapper[4754]: I0126 17:03:06.079843 4754 scope.go:117] "RemoveContainer" containerID="52c1e7f68887ea695d19b4f82f1b23b7e0ec428bf7f42c1b107b010826bb3068" Jan 26 17:03:06 crc kubenswrapper[4754]: I0126 17:03:06.082919 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-zv866" event={"ID":"39f92bf3-b95c-4950-ad79-05dedec74b02","Type":"ContainerStarted","Data":"be729797987b651fc3d194a8805b92830573052759a9b7f904b2bf0e1217b5d4"} Jan 26 17:03:06 crc kubenswrapper[4754]: I0126 17:03:06.103558 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-zv866" podStartSLOduration=2.018439988 podStartE2EDuration="3.103534496s" podCreationTimestamp="2026-01-26 17:03:03 +0000 UTC" firstStartedPulling="2026-01-26 17:03:04.224169606 +0000 UTC m=+950.748350040" lastFinishedPulling="2026-01-26 17:03:05.309264114 +0000 UTC m=+951.833444548" observedRunningTime="2026-01-26 17:03:06.101057357 +0000 UTC m=+952.625237811" watchObservedRunningTime="2026-01-26 17:03:06.103534496 +0000 UTC m=+952.627714940" Jan 26 17:03:06 crc kubenswrapper[4754]: I0126 17:03:06.116608 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-k7g8c"] Jan 26 17:03:06 crc kubenswrapper[4754]: I0126 17:03:06.122583 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-index-k7g8c"] Jan 26 17:03:07 crc kubenswrapper[4754]: I0126 17:03:07.129189 4754 patch_prober.go:28] interesting pod/machine-config-daemon-x65wv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 26 17:03:07 crc kubenswrapper[4754]: I0126 17:03:07.129535 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 26 17:03:07 crc kubenswrapper[4754]: I0126 17:03:07.780914 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31e2139e-589c-4d52-8a92-5835f038d3b1" path="/var/lib/kubelet/pods/31e2139e-589c-4d52-8a92-5835f038d3b1/volumes" Jan 26 17:03:11 crc kubenswrapper[4754]: I0126 17:03:11.903881 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-qh8sw" Jan 26 17:03:13 crc kubenswrapper[4754]: I0126 17:03:13.048168 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-qh8sw"] Jan 26 17:03:13 crc kubenswrapper[4754]: I0126 17:03:13.048555 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-qh8sw" podUID="6666899f-4615-424e-a0c9-80bc260eef85" containerName="registry-server" containerID="cri-o://3dcbe30c19091114d3255a9235e0681ef1a75c3b62f4d34cf92146af875b2b60" gracePeriod=2 Jan 26 17:03:13 crc kubenswrapper[4754]: I0126 17:03:13.731727 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-5h7fb" Jan 26 17:03:13 crc kubenswrapper[4754]: I0126 17:03:13.781469 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-zv866" Jan 26 17:03:13 crc kubenswrapper[4754]: I0126 17:03:13.781537 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-zv866" Jan 26 17:03:13 crc kubenswrapper[4754]: I0126 17:03:13.803476 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-zv866" Jan 26 17:03:13 crc kubenswrapper[4754]: I0126 17:03:13.944171 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qh8sw" Jan 26 17:03:14 crc kubenswrapper[4754]: I0126 17:03:14.132135 4754 generic.go:334] "Generic (PLEG): container finished" podID="6666899f-4615-424e-a0c9-80bc260eef85" containerID="3dcbe30c19091114d3255a9235e0681ef1a75c3b62f4d34cf92146af875b2b60" exitCode=0 Jan 26 17:03:14 crc kubenswrapper[4754]: I0126 17:03:14.132205 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qh8sw" Jan 26 17:03:14 crc kubenswrapper[4754]: I0126 17:03:14.132195 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qh8sw" event={"ID":"6666899f-4615-424e-a0c9-80bc260eef85","Type":"ContainerDied","Data":"3dcbe30c19091114d3255a9235e0681ef1a75c3b62f4d34cf92146af875b2b60"} Jan 26 17:03:14 crc kubenswrapper[4754]: I0126 17:03:14.132250 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qh8sw" event={"ID":"6666899f-4615-424e-a0c9-80bc260eef85","Type":"ContainerDied","Data":"f2c7a7c3e877b50aa6a3342098b754bb82fbcae455a432c1500a9ba539a0abf4"} Jan 26 17:03:14 crc kubenswrapper[4754]: I0126 17:03:14.132414 4754 scope.go:117] "RemoveContainer" containerID="3dcbe30c19091114d3255a9235e0681ef1a75c3b62f4d34cf92146af875b2b60" Jan 26 17:03:14 crc kubenswrapper[4754]: I0126 17:03:14.144247 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sgmxz\" (UniqueName: \"kubernetes.io/projected/6666899f-4615-424e-a0c9-80bc260eef85-kube-api-access-sgmxz\") pod \"6666899f-4615-424e-a0c9-80bc260eef85\" (UID: \"6666899f-4615-424e-a0c9-80bc260eef85\") " Jan 26 17:03:14 crc kubenswrapper[4754]: I0126 17:03:14.144304 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6666899f-4615-424e-a0c9-80bc260eef85-utilities\") pod \"6666899f-4615-424e-a0c9-80bc260eef85\" (UID: \"6666899f-4615-424e-a0c9-80bc260eef85\") " Jan 26 17:03:14 crc kubenswrapper[4754]: I0126 17:03:14.144361 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6666899f-4615-424e-a0c9-80bc260eef85-catalog-content\") pod \"6666899f-4615-424e-a0c9-80bc260eef85\" (UID: \"6666899f-4615-424e-a0c9-80bc260eef85\") " Jan 26 17:03:14 crc kubenswrapper[4754]: I0126 17:03:14.145416 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6666899f-4615-424e-a0c9-80bc260eef85-utilities" (OuterVolumeSpecName: "utilities") pod "6666899f-4615-424e-a0c9-80bc260eef85" (UID: "6666899f-4615-424e-a0c9-80bc260eef85"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:03:14 crc kubenswrapper[4754]: I0126 17:03:14.151858 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6666899f-4615-424e-a0c9-80bc260eef85-kube-api-access-sgmxz" (OuterVolumeSpecName: "kube-api-access-sgmxz") pod "6666899f-4615-424e-a0c9-80bc260eef85" (UID: "6666899f-4615-424e-a0c9-80bc260eef85"). InnerVolumeSpecName "kube-api-access-sgmxz". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:03:14 crc kubenswrapper[4754]: I0126 17:03:14.152899 4754 scope.go:117] "RemoveContainer" containerID="9c69138b331561fc2ce032914e28cffc9f28c66a8dcadd05ada91c5c0a8553e4" Jan 26 17:03:14 crc kubenswrapper[4754]: I0126 17:03:14.167181 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6666899f-4615-424e-a0c9-80bc260eef85-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6666899f-4615-424e-a0c9-80bc260eef85" (UID: "6666899f-4615-424e-a0c9-80bc260eef85"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:03:14 crc kubenswrapper[4754]: I0126 17:03:14.167415 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-zv866" Jan 26 17:03:14 crc kubenswrapper[4754]: I0126 17:03:14.184726 4754 scope.go:117] "RemoveContainer" containerID="6ce40029bc2f482d3f8847d8deec9eb922afb7272ab37ed70c016b805d748912" Jan 26 17:03:14 crc kubenswrapper[4754]: I0126 17:03:14.204751 4754 scope.go:117] "RemoveContainer" containerID="3dcbe30c19091114d3255a9235e0681ef1a75c3b62f4d34cf92146af875b2b60" Jan 26 17:03:14 crc kubenswrapper[4754]: E0126 17:03:14.205159 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3dcbe30c19091114d3255a9235e0681ef1a75c3b62f4d34cf92146af875b2b60\": container with ID starting with 3dcbe30c19091114d3255a9235e0681ef1a75c3b62f4d34cf92146af875b2b60 not found: ID does not exist" containerID="3dcbe30c19091114d3255a9235e0681ef1a75c3b62f4d34cf92146af875b2b60" Jan 26 17:03:14 crc kubenswrapper[4754]: I0126 17:03:14.205199 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3dcbe30c19091114d3255a9235e0681ef1a75c3b62f4d34cf92146af875b2b60"} err="failed to get container status \"3dcbe30c19091114d3255a9235e0681ef1a75c3b62f4d34cf92146af875b2b60\": rpc error: code = NotFound desc = could not find container \"3dcbe30c19091114d3255a9235e0681ef1a75c3b62f4d34cf92146af875b2b60\": container with ID starting with 3dcbe30c19091114d3255a9235e0681ef1a75c3b62f4d34cf92146af875b2b60 not found: ID does not exist" Jan 26 17:03:14 crc kubenswrapper[4754]: I0126 17:03:14.205223 4754 scope.go:117] "RemoveContainer" containerID="9c69138b331561fc2ce032914e28cffc9f28c66a8dcadd05ada91c5c0a8553e4" Jan 26 17:03:14 crc kubenswrapper[4754]: E0126 17:03:14.205534 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9c69138b331561fc2ce032914e28cffc9f28c66a8dcadd05ada91c5c0a8553e4\": container with ID starting with 9c69138b331561fc2ce032914e28cffc9f28c66a8dcadd05ada91c5c0a8553e4 not found: ID does not exist" containerID="9c69138b331561fc2ce032914e28cffc9f28c66a8dcadd05ada91c5c0a8553e4" Jan 26 17:03:14 crc kubenswrapper[4754]: I0126 17:03:14.205563 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9c69138b331561fc2ce032914e28cffc9f28c66a8dcadd05ada91c5c0a8553e4"} err="failed to get container status \"9c69138b331561fc2ce032914e28cffc9f28c66a8dcadd05ada91c5c0a8553e4\": rpc error: code = NotFound desc = could not find container \"9c69138b331561fc2ce032914e28cffc9f28c66a8dcadd05ada91c5c0a8553e4\": container with ID starting with 9c69138b331561fc2ce032914e28cffc9f28c66a8dcadd05ada91c5c0a8553e4 not found: ID does not exist" Jan 26 17:03:14 crc kubenswrapper[4754]: I0126 17:03:14.205579 4754 scope.go:117] "RemoveContainer" containerID="6ce40029bc2f482d3f8847d8deec9eb922afb7272ab37ed70c016b805d748912" Jan 26 17:03:14 crc kubenswrapper[4754]: E0126 17:03:14.206127 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6ce40029bc2f482d3f8847d8deec9eb922afb7272ab37ed70c016b805d748912\": container with ID starting with 6ce40029bc2f482d3f8847d8deec9eb922afb7272ab37ed70c016b805d748912 not found: ID does not exist" containerID="6ce40029bc2f482d3f8847d8deec9eb922afb7272ab37ed70c016b805d748912" Jan 26 17:03:14 crc kubenswrapper[4754]: I0126 17:03:14.206181 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6ce40029bc2f482d3f8847d8deec9eb922afb7272ab37ed70c016b805d748912"} err="failed to get container status \"6ce40029bc2f482d3f8847d8deec9eb922afb7272ab37ed70c016b805d748912\": rpc error: code = NotFound desc = could not find container \"6ce40029bc2f482d3f8847d8deec9eb922afb7272ab37ed70c016b805d748912\": container with ID starting with 6ce40029bc2f482d3f8847d8deec9eb922afb7272ab37ed70c016b805d748912 not found: ID does not exist" Jan 26 17:03:14 crc kubenswrapper[4754]: I0126 17:03:14.246573 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sgmxz\" (UniqueName: \"kubernetes.io/projected/6666899f-4615-424e-a0c9-80bc260eef85-kube-api-access-sgmxz\") on node \"crc\" DevicePath \"\"" Jan 26 17:03:14 crc kubenswrapper[4754]: I0126 17:03:14.246722 4754 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6666899f-4615-424e-a0c9-80bc260eef85-utilities\") on node \"crc\" DevicePath \"\"" Jan 26 17:03:14 crc kubenswrapper[4754]: I0126 17:03:14.246740 4754 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6666899f-4615-424e-a0c9-80bc260eef85-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 26 17:03:14 crc kubenswrapper[4754]: I0126 17:03:14.489306 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-qh8sw"] Jan 26 17:03:14 crc kubenswrapper[4754]: I0126 17:03:14.494604 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-qh8sw"] Jan 26 17:03:15 crc kubenswrapper[4754]: I0126 17:03:15.485471 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/80ea4f19923aee8b3bdba72b35be952869c3fe1246438cfe0124a675dcvvs8k"] Jan 26 17:03:15 crc kubenswrapper[4754]: E0126 17:03:15.487624 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6666899f-4615-424e-a0c9-80bc260eef85" containerName="extract-content" Jan 26 17:03:15 crc kubenswrapper[4754]: I0126 17:03:15.487648 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="6666899f-4615-424e-a0c9-80bc260eef85" containerName="extract-content" Jan 26 17:03:15 crc kubenswrapper[4754]: E0126 17:03:15.487689 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="31e2139e-589c-4d52-8a92-5835f038d3b1" containerName="registry-server" Jan 26 17:03:15 crc kubenswrapper[4754]: I0126 17:03:15.487695 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="31e2139e-589c-4d52-8a92-5835f038d3b1" containerName="registry-server" Jan 26 17:03:15 crc kubenswrapper[4754]: E0126 17:03:15.487706 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6666899f-4615-424e-a0c9-80bc260eef85" containerName="extract-utilities" Jan 26 17:03:15 crc kubenswrapper[4754]: I0126 17:03:15.487713 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="6666899f-4615-424e-a0c9-80bc260eef85" containerName="extract-utilities" Jan 26 17:03:15 crc kubenswrapper[4754]: E0126 17:03:15.487723 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6666899f-4615-424e-a0c9-80bc260eef85" containerName="registry-server" Jan 26 17:03:15 crc kubenswrapper[4754]: I0126 17:03:15.487729 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="6666899f-4615-424e-a0c9-80bc260eef85" containerName="registry-server" Jan 26 17:03:15 crc kubenswrapper[4754]: I0126 17:03:15.487845 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="31e2139e-589c-4d52-8a92-5835f038d3b1" containerName="registry-server" Jan 26 17:03:15 crc kubenswrapper[4754]: I0126 17:03:15.487857 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="6666899f-4615-424e-a0c9-80bc260eef85" containerName="registry-server" Jan 26 17:03:15 crc kubenswrapper[4754]: I0126 17:03:15.488745 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/80ea4f19923aee8b3bdba72b35be952869c3fe1246438cfe0124a675dcvvs8k" Jan 26 17:03:15 crc kubenswrapper[4754]: I0126 17:03:15.492123 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-xxfnh" Jan 26 17:03:15 crc kubenswrapper[4754]: I0126 17:03:15.508354 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/80ea4f19923aee8b3bdba72b35be952869c3fe1246438cfe0124a675dcvvs8k"] Jan 26 17:03:15 crc kubenswrapper[4754]: I0126 17:03:15.663396 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bl4kx\" (UniqueName: \"kubernetes.io/projected/187f901c-e262-4c66-9104-7a40cb64d0e4-kube-api-access-bl4kx\") pod \"80ea4f19923aee8b3bdba72b35be952869c3fe1246438cfe0124a675dcvvs8k\" (UID: \"187f901c-e262-4c66-9104-7a40cb64d0e4\") " pod="openstack-operators/80ea4f19923aee8b3bdba72b35be952869c3fe1246438cfe0124a675dcvvs8k" Jan 26 17:03:15 crc kubenswrapper[4754]: I0126 17:03:15.663537 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/187f901c-e262-4c66-9104-7a40cb64d0e4-util\") pod \"80ea4f19923aee8b3bdba72b35be952869c3fe1246438cfe0124a675dcvvs8k\" (UID: \"187f901c-e262-4c66-9104-7a40cb64d0e4\") " pod="openstack-operators/80ea4f19923aee8b3bdba72b35be952869c3fe1246438cfe0124a675dcvvs8k" Jan 26 17:03:15 crc kubenswrapper[4754]: I0126 17:03:15.663608 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/187f901c-e262-4c66-9104-7a40cb64d0e4-bundle\") pod \"80ea4f19923aee8b3bdba72b35be952869c3fe1246438cfe0124a675dcvvs8k\" (UID: \"187f901c-e262-4c66-9104-7a40cb64d0e4\") " pod="openstack-operators/80ea4f19923aee8b3bdba72b35be952869c3fe1246438cfe0124a675dcvvs8k" Jan 26 17:03:15 crc kubenswrapper[4754]: I0126 17:03:15.765039 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/187f901c-e262-4c66-9104-7a40cb64d0e4-bundle\") pod \"80ea4f19923aee8b3bdba72b35be952869c3fe1246438cfe0124a675dcvvs8k\" (UID: \"187f901c-e262-4c66-9104-7a40cb64d0e4\") " pod="openstack-operators/80ea4f19923aee8b3bdba72b35be952869c3fe1246438cfe0124a675dcvvs8k" Jan 26 17:03:15 crc kubenswrapper[4754]: I0126 17:03:15.765192 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bl4kx\" (UniqueName: \"kubernetes.io/projected/187f901c-e262-4c66-9104-7a40cb64d0e4-kube-api-access-bl4kx\") pod \"80ea4f19923aee8b3bdba72b35be952869c3fe1246438cfe0124a675dcvvs8k\" (UID: \"187f901c-e262-4c66-9104-7a40cb64d0e4\") " pod="openstack-operators/80ea4f19923aee8b3bdba72b35be952869c3fe1246438cfe0124a675dcvvs8k" Jan 26 17:03:15 crc kubenswrapper[4754]: I0126 17:03:15.765265 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/187f901c-e262-4c66-9104-7a40cb64d0e4-util\") pod \"80ea4f19923aee8b3bdba72b35be952869c3fe1246438cfe0124a675dcvvs8k\" (UID: \"187f901c-e262-4c66-9104-7a40cb64d0e4\") " pod="openstack-operators/80ea4f19923aee8b3bdba72b35be952869c3fe1246438cfe0124a675dcvvs8k" Jan 26 17:03:15 crc kubenswrapper[4754]: I0126 17:03:15.765625 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/187f901c-e262-4c66-9104-7a40cb64d0e4-bundle\") pod \"80ea4f19923aee8b3bdba72b35be952869c3fe1246438cfe0124a675dcvvs8k\" (UID: \"187f901c-e262-4c66-9104-7a40cb64d0e4\") " pod="openstack-operators/80ea4f19923aee8b3bdba72b35be952869c3fe1246438cfe0124a675dcvvs8k" Jan 26 17:03:15 crc kubenswrapper[4754]: I0126 17:03:15.765816 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/187f901c-e262-4c66-9104-7a40cb64d0e4-util\") pod \"80ea4f19923aee8b3bdba72b35be952869c3fe1246438cfe0124a675dcvvs8k\" (UID: \"187f901c-e262-4c66-9104-7a40cb64d0e4\") " pod="openstack-operators/80ea4f19923aee8b3bdba72b35be952869c3fe1246438cfe0124a675dcvvs8k" Jan 26 17:03:15 crc kubenswrapper[4754]: I0126 17:03:15.776487 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6666899f-4615-424e-a0c9-80bc260eef85" path="/var/lib/kubelet/pods/6666899f-4615-424e-a0c9-80bc260eef85/volumes" Jan 26 17:03:15 crc kubenswrapper[4754]: I0126 17:03:15.786732 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bl4kx\" (UniqueName: \"kubernetes.io/projected/187f901c-e262-4c66-9104-7a40cb64d0e4-kube-api-access-bl4kx\") pod \"80ea4f19923aee8b3bdba72b35be952869c3fe1246438cfe0124a675dcvvs8k\" (UID: \"187f901c-e262-4c66-9104-7a40cb64d0e4\") " pod="openstack-operators/80ea4f19923aee8b3bdba72b35be952869c3fe1246438cfe0124a675dcvvs8k" Jan 26 17:03:15 crc kubenswrapper[4754]: I0126 17:03:15.806977 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/80ea4f19923aee8b3bdba72b35be952869c3fe1246438cfe0124a675dcvvs8k" Jan 26 17:03:16 crc kubenswrapper[4754]: I0126 17:03:16.230144 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/80ea4f19923aee8b3bdba72b35be952869c3fe1246438cfe0124a675dcvvs8k"] Jan 26 17:03:16 crc kubenswrapper[4754]: W0126 17:03:16.233209 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod187f901c_e262_4c66_9104_7a40cb64d0e4.slice/crio-45e2f57db86bb52278873aa4bf2450d748d14e12d934a044ab9d0d9d8c88ccd5 WatchSource:0}: Error finding container 45e2f57db86bb52278873aa4bf2450d748d14e12d934a044ab9d0d9d8c88ccd5: Status 404 returned error can't find the container with id 45e2f57db86bb52278873aa4bf2450d748d14e12d934a044ab9d0d9d8c88ccd5 Jan 26 17:03:17 crc kubenswrapper[4754]: I0126 17:03:17.161935 4754 generic.go:334] "Generic (PLEG): container finished" podID="187f901c-e262-4c66-9104-7a40cb64d0e4" containerID="6ae5631b008a80349b0504416caefaa51a44428511cfefcb4e195dacdc92af6d" exitCode=0 Jan 26 17:03:17 crc kubenswrapper[4754]: I0126 17:03:17.162009 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/80ea4f19923aee8b3bdba72b35be952869c3fe1246438cfe0124a675dcvvs8k" event={"ID":"187f901c-e262-4c66-9104-7a40cb64d0e4","Type":"ContainerDied","Data":"6ae5631b008a80349b0504416caefaa51a44428511cfefcb4e195dacdc92af6d"} Jan 26 17:03:17 crc kubenswrapper[4754]: I0126 17:03:17.162296 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/80ea4f19923aee8b3bdba72b35be952869c3fe1246438cfe0124a675dcvvs8k" event={"ID":"187f901c-e262-4c66-9104-7a40cb64d0e4","Type":"ContainerStarted","Data":"45e2f57db86bb52278873aa4bf2450d748d14e12d934a044ab9d0d9d8c88ccd5"} Jan 26 17:03:18 crc kubenswrapper[4754]: I0126 17:03:18.171619 4754 generic.go:334] "Generic (PLEG): container finished" podID="187f901c-e262-4c66-9104-7a40cb64d0e4" containerID="d815211fc9874c5f4e1ef0e7e04ef27d75c97eacd5a60571fcd1e1a5ae2cbf92" exitCode=0 Jan 26 17:03:18 crc kubenswrapper[4754]: I0126 17:03:18.171726 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/80ea4f19923aee8b3bdba72b35be952869c3fe1246438cfe0124a675dcvvs8k" event={"ID":"187f901c-e262-4c66-9104-7a40cb64d0e4","Type":"ContainerDied","Data":"d815211fc9874c5f4e1ef0e7e04ef27d75c97eacd5a60571fcd1e1a5ae2cbf92"} Jan 26 17:03:19 crc kubenswrapper[4754]: I0126 17:03:19.184033 4754 generic.go:334] "Generic (PLEG): container finished" podID="187f901c-e262-4c66-9104-7a40cb64d0e4" containerID="e4ab2f4ecdd2cf8c7e65c6c3fd3bc8de65f2deb801fdda38b6fb77c7867f369a" exitCode=0 Jan 26 17:03:19 crc kubenswrapper[4754]: I0126 17:03:19.184126 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/80ea4f19923aee8b3bdba72b35be952869c3fe1246438cfe0124a675dcvvs8k" event={"ID":"187f901c-e262-4c66-9104-7a40cb64d0e4","Type":"ContainerDied","Data":"e4ab2f4ecdd2cf8c7e65c6c3fd3bc8de65f2deb801fdda38b6fb77c7867f369a"} Jan 26 17:03:20 crc kubenswrapper[4754]: I0126 17:03:20.413225 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/80ea4f19923aee8b3bdba72b35be952869c3fe1246438cfe0124a675dcvvs8k" Jan 26 17:03:20 crc kubenswrapper[4754]: I0126 17:03:20.524487 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/187f901c-e262-4c66-9104-7a40cb64d0e4-util\") pod \"187f901c-e262-4c66-9104-7a40cb64d0e4\" (UID: \"187f901c-e262-4c66-9104-7a40cb64d0e4\") " Jan 26 17:03:20 crc kubenswrapper[4754]: I0126 17:03:20.524569 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/187f901c-e262-4c66-9104-7a40cb64d0e4-bundle\") pod \"187f901c-e262-4c66-9104-7a40cb64d0e4\" (UID: \"187f901c-e262-4c66-9104-7a40cb64d0e4\") " Jan 26 17:03:20 crc kubenswrapper[4754]: I0126 17:03:20.524593 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bl4kx\" (UniqueName: \"kubernetes.io/projected/187f901c-e262-4c66-9104-7a40cb64d0e4-kube-api-access-bl4kx\") pod \"187f901c-e262-4c66-9104-7a40cb64d0e4\" (UID: \"187f901c-e262-4c66-9104-7a40cb64d0e4\") " Jan 26 17:03:20 crc kubenswrapper[4754]: I0126 17:03:20.525875 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/187f901c-e262-4c66-9104-7a40cb64d0e4-bundle" (OuterVolumeSpecName: "bundle") pod "187f901c-e262-4c66-9104-7a40cb64d0e4" (UID: "187f901c-e262-4c66-9104-7a40cb64d0e4"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:03:20 crc kubenswrapper[4754]: I0126 17:03:20.531707 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/187f901c-e262-4c66-9104-7a40cb64d0e4-kube-api-access-bl4kx" (OuterVolumeSpecName: "kube-api-access-bl4kx") pod "187f901c-e262-4c66-9104-7a40cb64d0e4" (UID: "187f901c-e262-4c66-9104-7a40cb64d0e4"). InnerVolumeSpecName "kube-api-access-bl4kx". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:03:20 crc kubenswrapper[4754]: I0126 17:03:20.544954 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/187f901c-e262-4c66-9104-7a40cb64d0e4-util" (OuterVolumeSpecName: "util") pod "187f901c-e262-4c66-9104-7a40cb64d0e4" (UID: "187f901c-e262-4c66-9104-7a40cb64d0e4"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:03:20 crc kubenswrapper[4754]: I0126 17:03:20.626170 4754 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/187f901c-e262-4c66-9104-7a40cb64d0e4-util\") on node \"crc\" DevicePath \"\"" Jan 26 17:03:20 crc kubenswrapper[4754]: I0126 17:03:20.626199 4754 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/187f901c-e262-4c66-9104-7a40cb64d0e4-bundle\") on node \"crc\" DevicePath \"\"" Jan 26 17:03:20 crc kubenswrapper[4754]: I0126 17:03:20.626209 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bl4kx\" (UniqueName: \"kubernetes.io/projected/187f901c-e262-4c66-9104-7a40cb64d0e4-kube-api-access-bl4kx\") on node \"crc\" DevicePath \"\"" Jan 26 17:03:21 crc kubenswrapper[4754]: I0126 17:03:21.200537 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/80ea4f19923aee8b3bdba72b35be952869c3fe1246438cfe0124a675dcvvs8k" event={"ID":"187f901c-e262-4c66-9104-7a40cb64d0e4","Type":"ContainerDied","Data":"45e2f57db86bb52278873aa4bf2450d748d14e12d934a044ab9d0d9d8c88ccd5"} Jan 26 17:03:21 crc kubenswrapper[4754]: I0126 17:03:21.200597 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="45e2f57db86bb52278873aa4bf2450d748d14e12d934a044ab9d0d9d8c88ccd5" Jan 26 17:03:21 crc kubenswrapper[4754]: I0126 17:03:21.200652 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/80ea4f19923aee8b3bdba72b35be952869c3fe1246438cfe0124a675dcvvs8k" Jan 26 17:03:27 crc kubenswrapper[4754]: I0126 17:03:27.036335 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-init-75cd685694-7x4dn"] Jan 26 17:03:27 crc kubenswrapper[4754]: E0126 17:03:27.037004 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="187f901c-e262-4c66-9104-7a40cb64d0e4" containerName="extract" Jan 26 17:03:27 crc kubenswrapper[4754]: I0126 17:03:27.037015 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="187f901c-e262-4c66-9104-7a40cb64d0e4" containerName="extract" Jan 26 17:03:27 crc kubenswrapper[4754]: E0126 17:03:27.037030 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="187f901c-e262-4c66-9104-7a40cb64d0e4" containerName="util" Jan 26 17:03:27 crc kubenswrapper[4754]: I0126 17:03:27.037036 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="187f901c-e262-4c66-9104-7a40cb64d0e4" containerName="util" Jan 26 17:03:27 crc kubenswrapper[4754]: E0126 17:03:27.037042 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="187f901c-e262-4c66-9104-7a40cb64d0e4" containerName="pull" Jan 26 17:03:27 crc kubenswrapper[4754]: I0126 17:03:27.037048 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="187f901c-e262-4c66-9104-7a40cb64d0e4" containerName="pull" Jan 26 17:03:27 crc kubenswrapper[4754]: I0126 17:03:27.037144 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="187f901c-e262-4c66-9104-7a40cb64d0e4" containerName="extract" Jan 26 17:03:27 crc kubenswrapper[4754]: I0126 17:03:27.037575 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-init-75cd685694-7x4dn" Jan 26 17:03:27 crc kubenswrapper[4754]: I0126 17:03:27.040457 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-init-dockercfg-pp5tt" Jan 26 17:03:27 crc kubenswrapper[4754]: I0126 17:03:27.110026 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-init-75cd685694-7x4dn"] Jan 26 17:03:27 crc kubenswrapper[4754]: I0126 17:03:27.133218 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fxhlm\" (UniqueName: \"kubernetes.io/projected/bfc14b7c-18c1-4442-bbb2-978889145894-kube-api-access-fxhlm\") pod \"openstack-operator-controller-init-75cd685694-7x4dn\" (UID: \"bfc14b7c-18c1-4442-bbb2-978889145894\") " pod="openstack-operators/openstack-operator-controller-init-75cd685694-7x4dn" Jan 26 17:03:27 crc kubenswrapper[4754]: I0126 17:03:27.234432 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fxhlm\" (UniqueName: \"kubernetes.io/projected/bfc14b7c-18c1-4442-bbb2-978889145894-kube-api-access-fxhlm\") pod \"openstack-operator-controller-init-75cd685694-7x4dn\" (UID: \"bfc14b7c-18c1-4442-bbb2-978889145894\") " pod="openstack-operators/openstack-operator-controller-init-75cd685694-7x4dn" Jan 26 17:03:27 crc kubenswrapper[4754]: I0126 17:03:27.256797 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fxhlm\" (UniqueName: \"kubernetes.io/projected/bfc14b7c-18c1-4442-bbb2-978889145894-kube-api-access-fxhlm\") pod \"openstack-operator-controller-init-75cd685694-7x4dn\" (UID: \"bfc14b7c-18c1-4442-bbb2-978889145894\") " pod="openstack-operators/openstack-operator-controller-init-75cd685694-7x4dn" Jan 26 17:03:27 crc kubenswrapper[4754]: I0126 17:03:27.353517 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-init-75cd685694-7x4dn" Jan 26 17:03:27 crc kubenswrapper[4754]: I0126 17:03:27.561225 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-init-75cd685694-7x4dn"] Jan 26 17:03:28 crc kubenswrapper[4754]: I0126 17:03:28.273471 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-init-75cd685694-7x4dn" event={"ID":"bfc14b7c-18c1-4442-bbb2-978889145894","Type":"ContainerStarted","Data":"7f97744f42d868e4a937d46cb308de962538becf58d2d5758b57f84438e936bf"} Jan 26 17:03:32 crc kubenswrapper[4754]: I0126 17:03:32.298892 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-init-75cd685694-7x4dn" event={"ID":"bfc14b7c-18c1-4442-bbb2-978889145894","Type":"ContainerStarted","Data":"40da022f398bace5b913608afd805f0ffaccdbd2b3f4dbde3b9c5ffa09d21d4d"} Jan 26 17:03:32 crc kubenswrapper[4754]: I0126 17:03:32.299385 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-init-75cd685694-7x4dn" Jan 26 17:03:32 crc kubenswrapper[4754]: I0126 17:03:32.354012 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-init-75cd685694-7x4dn" podStartSLOduration=1.730511664 podStartE2EDuration="5.353987785s" podCreationTimestamp="2026-01-26 17:03:27 +0000 UTC" firstStartedPulling="2026-01-26 17:03:27.573606953 +0000 UTC m=+974.097787387" lastFinishedPulling="2026-01-26 17:03:31.197083074 +0000 UTC m=+977.721263508" observedRunningTime="2026-01-26 17:03:32.345630989 +0000 UTC m=+978.869811443" watchObservedRunningTime="2026-01-26 17:03:32.353987785 +0000 UTC m=+978.878168239" Jan 26 17:03:37 crc kubenswrapper[4754]: I0126 17:03:37.129250 4754 patch_prober.go:28] interesting pod/machine-config-daemon-x65wv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 26 17:03:37 crc kubenswrapper[4754]: I0126 17:03:37.129827 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 26 17:03:37 crc kubenswrapper[4754]: I0126 17:03:37.129876 4754 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" Jan 26 17:03:37 crc kubenswrapper[4754]: I0126 17:03:37.130516 4754 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"2d157bed70bdc2504afc3419bb56bdffbceb22c7b7e8e1c5315d44974048523d"} pod="openshift-machine-config-operator/machine-config-daemon-x65wv" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 26 17:03:37 crc kubenswrapper[4754]: I0126 17:03:37.130584 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" containerName="machine-config-daemon" containerID="cri-o://2d157bed70bdc2504afc3419bb56bdffbceb22c7b7e8e1c5315d44974048523d" gracePeriod=600 Jan 26 17:03:37 crc kubenswrapper[4754]: I0126 17:03:37.358523 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-init-75cd685694-7x4dn" Jan 26 17:03:38 crc kubenswrapper[4754]: I0126 17:03:38.333079 4754 generic.go:334] "Generic (PLEG): container finished" podID="8c3718a4-f354-4284-92e0-fdfb45a692bd" containerID="2d157bed70bdc2504afc3419bb56bdffbceb22c7b7e8e1c5315d44974048523d" exitCode=0 Jan 26 17:03:38 crc kubenswrapper[4754]: I0126 17:03:38.333169 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" event={"ID":"8c3718a4-f354-4284-92e0-fdfb45a692bd","Type":"ContainerDied","Data":"2d157bed70bdc2504afc3419bb56bdffbceb22c7b7e8e1c5315d44974048523d"} Jan 26 17:03:38 crc kubenswrapper[4754]: I0126 17:03:38.333718 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" event={"ID":"8c3718a4-f354-4284-92e0-fdfb45a692bd","Type":"ContainerStarted","Data":"a1662bcf9213c8129b38140c4208b765dd0d2dd18def4d53852919fa5dab4d3e"} Jan 26 17:03:38 crc kubenswrapper[4754]: I0126 17:03:38.333745 4754 scope.go:117] "RemoveContainer" containerID="2258d22dac1a41453a2ed61fe65a0c8edeaa3d706e7abb2b8003f3eba6c44d91" Jan 26 17:03:56 crc kubenswrapper[4754]: I0126 17:03:56.858799 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7f86f8796f-5kbpl"] Jan 26 17:03:56 crc kubenswrapper[4754]: I0126 17:03:56.859996 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-7f86f8796f-5kbpl" Jan 26 17:03:56 crc kubenswrapper[4754]: I0126 17:03:56.862424 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-wwnnv" Jan 26 17:03:56 crc kubenswrapper[4754]: I0126 17:03:56.872998 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-7478f7dbf9-6vmhq"] Jan 26 17:03:56 crc kubenswrapper[4754]: I0126 17:03:56.874091 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-7478f7dbf9-6vmhq" Jan 26 17:03:56 crc kubenswrapper[4754]: I0126 17:03:56.876980 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-sp6bt" Jan 26 17:03:56 crc kubenswrapper[4754]: I0126 17:03:56.877179 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-b45d7bf98-bggfb"] Jan 26 17:03:56 crc kubenswrapper[4754]: I0126 17:03:56.878296 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-b45d7bf98-bggfb" Jan 26 17:03:56 crc kubenswrapper[4754]: I0126 17:03:56.881719 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-tqv2r" Jan 26 17:03:56 crc kubenswrapper[4754]: I0126 17:03:56.888209 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7f86f8796f-5kbpl"] Jan 26 17:03:56 crc kubenswrapper[4754]: I0126 17:03:56.892647 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-7478f7dbf9-6vmhq"] Jan 26 17:03:56 crc kubenswrapper[4754]: I0126 17:03:56.903970 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-78fdd796fd-c5kf9"] Jan 26 17:03:56 crc kubenswrapper[4754]: I0126 17:03:56.904652 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-78fdd796fd-c5kf9" Jan 26 17:03:56 crc kubenswrapper[4754]: I0126 17:03:56.906727 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-94t8d" Jan 26 17:03:56 crc kubenswrapper[4754]: I0126 17:03:56.928722 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-b45d7bf98-bggfb"] Jan 26 17:03:56 crc kubenswrapper[4754]: I0126 17:03:56.936806 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-594c8c9d5d-hgnm9"] Jan 26 17:03:56 crc kubenswrapper[4754]: I0126 17:03:56.937515 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-594c8c9d5d-hgnm9" Jan 26 17:03:56 crc kubenswrapper[4754]: I0126 17:03:56.949375 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-56wkx\" (UniqueName: \"kubernetes.io/projected/5e071de1-60dc-49d8-b965-90c2f99a6e02-kube-api-access-56wkx\") pod \"designate-operator-controller-manager-b45d7bf98-bggfb\" (UID: \"5e071de1-60dc-49d8-b965-90c2f99a6e02\") " pod="openstack-operators/designate-operator-controller-manager-b45d7bf98-bggfb" Jan 26 17:03:56 crc kubenswrapper[4754]: I0126 17:03:56.949518 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dbxrg\" (UniqueName: \"kubernetes.io/projected/cb2628e4-680e-489a-8fc9-d39986c74301-kube-api-access-dbxrg\") pod \"cinder-operator-controller-manager-7478f7dbf9-6vmhq\" (UID: \"cb2628e4-680e-489a-8fc9-d39986c74301\") " pod="openstack-operators/cinder-operator-controller-manager-7478f7dbf9-6vmhq" Jan 26 17:03:56 crc kubenswrapper[4754]: I0126 17:03:56.949557 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-765pt\" (UniqueName: \"kubernetes.io/projected/1a9a2d55-592f-4320-8e2e-49f65ca72dfc-kube-api-access-765pt\") pod \"barbican-operator-controller-manager-7f86f8796f-5kbpl\" (UID: \"1a9a2d55-592f-4320-8e2e-49f65ca72dfc\") " pod="openstack-operators/barbican-operator-controller-manager-7f86f8796f-5kbpl" Jan 26 17:03:56 crc kubenswrapper[4754]: I0126 17:03:56.949580 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-twqnx\" (UniqueName: \"kubernetes.io/projected/ee98073e-2fdb-4b3e-acb8-00c71df55fa7-kube-api-access-twqnx\") pod \"glance-operator-controller-manager-78fdd796fd-c5kf9\" (UID: \"ee98073e-2fdb-4b3e-acb8-00c71df55fa7\") " pod="openstack-operators/glance-operator-controller-manager-78fdd796fd-c5kf9" Jan 26 17:03:56 crc kubenswrapper[4754]: I0126 17:03:56.957318 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-v46nz" Jan 26 17:03:56 crc kubenswrapper[4754]: I0126 17:03:56.959616 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-78fdd796fd-c5kf9"] Jan 26 17:03:57 crc kubenswrapper[4754]: I0126 17:03:57.029740 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-594c8c9d5d-hgnm9"] Jan 26 17:03:57 crc kubenswrapper[4754]: I0126 17:03:57.050328 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-56wkx\" (UniqueName: \"kubernetes.io/projected/5e071de1-60dc-49d8-b965-90c2f99a6e02-kube-api-access-56wkx\") pod \"designate-operator-controller-manager-b45d7bf98-bggfb\" (UID: \"5e071de1-60dc-49d8-b965-90c2f99a6e02\") " pod="openstack-operators/designate-operator-controller-manager-b45d7bf98-bggfb" Jan 26 17:03:57 crc kubenswrapper[4754]: I0126 17:03:57.050392 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v8vwz\" (UniqueName: \"kubernetes.io/projected/f747ae1d-8181-4ef2-b332-b14db483aab6-kube-api-access-v8vwz\") pod \"heat-operator-controller-manager-594c8c9d5d-hgnm9\" (UID: \"f747ae1d-8181-4ef2-b332-b14db483aab6\") " pod="openstack-operators/heat-operator-controller-manager-594c8c9d5d-hgnm9" Jan 26 17:03:57 crc kubenswrapper[4754]: I0126 17:03:57.050470 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dbxrg\" (UniqueName: \"kubernetes.io/projected/cb2628e4-680e-489a-8fc9-d39986c74301-kube-api-access-dbxrg\") pod \"cinder-operator-controller-manager-7478f7dbf9-6vmhq\" (UID: \"cb2628e4-680e-489a-8fc9-d39986c74301\") " pod="openstack-operators/cinder-operator-controller-manager-7478f7dbf9-6vmhq" Jan 26 17:03:57 crc kubenswrapper[4754]: I0126 17:03:57.050496 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-765pt\" (UniqueName: \"kubernetes.io/projected/1a9a2d55-592f-4320-8e2e-49f65ca72dfc-kube-api-access-765pt\") pod \"barbican-operator-controller-manager-7f86f8796f-5kbpl\" (UID: \"1a9a2d55-592f-4320-8e2e-49f65ca72dfc\") " pod="openstack-operators/barbican-operator-controller-manager-7f86f8796f-5kbpl" Jan 26 17:03:57 crc kubenswrapper[4754]: I0126 17:03:57.050519 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-twqnx\" (UniqueName: \"kubernetes.io/projected/ee98073e-2fdb-4b3e-acb8-00c71df55fa7-kube-api-access-twqnx\") pod \"glance-operator-controller-manager-78fdd796fd-c5kf9\" (UID: \"ee98073e-2fdb-4b3e-acb8-00c71df55fa7\") " pod="openstack-operators/glance-operator-controller-manager-78fdd796fd-c5kf9" Jan 26 17:03:57 crc kubenswrapper[4754]: I0126 17:03:57.066714 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-77d5c5b54f-8sz77"] Jan 26 17:03:57 crc kubenswrapper[4754]: I0126 17:03:57.067516 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-77d5c5b54f-8sz77" Jan 26 17:03:57 crc kubenswrapper[4754]: I0126 17:03:57.080009 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-jlbwc" Jan 26 17:03:57 crc kubenswrapper[4754]: I0126 17:03:57.080389 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-694cf4f878-p66zm"] Jan 26 17:03:57 crc kubenswrapper[4754]: I0126 17:03:57.081146 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-694cf4f878-p66zm" Jan 26 17:03:57 crc kubenswrapper[4754]: I0126 17:03:57.080401 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-56wkx\" (UniqueName: \"kubernetes.io/projected/5e071de1-60dc-49d8-b965-90c2f99a6e02-kube-api-access-56wkx\") pod \"designate-operator-controller-manager-b45d7bf98-bggfb\" (UID: \"5e071de1-60dc-49d8-b965-90c2f99a6e02\") " pod="openstack-operators/designate-operator-controller-manager-b45d7bf98-bggfb" Jan 26 17:03:57 crc kubenswrapper[4754]: I0126 17:03:57.087519 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-dlczc" Jan 26 17:03:57 crc kubenswrapper[4754]: I0126 17:03:57.090020 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-twqnx\" (UniqueName: \"kubernetes.io/projected/ee98073e-2fdb-4b3e-acb8-00c71df55fa7-kube-api-access-twqnx\") pod \"glance-operator-controller-manager-78fdd796fd-c5kf9\" (UID: \"ee98073e-2fdb-4b3e-acb8-00c71df55fa7\") " pod="openstack-operators/glance-operator-controller-manager-78fdd796fd-c5kf9" Jan 26 17:03:57 crc kubenswrapper[4754]: I0126 17:03:57.090088 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-765pt\" (UniqueName: \"kubernetes.io/projected/1a9a2d55-592f-4320-8e2e-49f65ca72dfc-kube-api-access-765pt\") pod \"barbican-operator-controller-manager-7f86f8796f-5kbpl\" (UID: \"1a9a2d55-592f-4320-8e2e-49f65ca72dfc\") " pod="openstack-operators/barbican-operator-controller-manager-7f86f8796f-5kbpl" Jan 26 17:03:57 crc kubenswrapper[4754]: I0126 17:03:57.094433 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Jan 26 17:03:57 crc kubenswrapper[4754]: I0126 17:03:57.095265 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dbxrg\" (UniqueName: \"kubernetes.io/projected/cb2628e4-680e-489a-8fc9-d39986c74301-kube-api-access-dbxrg\") pod \"cinder-operator-controller-manager-7478f7dbf9-6vmhq\" (UID: \"cb2628e4-680e-489a-8fc9-d39986c74301\") " pod="openstack-operators/cinder-operator-controller-manager-7478f7dbf9-6vmhq" Jan 26 17:03:57 crc kubenswrapper[4754]: I0126 17:03:57.103735 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-77d5c5b54f-8sz77"] Jan 26 17:03:57 crc kubenswrapper[4754]: I0126 17:03:57.122758 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-598f7747c9-fvcgg"] Jan 26 17:03:57 crc kubenswrapper[4754]: I0126 17:03:57.123560 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-598f7747c9-fvcgg" Jan 26 17:03:57 crc kubenswrapper[4754]: I0126 17:03:57.142956 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-b6kpc" Jan 26 17:03:57 crc kubenswrapper[4754]: I0126 17:03:57.147720 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-694cf4f878-p66zm"] Jan 26 17:03:57 crc kubenswrapper[4754]: I0126 17:03:57.151440 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-smt7h\" (UniqueName: \"kubernetes.io/projected/28161331-7731-433f-845f-2ebe1daf5fd0-kube-api-access-smt7h\") pod \"ironic-operator-controller-manager-598f7747c9-fvcgg\" (UID: \"28161331-7731-433f-845f-2ebe1daf5fd0\") " pod="openstack-operators/ironic-operator-controller-manager-598f7747c9-fvcgg" Jan 26 17:03:57 crc kubenswrapper[4754]: I0126 17:03:57.151587 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-95tv6\" (UniqueName: \"kubernetes.io/projected/5702a910-71d1-4acd-93aa-9379bc3147ce-kube-api-access-95tv6\") pod \"infra-operator-controller-manager-694cf4f878-p66zm\" (UID: \"5702a910-71d1-4acd-93aa-9379bc3147ce\") " pod="openstack-operators/infra-operator-controller-manager-694cf4f878-p66zm" Jan 26 17:03:57 crc kubenswrapper[4754]: I0126 17:03:57.151715 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-45qhs\" (UniqueName: \"kubernetes.io/projected/fd6b6ccf-b7e9-41fb-b663-a3392d075880-kube-api-access-45qhs\") pod \"horizon-operator-controller-manager-77d5c5b54f-8sz77\" (UID: \"fd6b6ccf-b7e9-41fb-b663-a3392d075880\") " pod="openstack-operators/horizon-operator-controller-manager-77d5c5b54f-8sz77" Jan 26 17:03:57 crc kubenswrapper[4754]: I0126 17:03:57.151803 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v8vwz\" (UniqueName: \"kubernetes.io/projected/f747ae1d-8181-4ef2-b332-b14db483aab6-kube-api-access-v8vwz\") pod \"heat-operator-controller-manager-594c8c9d5d-hgnm9\" (UID: \"f747ae1d-8181-4ef2-b332-b14db483aab6\") " pod="openstack-operators/heat-operator-controller-manager-594c8c9d5d-hgnm9" Jan 26 17:03:57 crc kubenswrapper[4754]: I0126 17:03:57.151894 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5702a910-71d1-4acd-93aa-9379bc3147ce-cert\") pod \"infra-operator-controller-manager-694cf4f878-p66zm\" (UID: \"5702a910-71d1-4acd-93aa-9379bc3147ce\") " pod="openstack-operators/infra-operator-controller-manager-694cf4f878-p66zm" Jan 26 17:03:57 crc kubenswrapper[4754]: I0126 17:03:57.160726 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-b8b6d4659-d8db9"] Jan 26 17:03:57 crc kubenswrapper[4754]: I0126 17:03:57.161514 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-b8b6d4659-d8db9" Jan 26 17:03:57 crc kubenswrapper[4754]: I0126 17:03:57.163564 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-srkh7" Jan 26 17:03:57 crc kubenswrapper[4754]: I0126 17:03:57.168359 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-598f7747c9-fvcgg"] Jan 26 17:03:57 crc kubenswrapper[4754]: I0126 17:03:57.185873 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-78c6999f6f-jb5ps"] Jan 26 17:03:57 crc kubenswrapper[4754]: I0126 17:03:57.186544 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-b8b6d4659-d8db9"] Jan 26 17:03:57 crc kubenswrapper[4754]: I0126 17:03:57.186618 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-78c6999f6f-jb5ps" Jan 26 17:03:57 crc kubenswrapper[4754]: I0126 17:03:57.187227 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-7f86f8796f-5kbpl" Jan 26 17:03:57 crc kubenswrapper[4754]: I0126 17:03:57.199207 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-ztp9k" Jan 26 17:03:57 crc kubenswrapper[4754]: I0126 17:03:57.205464 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v8vwz\" (UniqueName: \"kubernetes.io/projected/f747ae1d-8181-4ef2-b332-b14db483aab6-kube-api-access-v8vwz\") pod \"heat-operator-controller-manager-594c8c9d5d-hgnm9\" (UID: \"f747ae1d-8181-4ef2-b332-b14db483aab6\") " pod="openstack-operators/heat-operator-controller-manager-594c8c9d5d-hgnm9" Jan 26 17:03:57 crc kubenswrapper[4754]: I0126 17:03:57.208113 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-7478f7dbf9-6vmhq" Jan 26 17:03:57 crc kubenswrapper[4754]: I0126 17:03:57.219950 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-78c6999f6f-jb5ps"] Jan 26 17:03:57 crc kubenswrapper[4754]: I0126 17:03:57.222207 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-b45d7bf98-bggfb" Jan 26 17:03:57 crc kubenswrapper[4754]: I0126 17:03:57.238797 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-78fdd796fd-c5kf9" Jan 26 17:03:57 crc kubenswrapper[4754]: I0126 17:03:57.254527 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-6b9fb5fdcb-fhwr4"] Jan 26 17:03:57 crc kubenswrapper[4754]: I0126 17:03:57.255375 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-45qhs\" (UniqueName: \"kubernetes.io/projected/fd6b6ccf-b7e9-41fb-b663-a3392d075880-kube-api-access-45qhs\") pod \"horizon-operator-controller-manager-77d5c5b54f-8sz77\" (UID: \"fd6b6ccf-b7e9-41fb-b663-a3392d075880\") " pod="openstack-operators/horizon-operator-controller-manager-77d5c5b54f-8sz77" Jan 26 17:03:57 crc kubenswrapper[4754]: I0126 17:03:57.255502 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wnr8c\" (UniqueName: \"kubernetes.io/projected/e3aac642-b3ee-4394-9f1d-bfac315bf162-kube-api-access-wnr8c\") pod \"keystone-operator-controller-manager-b8b6d4659-d8db9\" (UID: \"e3aac642-b3ee-4394-9f1d-bfac315bf162\") " pod="openstack-operators/keystone-operator-controller-manager-b8b6d4659-d8db9" Jan 26 17:03:57 crc kubenswrapper[4754]: I0126 17:03:57.255608 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5702a910-71d1-4acd-93aa-9379bc3147ce-cert\") pod \"infra-operator-controller-manager-694cf4f878-p66zm\" (UID: \"5702a910-71d1-4acd-93aa-9379bc3147ce\") " pod="openstack-operators/infra-operator-controller-manager-694cf4f878-p66zm" Jan 26 17:03:57 crc kubenswrapper[4754]: I0126 17:03:57.255741 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-smt7h\" (UniqueName: \"kubernetes.io/projected/28161331-7731-433f-845f-2ebe1daf5fd0-kube-api-access-smt7h\") pod \"ironic-operator-controller-manager-598f7747c9-fvcgg\" (UID: \"28161331-7731-433f-845f-2ebe1daf5fd0\") " pod="openstack-operators/ironic-operator-controller-manager-598f7747c9-fvcgg" Jan 26 17:03:57 crc kubenswrapper[4754]: I0126 17:03:57.255863 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-95tv6\" (UniqueName: \"kubernetes.io/projected/5702a910-71d1-4acd-93aa-9379bc3147ce-kube-api-access-95tv6\") pod \"infra-operator-controller-manager-694cf4f878-p66zm\" (UID: \"5702a910-71d1-4acd-93aa-9379bc3147ce\") " pod="openstack-operators/infra-operator-controller-manager-694cf4f878-p66zm" Jan 26 17:03:57 crc kubenswrapper[4754]: I0126 17:03:57.256155 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sr67m\" (UniqueName: \"kubernetes.io/projected/16e535a9-3fc0-4385-b809-51a2bf719657-kube-api-access-sr67m\") pod \"manila-operator-controller-manager-78c6999f6f-jb5ps\" (UID: \"16e535a9-3fc0-4385-b809-51a2bf719657\") " pod="openstack-operators/manila-operator-controller-manager-78c6999f6f-jb5ps" Jan 26 17:03:57 crc kubenswrapper[4754]: E0126 17:03:57.256558 4754 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Jan 26 17:03:57 crc kubenswrapper[4754]: E0126 17:03:57.256854 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5702a910-71d1-4acd-93aa-9379bc3147ce-cert podName:5702a910-71d1-4acd-93aa-9379bc3147ce nodeName:}" failed. No retries permitted until 2026-01-26 17:03:57.756838222 +0000 UTC m=+1004.281018656 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/5702a910-71d1-4acd-93aa-9379bc3147ce-cert") pod "infra-operator-controller-manager-694cf4f878-p66zm" (UID: "5702a910-71d1-4acd-93aa-9379bc3147ce") : secret "infra-operator-webhook-server-cert" not found Jan 26 17:03:57 crc kubenswrapper[4754]: I0126 17:03:57.268391 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-6b9fb5fdcb-fhwr4" Jan 26 17:03:57 crc kubenswrapper[4754]: I0126 17:03:57.293536 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-wbbsm" Jan 26 17:03:57 crc kubenswrapper[4754]: I0126 17:03:57.294206 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-594c8c9d5d-hgnm9" Jan 26 17:03:57 crc kubenswrapper[4754]: I0126 17:03:57.306394 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-smt7h\" (UniqueName: \"kubernetes.io/projected/28161331-7731-433f-845f-2ebe1daf5fd0-kube-api-access-smt7h\") pod \"ironic-operator-controller-manager-598f7747c9-fvcgg\" (UID: \"28161331-7731-433f-845f-2ebe1daf5fd0\") " pod="openstack-operators/ironic-operator-controller-manager-598f7747c9-fvcgg" Jan 26 17:03:57 crc kubenswrapper[4754]: I0126 17:03:57.308285 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-45qhs\" (UniqueName: \"kubernetes.io/projected/fd6b6ccf-b7e9-41fb-b663-a3392d075880-kube-api-access-45qhs\") pod \"horizon-operator-controller-manager-77d5c5b54f-8sz77\" (UID: \"fd6b6ccf-b7e9-41fb-b663-a3392d075880\") " pod="openstack-operators/horizon-operator-controller-manager-77d5c5b54f-8sz77" Jan 26 17:03:57 crc kubenswrapper[4754]: I0126 17:03:57.313577 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-95tv6\" (UniqueName: \"kubernetes.io/projected/5702a910-71d1-4acd-93aa-9379bc3147ce-kube-api-access-95tv6\") pod \"infra-operator-controller-manager-694cf4f878-p66zm\" (UID: \"5702a910-71d1-4acd-93aa-9379bc3147ce\") " pod="openstack-operators/infra-operator-controller-manager-694cf4f878-p66zm" Jan 26 17:03:57 crc kubenswrapper[4754]: I0126 17:03:57.323605 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-6b9fb5fdcb-fhwr4"] Jan 26 17:03:57 crc kubenswrapper[4754]: I0126 17:03:57.328054 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-78d58447c5-pgmv6"] Jan 26 17:03:57 crc kubenswrapper[4754]: I0126 17:03:57.328820 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-78d58447c5-pgmv6" Jan 26 17:03:57 crc kubenswrapper[4754]: I0126 17:03:57.347075 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-hphgs" Jan 26 17:03:57 crc kubenswrapper[4754]: I0126 17:03:57.347315 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-7bdb645866-27hg7"] Jan 26 17:03:57 crc kubenswrapper[4754]: I0126 17:03:57.348263 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-7bdb645866-27hg7" Jan 26 17:03:57 crc kubenswrapper[4754]: I0126 17:03:57.350200 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-z67d6" Jan 26 17:03:57 crc kubenswrapper[4754]: I0126 17:03:57.359370 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wnr8c\" (UniqueName: \"kubernetes.io/projected/e3aac642-b3ee-4394-9f1d-bfac315bf162-kube-api-access-wnr8c\") pod \"keystone-operator-controller-manager-b8b6d4659-d8db9\" (UID: \"e3aac642-b3ee-4394-9f1d-bfac315bf162\") " pod="openstack-operators/keystone-operator-controller-manager-b8b6d4659-d8db9" Jan 26 17:03:57 crc kubenswrapper[4754]: I0126 17:03:57.359460 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qc6lm\" (UniqueName: \"kubernetes.io/projected/d23565a1-cb57-4cbe-88a2-48a01e8056f3-kube-api-access-qc6lm\") pod \"nova-operator-controller-manager-7bdb645866-27hg7\" (UID: \"d23565a1-cb57-4cbe-88a2-48a01e8056f3\") " pod="openstack-operators/nova-operator-controller-manager-7bdb645866-27hg7" Jan 26 17:03:57 crc kubenswrapper[4754]: I0126 17:03:57.359505 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sr67m\" (UniqueName: \"kubernetes.io/projected/16e535a9-3fc0-4385-b809-51a2bf719657-kube-api-access-sr67m\") pod \"manila-operator-controller-manager-78c6999f6f-jb5ps\" (UID: \"16e535a9-3fc0-4385-b809-51a2bf719657\") " pod="openstack-operators/manila-operator-controller-manager-78c6999f6f-jb5ps" Jan 26 17:03:57 crc kubenswrapper[4754]: I0126 17:03:57.359546 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l2x9c\" (UniqueName: \"kubernetes.io/projected/2ab80530-a54c-41bf-8e4f-c895dbacb368-kube-api-access-l2x9c\") pod \"neutron-operator-controller-manager-78d58447c5-pgmv6\" (UID: \"2ab80530-a54c-41bf-8e4f-c895dbacb368\") " pod="openstack-operators/neutron-operator-controller-manager-78d58447c5-pgmv6" Jan 26 17:03:57 crc kubenswrapper[4754]: I0126 17:03:57.359578 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9hbrr\" (UniqueName: \"kubernetes.io/projected/add7fe6b-a864-439d-a3f4-80c9cc80ddc1-kube-api-access-9hbrr\") pod \"mariadb-operator-controller-manager-6b9fb5fdcb-fhwr4\" (UID: \"add7fe6b-a864-439d-a3f4-80c9cc80ddc1\") " pod="openstack-operators/mariadb-operator-controller-manager-6b9fb5fdcb-fhwr4" Jan 26 17:03:57 crc kubenswrapper[4754]: I0126 17:03:57.363441 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-78d58447c5-pgmv6"] Jan 26 17:03:57 crc kubenswrapper[4754]: I0126 17:03:57.377282 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-5f4cd88d46-fkz84"] Jan 26 17:03:57 crc kubenswrapper[4754]: I0126 17:03:57.378288 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-5f4cd88d46-fkz84" Jan 26 17:03:57 crc kubenswrapper[4754]: I0126 17:03:57.380900 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-kqzdr" Jan 26 17:03:57 crc kubenswrapper[4754]: I0126 17:03:57.397809 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-7bdb645866-27hg7"] Jan 26 17:03:57 crc kubenswrapper[4754]: I0126 17:03:57.416094 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-5f4cd88d46-fkz84"] Jan 26 17:03:57 crc kubenswrapper[4754]: I0126 17:03:57.422163 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-6f75f45d54-kppds"] Jan 26 17:03:57 crc kubenswrapper[4754]: I0126 17:03:57.423059 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sr67m\" (UniqueName: \"kubernetes.io/projected/16e535a9-3fc0-4385-b809-51a2bf719657-kube-api-access-sr67m\") pod \"manila-operator-controller-manager-78c6999f6f-jb5ps\" (UID: \"16e535a9-3fc0-4385-b809-51a2bf719657\") " pod="openstack-operators/manila-operator-controller-manager-78c6999f6f-jb5ps" Jan 26 17:03:57 crc kubenswrapper[4754]: I0126 17:03:57.423408 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-6f75f45d54-kppds" Jan 26 17:03:57 crc kubenswrapper[4754]: I0126 17:03:57.432033 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-tsdnq" Jan 26 17:03:57 crc kubenswrapper[4754]: I0126 17:03:57.436286 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wnr8c\" (UniqueName: \"kubernetes.io/projected/e3aac642-b3ee-4394-9f1d-bfac315bf162-kube-api-access-wnr8c\") pod \"keystone-operator-controller-manager-b8b6d4659-d8db9\" (UID: \"e3aac642-b3ee-4394-9f1d-bfac315bf162\") " pod="openstack-operators/keystone-operator-controller-manager-b8b6d4659-d8db9" Jan 26 17:03:57 crc kubenswrapper[4754]: I0126 17:03:57.452261 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-79d5ccc684-gqtkb"] Jan 26 17:03:57 crc kubenswrapper[4754]: I0126 17:03:57.452933 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-79d5ccc684-gqtkb" Jan 26 17:03:57 crc kubenswrapper[4754]: I0126 17:03:57.462378 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-77d5c5b54f-8sz77" Jan 26 17:03:57 crc kubenswrapper[4754]: I0126 17:03:57.462497 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-547cbdb99f-fmvdv"] Jan 26 17:03:57 crc kubenswrapper[4754]: I0126 17:03:57.462638 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l2x9c\" (UniqueName: \"kubernetes.io/projected/2ab80530-a54c-41bf-8e4f-c895dbacb368-kube-api-access-l2x9c\") pod \"neutron-operator-controller-manager-78d58447c5-pgmv6\" (UID: \"2ab80530-a54c-41bf-8e4f-c895dbacb368\") " pod="openstack-operators/neutron-operator-controller-manager-78d58447c5-pgmv6" Jan 26 17:03:57 crc kubenswrapper[4754]: I0126 17:03:57.462744 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9hbrr\" (UniqueName: \"kubernetes.io/projected/add7fe6b-a864-439d-a3f4-80c9cc80ddc1-kube-api-access-9hbrr\") pod \"mariadb-operator-controller-manager-6b9fb5fdcb-fhwr4\" (UID: \"add7fe6b-a864-439d-a3f4-80c9cc80ddc1\") " pod="openstack-operators/mariadb-operator-controller-manager-6b9fb5fdcb-fhwr4" Jan 26 17:03:57 crc kubenswrapper[4754]: I0126 17:03:57.463443 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-547cbdb99f-fmvdv" Jan 26 17:03:57 crc kubenswrapper[4754]: I0126 17:03:57.463522 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qc6lm\" (UniqueName: \"kubernetes.io/projected/d23565a1-cb57-4cbe-88a2-48a01e8056f3-kube-api-access-qc6lm\") pod \"nova-operator-controller-manager-7bdb645866-27hg7\" (UID: \"d23565a1-cb57-4cbe-88a2-48a01e8056f3\") " pod="openstack-operators/nova-operator-controller-manager-7bdb645866-27hg7" Jan 26 17:03:57 crc kubenswrapper[4754]: I0126 17:03:57.463905 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-nbrbh" Jan 26 17:03:57 crc kubenswrapper[4754]: I0126 17:03:57.469251 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-t5xh5" Jan 26 17:03:57 crc kubenswrapper[4754]: I0126 17:03:57.473862 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-6b68b8b854sclfc"] Jan 26 17:03:57 crc kubenswrapper[4754]: I0126 17:03:57.475213 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6b68b8b854sclfc" Jan 26 17:03:57 crc kubenswrapper[4754]: I0126 17:03:57.480131 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-nxv4s" Jan 26 17:03:57 crc kubenswrapper[4754]: I0126 17:03:57.480322 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Jan 26 17:03:57 crc kubenswrapper[4754]: I0126 17:03:57.480610 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-6f75f45d54-kppds"] Jan 26 17:03:57 crc kubenswrapper[4754]: I0126 17:03:57.489892 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9hbrr\" (UniqueName: \"kubernetes.io/projected/add7fe6b-a864-439d-a3f4-80c9cc80ddc1-kube-api-access-9hbrr\") pod \"mariadb-operator-controller-manager-6b9fb5fdcb-fhwr4\" (UID: \"add7fe6b-a864-439d-a3f4-80c9cc80ddc1\") " pod="openstack-operators/mariadb-operator-controller-manager-6b9fb5fdcb-fhwr4" Jan 26 17:03:57 crc kubenswrapper[4754]: I0126 17:03:57.527460 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qc6lm\" (UniqueName: \"kubernetes.io/projected/d23565a1-cb57-4cbe-88a2-48a01e8056f3-kube-api-access-qc6lm\") pod \"nova-operator-controller-manager-7bdb645866-27hg7\" (UID: \"d23565a1-cb57-4cbe-88a2-48a01e8056f3\") " pod="openstack-operators/nova-operator-controller-manager-7bdb645866-27hg7" Jan 26 17:03:57 crc kubenswrapper[4754]: I0126 17:03:57.534415 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l2x9c\" (UniqueName: \"kubernetes.io/projected/2ab80530-a54c-41bf-8e4f-c895dbacb368-kube-api-access-l2x9c\") pod \"neutron-operator-controller-manager-78d58447c5-pgmv6\" (UID: \"2ab80530-a54c-41bf-8e4f-c895dbacb368\") " pod="openstack-operators/neutron-operator-controller-manager-78d58447c5-pgmv6" Jan 26 17:03:57 crc kubenswrapper[4754]: I0126 17:03:57.537099 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-598f7747c9-fvcgg" Jan 26 17:03:57 crc kubenswrapper[4754]: I0126 17:03:57.548417 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-85cd9769bb-dq4qw"] Jan 26 17:03:57 crc kubenswrapper[4754]: I0126 17:03:57.553140 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-85cd9769bb-dq4qw" Jan 26 17:03:57 crc kubenswrapper[4754]: I0126 17:03:57.558410 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-msgb4" Jan 26 17:03:57 crc kubenswrapper[4754]: I0126 17:03:57.573968 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-shvr5\" (UniqueName: \"kubernetes.io/projected/8353f62d-eb53-46c3-ba4d-2d643f4f960b-kube-api-access-shvr5\") pod \"ovn-operator-controller-manager-6f75f45d54-kppds\" (UID: \"8353f62d-eb53-46c3-ba4d-2d643f4f960b\") " pod="openstack-operators/ovn-operator-controller-manager-6f75f45d54-kppds" Jan 26 17:03:57 crc kubenswrapper[4754]: I0126 17:03:57.574176 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rtgsw\" (UniqueName: \"kubernetes.io/projected/56addd62-aeed-4139-a0db-37292a6acf8a-kube-api-access-rtgsw\") pod \"placement-operator-controller-manager-79d5ccc684-gqtkb\" (UID: \"56addd62-aeed-4139-a0db-37292a6acf8a\") " pod="openstack-operators/placement-operator-controller-manager-79d5ccc684-gqtkb" Jan 26 17:03:57 crc kubenswrapper[4754]: I0126 17:03:57.574257 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7hzhd\" (UniqueName: \"kubernetes.io/projected/4fdc4d46-264a-4689-8ad5-0ed253f805df-kube-api-access-7hzhd\") pod \"octavia-operator-controller-manager-5f4cd88d46-fkz84\" (UID: \"4fdc4d46-264a-4689-8ad5-0ed253f805df\") " pod="openstack-operators/octavia-operator-controller-manager-5f4cd88d46-fkz84" Jan 26 17:03:57 crc kubenswrapper[4754]: I0126 17:03:57.599392 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-547cbdb99f-fmvdv"] Jan 26 17:03:57 crc kubenswrapper[4754]: I0126 17:03:57.617779 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-79d5ccc684-gqtkb"] Jan 26 17:03:57 crc kubenswrapper[4754]: I0126 17:03:57.627743 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-6b68b8b854sclfc"] Jan 26 17:03:57 crc kubenswrapper[4754]: I0126 17:03:57.635378 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-85cd9769bb-dq4qw"] Jan 26 17:03:57 crc kubenswrapper[4754]: I0126 17:03:57.636380 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-b8b6d4659-d8db9" Jan 26 17:03:57 crc kubenswrapper[4754]: I0126 17:03:57.656782 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-69797bbcbd-mr8bp"] Jan 26 17:03:57 crc kubenswrapper[4754]: I0126 17:03:57.657738 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-69797bbcbd-mr8bp" Jan 26 17:03:57 crc kubenswrapper[4754]: I0126 17:03:57.658606 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-78c6999f6f-jb5ps" Jan 26 17:03:57 crc kubenswrapper[4754]: I0126 17:03:57.660842 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-gz5lc" Jan 26 17:03:57 crc kubenswrapper[4754]: I0126 17:03:57.678053 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-shvr5\" (UniqueName: \"kubernetes.io/projected/8353f62d-eb53-46c3-ba4d-2d643f4f960b-kube-api-access-shvr5\") pod \"ovn-operator-controller-manager-6f75f45d54-kppds\" (UID: \"8353f62d-eb53-46c3-ba4d-2d643f4f960b\") " pod="openstack-operators/ovn-operator-controller-manager-6f75f45d54-kppds" Jan 26 17:03:57 crc kubenswrapper[4754]: I0126 17:03:57.678300 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gvsb5\" (UniqueName: \"kubernetes.io/projected/f43cc0a3-41e9-4e09-837f-322a53221560-kube-api-access-gvsb5\") pod \"openstack-baremetal-operator-controller-manager-6b68b8b854sclfc\" (UID: \"f43cc0a3-41e9-4e09-837f-322a53221560\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6b68b8b854sclfc" Jan 26 17:03:57 crc kubenswrapper[4754]: I0126 17:03:57.678334 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rtgsw\" (UniqueName: \"kubernetes.io/projected/56addd62-aeed-4139-a0db-37292a6acf8a-kube-api-access-rtgsw\") pod \"placement-operator-controller-manager-79d5ccc684-gqtkb\" (UID: \"56addd62-aeed-4139-a0db-37292a6acf8a\") " pod="openstack-operators/placement-operator-controller-manager-79d5ccc684-gqtkb" Jan 26 17:03:57 crc kubenswrapper[4754]: I0126 17:03:57.678368 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-64c6f\" (UniqueName: \"kubernetes.io/projected/a8584d32-796d-47cd-8dd6-233374660688-kube-api-access-64c6f\") pod \"test-operator-controller-manager-69797bbcbd-mr8bp\" (UID: \"a8584d32-796d-47cd-8dd6-233374660688\") " pod="openstack-operators/test-operator-controller-manager-69797bbcbd-mr8bp" Jan 26 17:03:57 crc kubenswrapper[4754]: I0126 17:03:57.678396 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7hzhd\" (UniqueName: \"kubernetes.io/projected/4fdc4d46-264a-4689-8ad5-0ed253f805df-kube-api-access-7hzhd\") pod \"octavia-operator-controller-manager-5f4cd88d46-fkz84\" (UID: \"4fdc4d46-264a-4689-8ad5-0ed253f805df\") " pod="openstack-operators/octavia-operator-controller-manager-5f4cd88d46-fkz84" Jan 26 17:03:57 crc kubenswrapper[4754]: I0126 17:03:57.678426 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8hscl\" (UniqueName: \"kubernetes.io/projected/13391684-794c-48b2-8d7f-23a122f8acc6-kube-api-access-8hscl\") pod \"telemetry-operator-controller-manager-85cd9769bb-dq4qw\" (UID: \"13391684-794c-48b2-8d7f-23a122f8acc6\") " pod="openstack-operators/telemetry-operator-controller-manager-85cd9769bb-dq4qw" Jan 26 17:03:57 crc kubenswrapper[4754]: I0126 17:03:57.678459 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f43cc0a3-41e9-4e09-837f-322a53221560-cert\") pod \"openstack-baremetal-operator-controller-manager-6b68b8b854sclfc\" (UID: \"f43cc0a3-41e9-4e09-837f-322a53221560\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6b68b8b854sclfc" Jan 26 17:03:57 crc kubenswrapper[4754]: I0126 17:03:57.678478 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-57rp9\" (UniqueName: \"kubernetes.io/projected/aaee86fe-018f-4055-b8c5-98e3795c53d8-kube-api-access-57rp9\") pod \"swift-operator-controller-manager-547cbdb99f-fmvdv\" (UID: \"aaee86fe-018f-4055-b8c5-98e3795c53d8\") " pod="openstack-operators/swift-operator-controller-manager-547cbdb99f-fmvdv" Jan 26 17:03:57 crc kubenswrapper[4754]: I0126 17:03:57.681398 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-69797bbcbd-mr8bp"] Jan 26 17:03:57 crc kubenswrapper[4754]: I0126 17:03:57.687399 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-6b9fb5fdcb-fhwr4" Jan 26 17:03:57 crc kubenswrapper[4754]: I0126 17:03:57.701068 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rtgsw\" (UniqueName: \"kubernetes.io/projected/56addd62-aeed-4139-a0db-37292a6acf8a-kube-api-access-rtgsw\") pod \"placement-operator-controller-manager-79d5ccc684-gqtkb\" (UID: \"56addd62-aeed-4139-a0db-37292a6acf8a\") " pod="openstack-operators/placement-operator-controller-manager-79d5ccc684-gqtkb" Jan 26 17:03:57 crc kubenswrapper[4754]: I0126 17:03:57.701560 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7hzhd\" (UniqueName: \"kubernetes.io/projected/4fdc4d46-264a-4689-8ad5-0ed253f805df-kube-api-access-7hzhd\") pod \"octavia-operator-controller-manager-5f4cd88d46-fkz84\" (UID: \"4fdc4d46-264a-4689-8ad5-0ed253f805df\") " pod="openstack-operators/octavia-operator-controller-manager-5f4cd88d46-fkz84" Jan 26 17:03:57 crc kubenswrapper[4754]: I0126 17:03:57.702864 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-shvr5\" (UniqueName: \"kubernetes.io/projected/8353f62d-eb53-46c3-ba4d-2d643f4f960b-kube-api-access-shvr5\") pod \"ovn-operator-controller-manager-6f75f45d54-kppds\" (UID: \"8353f62d-eb53-46c3-ba4d-2d643f4f960b\") " pod="openstack-operators/ovn-operator-controller-manager-6f75f45d54-kppds" Jan 26 17:03:57 crc kubenswrapper[4754]: I0126 17:03:57.720155 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-78d58447c5-pgmv6" Jan 26 17:03:57 crc kubenswrapper[4754]: I0126 17:03:57.731028 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-564965969-k7nbl"] Jan 26 17:03:57 crc kubenswrapper[4754]: I0126 17:03:57.731972 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-564965969-k7nbl" Jan 26 17:03:57 crc kubenswrapper[4754]: I0126 17:03:57.733621 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-k6r7f" Jan 26 17:03:57 crc kubenswrapper[4754]: I0126 17:03:57.738647 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-564965969-k7nbl"] Jan 26 17:03:57 crc kubenswrapper[4754]: I0126 17:03:57.754343 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-7bdb645866-27hg7" Jan 26 17:03:57 crc kubenswrapper[4754]: I0126 17:03:57.779587 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f43cc0a3-41e9-4e09-837f-322a53221560-cert\") pod \"openstack-baremetal-operator-controller-manager-6b68b8b854sclfc\" (UID: \"f43cc0a3-41e9-4e09-837f-322a53221560\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6b68b8b854sclfc" Jan 26 17:03:57 crc kubenswrapper[4754]: I0126 17:03:57.779621 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-57rp9\" (UniqueName: \"kubernetes.io/projected/aaee86fe-018f-4055-b8c5-98e3795c53d8-kube-api-access-57rp9\") pod \"swift-operator-controller-manager-547cbdb99f-fmvdv\" (UID: \"aaee86fe-018f-4055-b8c5-98e3795c53d8\") " pod="openstack-operators/swift-operator-controller-manager-547cbdb99f-fmvdv" Jan 26 17:03:57 crc kubenswrapper[4754]: I0126 17:03:57.779657 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gvsb5\" (UniqueName: \"kubernetes.io/projected/f43cc0a3-41e9-4e09-837f-322a53221560-kube-api-access-gvsb5\") pod \"openstack-baremetal-operator-controller-manager-6b68b8b854sclfc\" (UID: \"f43cc0a3-41e9-4e09-837f-322a53221560\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6b68b8b854sclfc" Jan 26 17:03:57 crc kubenswrapper[4754]: I0126 17:03:57.779715 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-64c6f\" (UniqueName: \"kubernetes.io/projected/a8584d32-796d-47cd-8dd6-233374660688-kube-api-access-64c6f\") pod \"test-operator-controller-manager-69797bbcbd-mr8bp\" (UID: \"a8584d32-796d-47cd-8dd6-233374660688\") " pod="openstack-operators/test-operator-controller-manager-69797bbcbd-mr8bp" Jan 26 17:03:57 crc kubenswrapper[4754]: I0126 17:03:57.779747 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5702a910-71d1-4acd-93aa-9379bc3147ce-cert\") pod \"infra-operator-controller-manager-694cf4f878-p66zm\" (UID: \"5702a910-71d1-4acd-93aa-9379bc3147ce\") " pod="openstack-operators/infra-operator-controller-manager-694cf4f878-p66zm" Jan 26 17:03:57 crc kubenswrapper[4754]: I0126 17:03:57.779767 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8hscl\" (UniqueName: \"kubernetes.io/projected/13391684-794c-48b2-8d7f-23a122f8acc6-kube-api-access-8hscl\") pod \"telemetry-operator-controller-manager-85cd9769bb-dq4qw\" (UID: \"13391684-794c-48b2-8d7f-23a122f8acc6\") " pod="openstack-operators/telemetry-operator-controller-manager-85cd9769bb-dq4qw" Jan 26 17:03:57 crc kubenswrapper[4754]: E0126 17:03:57.781976 4754 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Jan 26 17:03:57 crc kubenswrapper[4754]: E0126 17:03:57.782034 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f43cc0a3-41e9-4e09-837f-322a53221560-cert podName:f43cc0a3-41e9-4e09-837f-322a53221560 nodeName:}" failed. No retries permitted until 2026-01-26 17:03:58.282019751 +0000 UTC m=+1004.806200185 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/f43cc0a3-41e9-4e09-837f-322a53221560-cert") pod "openstack-baremetal-operator-controller-manager-6b68b8b854sclfc" (UID: "f43cc0a3-41e9-4e09-837f-322a53221560") : secret "openstack-baremetal-operator-webhook-server-cert" not found Jan 26 17:03:57 crc kubenswrapper[4754]: E0126 17:03:57.782147 4754 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Jan 26 17:03:57 crc kubenswrapper[4754]: E0126 17:03:57.782169 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5702a910-71d1-4acd-93aa-9379bc3147ce-cert podName:5702a910-71d1-4acd-93aa-9379bc3147ce nodeName:}" failed. No retries permitted until 2026-01-26 17:03:58.782162885 +0000 UTC m=+1005.306343319 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/5702a910-71d1-4acd-93aa-9379bc3147ce-cert") pod "infra-operator-controller-manager-694cf4f878-p66zm" (UID: "5702a910-71d1-4acd-93aa-9379bc3147ce") : secret "infra-operator-webhook-server-cert" not found Jan 26 17:03:57 crc kubenswrapper[4754]: I0126 17:03:57.782300 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-5f4cd88d46-fkz84" Jan 26 17:03:57 crc kubenswrapper[4754]: I0126 17:03:57.798944 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-59fcd4bdb5-gvt4r"] Jan 26 17:03:57 crc kubenswrapper[4754]: I0126 17:03:57.799970 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8hscl\" (UniqueName: \"kubernetes.io/projected/13391684-794c-48b2-8d7f-23a122f8acc6-kube-api-access-8hscl\") pod \"telemetry-operator-controller-manager-85cd9769bb-dq4qw\" (UID: \"13391684-794c-48b2-8d7f-23a122f8acc6\") " pod="openstack-operators/telemetry-operator-controller-manager-85cd9769bb-dq4qw" Jan 26 17:03:57 crc kubenswrapper[4754]: I0126 17:03:57.801319 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-59fcd4bdb5-gvt4r" Jan 26 17:03:57 crc kubenswrapper[4754]: I0126 17:03:57.802757 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-59fcd4bdb5-gvt4r"] Jan 26 17:03:57 crc kubenswrapper[4754]: I0126 17:03:57.804343 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Jan 26 17:03:57 crc kubenswrapper[4754]: I0126 17:03:57.804711 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"metrics-server-cert" Jan 26 17:03:57 crc kubenswrapper[4754]: I0126 17:03:57.804880 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-ppmm7" Jan 26 17:03:57 crc kubenswrapper[4754]: I0126 17:03:57.804943 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-64c6f\" (UniqueName: \"kubernetes.io/projected/a8584d32-796d-47cd-8dd6-233374660688-kube-api-access-64c6f\") pod \"test-operator-controller-manager-69797bbcbd-mr8bp\" (UID: \"a8584d32-796d-47cd-8dd6-233374660688\") " pod="openstack-operators/test-operator-controller-manager-69797bbcbd-mr8bp" Jan 26 17:03:57 crc kubenswrapper[4754]: I0126 17:03:57.812739 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gvsb5\" (UniqueName: \"kubernetes.io/projected/f43cc0a3-41e9-4e09-837f-322a53221560-kube-api-access-gvsb5\") pod \"openstack-baremetal-operator-controller-manager-6b68b8b854sclfc\" (UID: \"f43cc0a3-41e9-4e09-837f-322a53221560\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6b68b8b854sclfc" Jan 26 17:03:57 crc kubenswrapper[4754]: I0126 17:03:57.823365 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-57rp9\" (UniqueName: \"kubernetes.io/projected/aaee86fe-018f-4055-b8c5-98e3795c53d8-kube-api-access-57rp9\") pod \"swift-operator-controller-manager-547cbdb99f-fmvdv\" (UID: \"aaee86fe-018f-4055-b8c5-98e3795c53d8\") " pod="openstack-operators/swift-operator-controller-manager-547cbdb99f-fmvdv" Jan 26 17:03:57 crc kubenswrapper[4754]: I0126 17:03:57.823432 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-mqmkw"] Jan 26 17:03:57 crc kubenswrapper[4754]: I0126 17:03:57.824290 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-mqmkw" Jan 26 17:03:57 crc kubenswrapper[4754]: I0126 17:03:57.832277 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-mqmkw"] Jan 26 17:03:57 crc kubenswrapper[4754]: I0126 17:03:57.833273 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-v95qk" Jan 26 17:03:57 crc kubenswrapper[4754]: I0126 17:03:57.847269 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-6f75f45d54-kppds" Jan 26 17:03:57 crc kubenswrapper[4754]: I0126 17:03:57.858093 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-79d5ccc684-gqtkb" Jan 26 17:03:57 crc kubenswrapper[4754]: I0126 17:03:57.873280 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-547cbdb99f-fmvdv" Jan 26 17:03:57 crc kubenswrapper[4754]: I0126 17:03:57.881809 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-95qgc\" (UniqueName: \"kubernetes.io/projected/138cc11d-6be6-43bb-994e-94ea39bb2e42-kube-api-access-95qgc\") pod \"watcher-operator-controller-manager-564965969-k7nbl\" (UID: \"138cc11d-6be6-43bb-994e-94ea39bb2e42\") " pod="openstack-operators/watcher-operator-controller-manager-564965969-k7nbl" Jan 26 17:03:57 crc kubenswrapper[4754]: I0126 17:03:57.909395 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-85cd9769bb-dq4qw" Jan 26 17:03:57 crc kubenswrapper[4754]: I0126 17:03:57.982734 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/93622668-0766-46f4-a216-83a7a17f36fc-metrics-certs\") pod \"openstack-operator-controller-manager-59fcd4bdb5-gvt4r\" (UID: \"93622668-0766-46f4-a216-83a7a17f36fc\") " pod="openstack-operators/openstack-operator-controller-manager-59fcd4bdb5-gvt4r" Jan 26 17:03:57 crc kubenswrapper[4754]: I0126 17:03:57.982809 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/93622668-0766-46f4-a216-83a7a17f36fc-webhook-certs\") pod \"openstack-operator-controller-manager-59fcd4bdb5-gvt4r\" (UID: \"93622668-0766-46f4-a216-83a7a17f36fc\") " pod="openstack-operators/openstack-operator-controller-manager-59fcd4bdb5-gvt4r" Jan 26 17:03:57 crc kubenswrapper[4754]: I0126 17:03:57.982898 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-95qgc\" (UniqueName: \"kubernetes.io/projected/138cc11d-6be6-43bb-994e-94ea39bb2e42-kube-api-access-95qgc\") pod \"watcher-operator-controller-manager-564965969-k7nbl\" (UID: \"138cc11d-6be6-43bb-994e-94ea39bb2e42\") " pod="openstack-operators/watcher-operator-controller-manager-564965969-k7nbl" Jan 26 17:03:57 crc kubenswrapper[4754]: I0126 17:03:57.982941 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z6tpd\" (UniqueName: \"kubernetes.io/projected/93622668-0766-46f4-a216-83a7a17f36fc-kube-api-access-z6tpd\") pod \"openstack-operator-controller-manager-59fcd4bdb5-gvt4r\" (UID: \"93622668-0766-46f4-a216-83a7a17f36fc\") " pod="openstack-operators/openstack-operator-controller-manager-59fcd4bdb5-gvt4r" Jan 26 17:03:57 crc kubenswrapper[4754]: I0126 17:03:57.982985 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5mb2z\" (UniqueName: \"kubernetes.io/projected/b48b5414-4034-4c0e-9d25-5053f84cf246-kube-api-access-5mb2z\") pod \"rabbitmq-cluster-operator-manager-668c99d594-mqmkw\" (UID: \"b48b5414-4034-4c0e-9d25-5053f84cf246\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-mqmkw" Jan 26 17:03:57 crc kubenswrapper[4754]: I0126 17:03:57.983134 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-7478f7dbf9-6vmhq"] Jan 26 17:03:58 crc kubenswrapper[4754]: I0126 17:03:58.006450 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-95qgc\" (UniqueName: \"kubernetes.io/projected/138cc11d-6be6-43bb-994e-94ea39bb2e42-kube-api-access-95qgc\") pod \"watcher-operator-controller-manager-564965969-k7nbl\" (UID: \"138cc11d-6be6-43bb-994e-94ea39bb2e42\") " pod="openstack-operators/watcher-operator-controller-manager-564965969-k7nbl" Jan 26 17:03:58 crc kubenswrapper[4754]: I0126 17:03:58.006986 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-69797bbcbd-mr8bp" Jan 26 17:03:58 crc kubenswrapper[4754]: I0126 17:03:58.084090 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-564965969-k7nbl" Jan 26 17:03:58 crc kubenswrapper[4754]: I0126 17:03:58.084485 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/93622668-0766-46f4-a216-83a7a17f36fc-webhook-certs\") pod \"openstack-operator-controller-manager-59fcd4bdb5-gvt4r\" (UID: \"93622668-0766-46f4-a216-83a7a17f36fc\") " pod="openstack-operators/openstack-operator-controller-manager-59fcd4bdb5-gvt4r" Jan 26 17:03:58 crc kubenswrapper[4754]: I0126 17:03:58.084586 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z6tpd\" (UniqueName: \"kubernetes.io/projected/93622668-0766-46f4-a216-83a7a17f36fc-kube-api-access-z6tpd\") pod \"openstack-operator-controller-manager-59fcd4bdb5-gvt4r\" (UID: \"93622668-0766-46f4-a216-83a7a17f36fc\") " pod="openstack-operators/openstack-operator-controller-manager-59fcd4bdb5-gvt4r" Jan 26 17:03:58 crc kubenswrapper[4754]: I0126 17:03:58.084624 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5mb2z\" (UniqueName: \"kubernetes.io/projected/b48b5414-4034-4c0e-9d25-5053f84cf246-kube-api-access-5mb2z\") pod \"rabbitmq-cluster-operator-manager-668c99d594-mqmkw\" (UID: \"b48b5414-4034-4c0e-9d25-5053f84cf246\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-mqmkw" Jan 26 17:03:58 crc kubenswrapper[4754]: I0126 17:03:58.084648 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/93622668-0766-46f4-a216-83a7a17f36fc-metrics-certs\") pod \"openstack-operator-controller-manager-59fcd4bdb5-gvt4r\" (UID: \"93622668-0766-46f4-a216-83a7a17f36fc\") " pod="openstack-operators/openstack-operator-controller-manager-59fcd4bdb5-gvt4r" Jan 26 17:03:58 crc kubenswrapper[4754]: E0126 17:03:58.084769 4754 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Jan 26 17:03:58 crc kubenswrapper[4754]: E0126 17:03:58.084819 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/93622668-0766-46f4-a216-83a7a17f36fc-metrics-certs podName:93622668-0766-46f4-a216-83a7a17f36fc nodeName:}" failed. No retries permitted until 2026-01-26 17:03:58.584805249 +0000 UTC m=+1005.108985673 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/93622668-0766-46f4-a216-83a7a17f36fc-metrics-certs") pod "openstack-operator-controller-manager-59fcd4bdb5-gvt4r" (UID: "93622668-0766-46f4-a216-83a7a17f36fc") : secret "metrics-server-cert" not found Jan 26 17:03:58 crc kubenswrapper[4754]: E0126 17:03:58.084935 4754 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Jan 26 17:03:58 crc kubenswrapper[4754]: E0126 17:03:58.085032 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/93622668-0766-46f4-a216-83a7a17f36fc-webhook-certs podName:93622668-0766-46f4-a216-83a7a17f36fc nodeName:}" failed. No retries permitted until 2026-01-26 17:03:58.585002934 +0000 UTC m=+1005.109183368 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/93622668-0766-46f4-a216-83a7a17f36fc-webhook-certs") pod "openstack-operator-controller-manager-59fcd4bdb5-gvt4r" (UID: "93622668-0766-46f4-a216-83a7a17f36fc") : secret "webhook-server-cert" not found Jan 26 17:03:58 crc kubenswrapper[4754]: I0126 17:03:58.106099 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z6tpd\" (UniqueName: \"kubernetes.io/projected/93622668-0766-46f4-a216-83a7a17f36fc-kube-api-access-z6tpd\") pod \"openstack-operator-controller-manager-59fcd4bdb5-gvt4r\" (UID: \"93622668-0766-46f4-a216-83a7a17f36fc\") " pod="openstack-operators/openstack-operator-controller-manager-59fcd4bdb5-gvt4r" Jan 26 17:03:58 crc kubenswrapper[4754]: I0126 17:03:58.115510 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5mb2z\" (UniqueName: \"kubernetes.io/projected/b48b5414-4034-4c0e-9d25-5053f84cf246-kube-api-access-5mb2z\") pod \"rabbitmq-cluster-operator-manager-668c99d594-mqmkw\" (UID: \"b48b5414-4034-4c0e-9d25-5053f84cf246\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-mqmkw" Jan 26 17:03:58 crc kubenswrapper[4754]: I0126 17:03:58.193241 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-mqmkw" Jan 26 17:03:58 crc kubenswrapper[4754]: I0126 17:03:58.288486 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f43cc0a3-41e9-4e09-837f-322a53221560-cert\") pod \"openstack-baremetal-operator-controller-manager-6b68b8b854sclfc\" (UID: \"f43cc0a3-41e9-4e09-837f-322a53221560\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6b68b8b854sclfc" Jan 26 17:03:58 crc kubenswrapper[4754]: E0126 17:03:58.288652 4754 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Jan 26 17:03:58 crc kubenswrapper[4754]: E0126 17:03:58.288743 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f43cc0a3-41e9-4e09-837f-322a53221560-cert podName:f43cc0a3-41e9-4e09-837f-322a53221560 nodeName:}" failed. No retries permitted until 2026-01-26 17:03:59.288728299 +0000 UTC m=+1005.812908733 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/f43cc0a3-41e9-4e09-837f-322a53221560-cert") pod "openstack-baremetal-operator-controller-manager-6b68b8b854sclfc" (UID: "f43cc0a3-41e9-4e09-837f-322a53221560") : secret "openstack-baremetal-operator-webhook-server-cert" not found Jan 26 17:03:58 crc kubenswrapper[4754]: I0126 17:03:58.340010 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-77d5c5b54f-8sz77"] Jan 26 17:03:58 crc kubenswrapper[4754]: W0126 17:03:58.354779 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod28161331_7731_433f_845f_2ebe1daf5fd0.slice/crio-29002b51963410fc166ee51e3931071dd7a53379cb832073772961cd131524b1 WatchSource:0}: Error finding container 29002b51963410fc166ee51e3931071dd7a53379cb832073772961cd131524b1: Status 404 returned error can't find the container with id 29002b51963410fc166ee51e3931071dd7a53379cb832073772961cd131524b1 Jan 26 17:03:58 crc kubenswrapper[4754]: I0126 17:03:58.357417 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-598f7747c9-fvcgg"] Jan 26 17:03:58 crc kubenswrapper[4754]: I0126 17:03:58.365928 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7f86f8796f-5kbpl"] Jan 26 17:03:58 crc kubenswrapper[4754]: I0126 17:03:58.373563 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-594c8c9d5d-hgnm9"] Jan 26 17:03:58 crc kubenswrapper[4754]: I0126 17:03:58.396006 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-78fdd796fd-c5kf9"] Jan 26 17:03:58 crc kubenswrapper[4754]: I0126 17:03:58.401950 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-b45d7bf98-bggfb"] Jan 26 17:03:58 crc kubenswrapper[4754]: W0126 17:03:58.404097 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5e071de1_60dc_49d8_b965_90c2f99a6e02.slice/crio-4d935a3aaf55d9ac8cae4a6da40633d3946876e58c76bca45237c40fce3ca851 WatchSource:0}: Error finding container 4d935a3aaf55d9ac8cae4a6da40633d3946876e58c76bca45237c40fce3ca851: Status 404 returned error can't find the container with id 4d935a3aaf55d9ac8cae4a6da40633d3946876e58c76bca45237c40fce3ca851 Jan 26 17:03:58 crc kubenswrapper[4754]: I0126 17:03:58.466013 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-b45d7bf98-bggfb" event={"ID":"5e071de1-60dc-49d8-b965-90c2f99a6e02","Type":"ContainerStarted","Data":"4d935a3aaf55d9ac8cae4a6da40633d3946876e58c76bca45237c40fce3ca851"} Jan 26 17:03:58 crc kubenswrapper[4754]: I0126 17:03:58.468507 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7f86f8796f-5kbpl" event={"ID":"1a9a2d55-592f-4320-8e2e-49f65ca72dfc","Type":"ContainerStarted","Data":"8fafb18e3e5cd0c71df865fa641ee86def949b308944c0ad350b158c4cf56af0"} Jan 26 17:03:58 crc kubenswrapper[4754]: I0126 17:03:58.470613 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-78fdd796fd-c5kf9" event={"ID":"ee98073e-2fdb-4b3e-acb8-00c71df55fa7","Type":"ContainerStarted","Data":"970401dabd078b1d669f8fd4f48be60ffc8ba48921d2cb6db2968663019de8d5"} Jan 26 17:03:58 crc kubenswrapper[4754]: I0126 17:03:58.471409 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-7478f7dbf9-6vmhq" event={"ID":"cb2628e4-680e-489a-8fc9-d39986c74301","Type":"ContainerStarted","Data":"1f739bccd372aa705c37ca54a37a313b545d1e8d9a67129fc6acf9ce97fd4c76"} Jan 26 17:03:58 crc kubenswrapper[4754]: I0126 17:03:58.478621 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-77d5c5b54f-8sz77" event={"ID":"fd6b6ccf-b7e9-41fb-b663-a3392d075880","Type":"ContainerStarted","Data":"88eeb31a86fd910911b6e193ad009b93e4ec47e065fd5c61a300606ce39cc5d1"} Jan 26 17:03:58 crc kubenswrapper[4754]: I0126 17:03:58.479787 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-594c8c9d5d-hgnm9" event={"ID":"f747ae1d-8181-4ef2-b332-b14db483aab6","Type":"ContainerStarted","Data":"0a906de9a73ee4cda77c7315c2da25d8d1028725427c484f3bddd073aeca725f"} Jan 26 17:03:58 crc kubenswrapper[4754]: I0126 17:03:58.480997 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-598f7747c9-fvcgg" event={"ID":"28161331-7731-433f-845f-2ebe1daf5fd0","Type":"ContainerStarted","Data":"29002b51963410fc166ee51e3931071dd7a53379cb832073772961cd131524b1"} Jan 26 17:03:58 crc kubenswrapper[4754]: I0126 17:03:58.593055 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/93622668-0766-46f4-a216-83a7a17f36fc-metrics-certs\") pod \"openstack-operator-controller-manager-59fcd4bdb5-gvt4r\" (UID: \"93622668-0766-46f4-a216-83a7a17f36fc\") " pod="openstack-operators/openstack-operator-controller-manager-59fcd4bdb5-gvt4r" Jan 26 17:03:58 crc kubenswrapper[4754]: I0126 17:03:58.593111 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/93622668-0766-46f4-a216-83a7a17f36fc-webhook-certs\") pod \"openstack-operator-controller-manager-59fcd4bdb5-gvt4r\" (UID: \"93622668-0766-46f4-a216-83a7a17f36fc\") " pod="openstack-operators/openstack-operator-controller-manager-59fcd4bdb5-gvt4r" Jan 26 17:03:58 crc kubenswrapper[4754]: E0126 17:03:58.593309 4754 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Jan 26 17:03:58 crc kubenswrapper[4754]: E0126 17:03:58.593357 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/93622668-0766-46f4-a216-83a7a17f36fc-webhook-certs podName:93622668-0766-46f4-a216-83a7a17f36fc nodeName:}" failed. No retries permitted until 2026-01-26 17:03:59.593342439 +0000 UTC m=+1006.117522873 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/93622668-0766-46f4-a216-83a7a17f36fc-webhook-certs") pod "openstack-operator-controller-manager-59fcd4bdb5-gvt4r" (UID: "93622668-0766-46f4-a216-83a7a17f36fc") : secret "webhook-server-cert" not found Jan 26 17:03:58 crc kubenswrapper[4754]: E0126 17:03:58.593398 4754 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Jan 26 17:03:58 crc kubenswrapper[4754]: E0126 17:03:58.593416 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/93622668-0766-46f4-a216-83a7a17f36fc-metrics-certs podName:93622668-0766-46f4-a216-83a7a17f36fc nodeName:}" failed. No retries permitted until 2026-01-26 17:03:59.59341021 +0000 UTC m=+1006.117590644 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/93622668-0766-46f4-a216-83a7a17f36fc-metrics-certs") pod "openstack-operator-controller-manager-59fcd4bdb5-gvt4r" (UID: "93622668-0766-46f4-a216-83a7a17f36fc") : secret "metrics-server-cert" not found Jan 26 17:03:58 crc kubenswrapper[4754]: I0126 17:03:58.735487 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-6f75f45d54-kppds"] Jan 26 17:03:58 crc kubenswrapper[4754]: I0126 17:03:58.743551 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-5f4cd88d46-fkz84"] Jan 26 17:03:58 crc kubenswrapper[4754]: W0126 17:03:58.745568 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4fdc4d46_264a_4689_8ad5_0ed253f805df.slice/crio-520c814741c4adc3a99d2bcdaa4e4dca4f184ac2fe0b017140b981617b959f8c WatchSource:0}: Error finding container 520c814741c4adc3a99d2bcdaa4e4dca4f184ac2fe0b017140b981617b959f8c: Status 404 returned error can't find the container with id 520c814741c4adc3a99d2bcdaa4e4dca4f184ac2fe0b017140b981617b959f8c Jan 26 17:03:58 crc kubenswrapper[4754]: I0126 17:03:58.749445 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-85cd9769bb-dq4qw"] Jan 26 17:03:58 crc kubenswrapper[4754]: W0126 17:03:58.762601 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod13391684_794c_48b2_8d7f_23a122f8acc6.slice/crio-67d248bd07abea1cc3298acef343c517394f4395cd26cfa5fee5981beb825da1 WatchSource:0}: Error finding container 67d248bd07abea1cc3298acef343c517394f4395cd26cfa5fee5981beb825da1: Status 404 returned error can't find the container with id 67d248bd07abea1cc3298acef343c517394f4395cd26cfa5fee5981beb825da1 Jan 26 17:03:58 crc kubenswrapper[4754]: I0126 17:03:58.764505 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-6b9fb5fdcb-fhwr4"] Jan 26 17:03:58 crc kubenswrapper[4754]: I0126 17:03:58.774778 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-7bdb645866-27hg7"] Jan 26 17:03:58 crc kubenswrapper[4754]: I0126 17:03:58.780456 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-78c6999f6f-jb5ps"] Jan 26 17:03:58 crc kubenswrapper[4754]: I0126 17:03:58.786780 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-78d58447c5-pgmv6"] Jan 26 17:03:58 crc kubenswrapper[4754]: I0126 17:03:58.795977 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5702a910-71d1-4acd-93aa-9379bc3147ce-cert\") pod \"infra-operator-controller-manager-694cf4f878-p66zm\" (UID: \"5702a910-71d1-4acd-93aa-9379bc3147ce\") " pod="openstack-operators/infra-operator-controller-manager-694cf4f878-p66zm" Jan 26 17:03:58 crc kubenswrapper[4754]: E0126 17:03:58.796307 4754 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Jan 26 17:03:58 crc kubenswrapper[4754]: E0126 17:03:58.796378 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5702a910-71d1-4acd-93aa-9379bc3147ce-cert podName:5702a910-71d1-4acd-93aa-9379bc3147ce nodeName:}" failed. No retries permitted until 2026-01-26 17:04:00.796360933 +0000 UTC m=+1007.320541357 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/5702a910-71d1-4acd-93aa-9379bc3147ce-cert") pod "infra-operator-controller-manager-694cf4f878-p66zm" (UID: "5702a910-71d1-4acd-93aa-9379bc3147ce") : secret "infra-operator-webhook-server-cert" not found Jan 26 17:03:58 crc kubenswrapper[4754]: I0126 17:03:58.805038 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-b8b6d4659-d8db9"] Jan 26 17:03:58 crc kubenswrapper[4754]: E0126 17:03:58.807499 4754 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/neutron-operator@sha256:816d474f502d730d6a2522a272b0e09a2d579ac63617817655d60c54bda4191e,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-l2x9c,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod neutron-operator-controller-manager-78d58447c5-pgmv6_openstack-operators(2ab80530-a54c-41bf-8e4f-c895dbacb368): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Jan 26 17:03:58 crc kubenswrapper[4754]: E0126 17:03:58.809169 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/neutron-operator-controller-manager-78d58447c5-pgmv6" podUID="2ab80530-a54c-41bf-8e4f-c895dbacb368" Jan 26 17:03:58 crc kubenswrapper[4754]: I0126 17:03:58.809826 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-79d5ccc684-gqtkb"] Jan 26 17:03:58 crc kubenswrapper[4754]: W0126 17:03:58.824772 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod56addd62_aeed_4139_a0db_37292a6acf8a.slice/crio-a432582586face83f575c074d8bfe1638872fe92ceb37bc54651c3b47bd8e30c WatchSource:0}: Error finding container a432582586face83f575c074d8bfe1638872fe92ceb37bc54651c3b47bd8e30c: Status 404 returned error can't find the container with id a432582586face83f575c074d8bfe1638872fe92ceb37bc54651c3b47bd8e30c Jan 26 17:03:58 crc kubenswrapper[4754]: E0126 17:03:58.828115 4754 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/placement-operator@sha256:013c0ad82d21a21c7eece5cd4b5d5c4b8eb410b6671ac33a6f3fb78c8510811d,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-rtgsw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-79d5ccc684-gqtkb_openstack-operators(56addd62-aeed-4139-a0db-37292a6acf8a): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Jan 26 17:03:58 crc kubenswrapper[4754]: E0126 17:03:58.829448 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/placement-operator-controller-manager-79d5ccc684-gqtkb" podUID="56addd62-aeed-4139-a0db-37292a6acf8a" Jan 26 17:03:58 crc kubenswrapper[4754]: W0126 17:03:58.834228 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode3aac642_b3ee_4394_9f1d_bfac315bf162.slice/crio-a280f574d0e4b4f985e74423927fc27f7807bfb15cfafc121d3e781c7789f49b WatchSource:0}: Error finding container a280f574d0e4b4f985e74423927fc27f7807bfb15cfafc121d3e781c7789f49b: Status 404 returned error can't find the container with id a280f574d0e4b4f985e74423927fc27f7807bfb15cfafc121d3e781c7789f49b Jan 26 17:03:58 crc kubenswrapper[4754]: E0126 17:03:58.838139 4754 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/keystone-operator@sha256:8e340ff11922b38e811261de96982e1aff5f4eb8f225d1d9f5973025a4fe8349,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-wnr8c,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod keystone-operator-controller-manager-b8b6d4659-d8db9_openstack-operators(e3aac642-b3ee-4394-9f1d-bfac315bf162): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Jan 26 17:03:58 crc kubenswrapper[4754]: E0126 17:03:58.839327 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/keystone-operator-controller-manager-b8b6d4659-d8db9" podUID="e3aac642-b3ee-4394-9f1d-bfac315bf162" Jan 26 17:03:58 crc kubenswrapper[4754]: I0126 17:03:58.852555 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-547cbdb99f-fmvdv"] Jan 26 17:03:58 crc kubenswrapper[4754]: I0126 17:03:58.861484 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-564965969-k7nbl"] Jan 26 17:03:58 crc kubenswrapper[4754]: I0126 17:03:58.868065 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-mqmkw"] Jan 26 17:03:58 crc kubenswrapper[4754]: E0126 17:03:58.871614 4754 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/swift-operator@sha256:445e951df2f21df6d33a466f75917e0f6103052ae751ae11887136e8ab165922,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-57rp9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-547cbdb99f-fmvdv_openstack-operators(aaee86fe-018f-4055-b8c5-98e3795c53d8): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Jan 26 17:03:58 crc kubenswrapper[4754]: E0126 17:03:58.872130 4754 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/watcher-operator@sha256:7869203f6f97de780368d507636031090fed3b658d2f7771acbd4481bdfc870b,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-95qgc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-564965969-k7nbl_openstack-operators(138cc11d-6be6-43bb-994e-94ea39bb2e42): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Jan 26 17:03:58 crc kubenswrapper[4754]: E0126 17:03:58.872818 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/swift-operator-controller-manager-547cbdb99f-fmvdv" podUID="aaee86fe-018f-4055-b8c5-98e3795c53d8" Jan 26 17:03:58 crc kubenswrapper[4754]: E0126 17:03:58.873212 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/watcher-operator-controller-manager-564965969-k7nbl" podUID="138cc11d-6be6-43bb-994e-94ea39bb2e42" Jan 26 17:03:58 crc kubenswrapper[4754]: I0126 17:03:58.876406 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-69797bbcbd-mr8bp"] Jan 26 17:03:58 crc kubenswrapper[4754]: W0126 17:03:58.881476 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda8584d32_796d_47cd_8dd6_233374660688.slice/crio-01a91643bc96978df5803b6d087f2b050721c3343bead7873bbe663a16534760 WatchSource:0}: Error finding container 01a91643bc96978df5803b6d087f2b050721c3343bead7873bbe663a16534760: Status 404 returned error can't find the container with id 01a91643bc96978df5803b6d087f2b050721c3343bead7873bbe663a16534760 Jan 26 17:03:58 crc kubenswrapper[4754]: E0126 17:03:58.884110 4754 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:c8dde42dafd41026ed2e4cfc26efc0fff63c4ba9d31326ae7dc644ccceaafa9d,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-64c6f,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-69797bbcbd-mr8bp_openstack-operators(a8584d32-796d-47cd-8dd6-233374660688): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Jan 26 17:03:58 crc kubenswrapper[4754]: W0126 17:03:58.884995 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb48b5414_4034_4c0e_9d25_5053f84cf246.slice/crio-7276a7ad81cd6b30a43608984beb7002d7f4422eb140c73a1d7ed7b6f6cef250 WatchSource:0}: Error finding container 7276a7ad81cd6b30a43608984beb7002d7f4422eb140c73a1d7ed7b6f6cef250: Status 404 returned error can't find the container with id 7276a7ad81cd6b30a43608984beb7002d7f4422eb140c73a1d7ed7b6f6cef250 Jan 26 17:03:58 crc kubenswrapper[4754]: E0126 17:03:58.886110 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/test-operator-controller-manager-69797bbcbd-mr8bp" podUID="a8584d32-796d-47cd-8dd6-233374660688" Jan 26 17:03:58 crc kubenswrapper[4754]: E0126 17:03:58.887595 4754 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-5mb2z,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-668c99d594-mqmkw_openstack-operators(b48b5414-4034-4c0e-9d25-5053f84cf246): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Jan 26 17:03:58 crc kubenswrapper[4754]: E0126 17:03:58.888966 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-mqmkw" podUID="b48b5414-4034-4c0e-9d25-5053f84cf246" Jan 26 17:03:59 crc kubenswrapper[4754]: I0126 17:03:59.307870 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f43cc0a3-41e9-4e09-837f-322a53221560-cert\") pod \"openstack-baremetal-operator-controller-manager-6b68b8b854sclfc\" (UID: \"f43cc0a3-41e9-4e09-837f-322a53221560\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6b68b8b854sclfc" Jan 26 17:03:59 crc kubenswrapper[4754]: E0126 17:03:59.308023 4754 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Jan 26 17:03:59 crc kubenswrapper[4754]: E0126 17:03:59.308099 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f43cc0a3-41e9-4e09-837f-322a53221560-cert podName:f43cc0a3-41e9-4e09-837f-322a53221560 nodeName:}" failed. No retries permitted until 2026-01-26 17:04:01.308081772 +0000 UTC m=+1007.832262196 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/f43cc0a3-41e9-4e09-837f-322a53221560-cert") pod "openstack-baremetal-operator-controller-manager-6b68b8b854sclfc" (UID: "f43cc0a3-41e9-4e09-837f-322a53221560") : secret "openstack-baremetal-operator-webhook-server-cert" not found Jan 26 17:03:59 crc kubenswrapper[4754]: I0126 17:03:59.493185 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-6f75f45d54-kppds" event={"ID":"8353f62d-eb53-46c3-ba4d-2d643f4f960b","Type":"ContainerStarted","Data":"243ad6a95d9ea3556ceabeaae4ff60d8a455b951960bf88e1d2e29c03fdd74f4"} Jan 26 17:03:59 crc kubenswrapper[4754]: I0126 17:03:59.495416 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-b8b6d4659-d8db9" event={"ID":"e3aac642-b3ee-4394-9f1d-bfac315bf162","Type":"ContainerStarted","Data":"a280f574d0e4b4f985e74423927fc27f7807bfb15cfafc121d3e781c7789f49b"} Jan 26 17:03:59 crc kubenswrapper[4754]: E0126 17:03:59.507009 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/keystone-operator@sha256:8e340ff11922b38e811261de96982e1aff5f4eb8f225d1d9f5973025a4fe8349\\\"\"" pod="openstack-operators/keystone-operator-controller-manager-b8b6d4659-d8db9" podUID="e3aac642-b3ee-4394-9f1d-bfac315bf162" Jan 26 17:03:59 crc kubenswrapper[4754]: I0126 17:03:59.507703 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-85cd9769bb-dq4qw" event={"ID":"13391684-794c-48b2-8d7f-23a122f8acc6","Type":"ContainerStarted","Data":"67d248bd07abea1cc3298acef343c517394f4395cd26cfa5fee5981beb825da1"} Jan 26 17:03:59 crc kubenswrapper[4754]: I0126 17:03:59.511626 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-79d5ccc684-gqtkb" event={"ID":"56addd62-aeed-4139-a0db-37292a6acf8a","Type":"ContainerStarted","Data":"a432582586face83f575c074d8bfe1638872fe92ceb37bc54651c3b47bd8e30c"} Jan 26 17:03:59 crc kubenswrapper[4754]: E0126 17:03:59.512623 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:013c0ad82d21a21c7eece5cd4b5d5c4b8eb410b6671ac33a6f3fb78c8510811d\\\"\"" pod="openstack-operators/placement-operator-controller-manager-79d5ccc684-gqtkb" podUID="56addd62-aeed-4139-a0db-37292a6acf8a" Jan 26 17:03:59 crc kubenswrapper[4754]: I0126 17:03:59.515825 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-69797bbcbd-mr8bp" event={"ID":"a8584d32-796d-47cd-8dd6-233374660688","Type":"ContainerStarted","Data":"01a91643bc96978df5803b6d087f2b050721c3343bead7873bbe663a16534760"} Jan 26 17:03:59 crc kubenswrapper[4754]: E0126 17:03:59.516747 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:c8dde42dafd41026ed2e4cfc26efc0fff63c4ba9d31326ae7dc644ccceaafa9d\\\"\"" pod="openstack-operators/test-operator-controller-manager-69797bbcbd-mr8bp" podUID="a8584d32-796d-47cd-8dd6-233374660688" Jan 26 17:03:59 crc kubenswrapper[4754]: I0126 17:03:59.517432 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-547cbdb99f-fmvdv" event={"ID":"aaee86fe-018f-4055-b8c5-98e3795c53d8","Type":"ContainerStarted","Data":"9de4f0a357730f8d12a6dac74b703f436af19c0d3ab02e3ef0731f03a508345f"} Jan 26 17:03:59 crc kubenswrapper[4754]: E0126 17:03:59.518532 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:445e951df2f21df6d33a466f75917e0f6103052ae751ae11887136e8ab165922\\\"\"" pod="openstack-operators/swift-operator-controller-manager-547cbdb99f-fmvdv" podUID="aaee86fe-018f-4055-b8c5-98e3795c53d8" Jan 26 17:03:59 crc kubenswrapper[4754]: I0126 17:03:59.519619 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-5f4cd88d46-fkz84" event={"ID":"4fdc4d46-264a-4689-8ad5-0ed253f805df","Type":"ContainerStarted","Data":"520c814741c4adc3a99d2bcdaa4e4dca4f184ac2fe0b017140b981617b959f8c"} Jan 26 17:03:59 crc kubenswrapper[4754]: I0126 17:03:59.521019 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-564965969-k7nbl" event={"ID":"138cc11d-6be6-43bb-994e-94ea39bb2e42","Type":"ContainerStarted","Data":"74f03b75211c581351fecaf2fa03e53c0aa67ece4c194b0b635fa9cff9e93ab2"} Jan 26 17:03:59 crc kubenswrapper[4754]: E0126 17:03:59.522889 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:7869203f6f97de780368d507636031090fed3b658d2f7771acbd4481bdfc870b\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-564965969-k7nbl" podUID="138cc11d-6be6-43bb-994e-94ea39bb2e42" Jan 26 17:03:59 crc kubenswrapper[4754]: I0126 17:03:59.533069 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-78d58447c5-pgmv6" event={"ID":"2ab80530-a54c-41bf-8e4f-c895dbacb368","Type":"ContainerStarted","Data":"14edf531e064857890be71e14b28db1a399cdc1689b9828d5391dea6c500a7f8"} Jan 26 17:03:59 crc kubenswrapper[4754]: E0126 17:03:59.534983 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/neutron-operator@sha256:816d474f502d730d6a2522a272b0e09a2d579ac63617817655d60c54bda4191e\\\"\"" pod="openstack-operators/neutron-operator-controller-manager-78d58447c5-pgmv6" podUID="2ab80530-a54c-41bf-8e4f-c895dbacb368" Jan 26 17:03:59 crc kubenswrapper[4754]: I0126 17:03:59.537036 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-mqmkw" event={"ID":"b48b5414-4034-4c0e-9d25-5053f84cf246","Type":"ContainerStarted","Data":"7276a7ad81cd6b30a43608984beb7002d7f4422eb140c73a1d7ed7b6f6cef250"} Jan 26 17:03:59 crc kubenswrapper[4754]: E0126 17:03:59.538650 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-mqmkw" podUID="b48b5414-4034-4c0e-9d25-5053f84cf246" Jan 26 17:03:59 crc kubenswrapper[4754]: I0126 17:03:59.539615 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-78c6999f6f-jb5ps" event={"ID":"16e535a9-3fc0-4385-b809-51a2bf719657","Type":"ContainerStarted","Data":"54a6c263c6ea9bdd75907c13c8329f739f6852802d59405daea185bf2cb3e899"} Jan 26 17:03:59 crc kubenswrapper[4754]: I0126 17:03:59.543616 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-7bdb645866-27hg7" event={"ID":"d23565a1-cb57-4cbe-88a2-48a01e8056f3","Type":"ContainerStarted","Data":"f931d0b02ddf94a7c6b4d5ddf69cfd6102c4f22a608a090d922a7445bb99c179"} Jan 26 17:03:59 crc kubenswrapper[4754]: I0126 17:03:59.548284 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-6b9fb5fdcb-fhwr4" event={"ID":"add7fe6b-a864-439d-a3f4-80c9cc80ddc1","Type":"ContainerStarted","Data":"6049bea16f7e74c40e1428bfc1fab31a33e5b31e892f1c77cc84ea4f6afd501f"} Jan 26 17:03:59 crc kubenswrapper[4754]: I0126 17:03:59.615392 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/93622668-0766-46f4-a216-83a7a17f36fc-metrics-certs\") pod \"openstack-operator-controller-manager-59fcd4bdb5-gvt4r\" (UID: \"93622668-0766-46f4-a216-83a7a17f36fc\") " pod="openstack-operators/openstack-operator-controller-manager-59fcd4bdb5-gvt4r" Jan 26 17:03:59 crc kubenswrapper[4754]: I0126 17:03:59.615468 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/93622668-0766-46f4-a216-83a7a17f36fc-webhook-certs\") pod \"openstack-operator-controller-manager-59fcd4bdb5-gvt4r\" (UID: \"93622668-0766-46f4-a216-83a7a17f36fc\") " pod="openstack-operators/openstack-operator-controller-manager-59fcd4bdb5-gvt4r" Jan 26 17:03:59 crc kubenswrapper[4754]: E0126 17:03:59.615689 4754 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Jan 26 17:03:59 crc kubenswrapper[4754]: E0126 17:03:59.615740 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/93622668-0766-46f4-a216-83a7a17f36fc-webhook-certs podName:93622668-0766-46f4-a216-83a7a17f36fc nodeName:}" failed. No retries permitted until 2026-01-26 17:04:01.615725426 +0000 UTC m=+1008.139905860 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/93622668-0766-46f4-a216-83a7a17f36fc-webhook-certs") pod "openstack-operator-controller-manager-59fcd4bdb5-gvt4r" (UID: "93622668-0766-46f4-a216-83a7a17f36fc") : secret "webhook-server-cert" not found Jan 26 17:03:59 crc kubenswrapper[4754]: E0126 17:03:59.616188 4754 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Jan 26 17:03:59 crc kubenswrapper[4754]: E0126 17:03:59.616223 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/93622668-0766-46f4-a216-83a7a17f36fc-metrics-certs podName:93622668-0766-46f4-a216-83a7a17f36fc nodeName:}" failed. No retries permitted until 2026-01-26 17:04:01.61621545 +0000 UTC m=+1008.140395884 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/93622668-0766-46f4-a216-83a7a17f36fc-metrics-certs") pod "openstack-operator-controller-manager-59fcd4bdb5-gvt4r" (UID: "93622668-0766-46f4-a216-83a7a17f36fc") : secret "metrics-server-cert" not found Jan 26 17:04:00 crc kubenswrapper[4754]: E0126 17:04:00.555070 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:013c0ad82d21a21c7eece5cd4b5d5c4b8eb410b6671ac33a6f3fb78c8510811d\\\"\"" pod="openstack-operators/placement-operator-controller-manager-79d5ccc684-gqtkb" podUID="56addd62-aeed-4139-a0db-37292a6acf8a" Jan 26 17:04:00 crc kubenswrapper[4754]: E0126 17:04:00.555265 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:445e951df2f21df6d33a466f75917e0f6103052ae751ae11887136e8ab165922\\\"\"" pod="openstack-operators/swift-operator-controller-manager-547cbdb99f-fmvdv" podUID="aaee86fe-018f-4055-b8c5-98e3795c53d8" Jan 26 17:04:00 crc kubenswrapper[4754]: E0126 17:04:00.555377 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:c8dde42dafd41026ed2e4cfc26efc0fff63c4ba9d31326ae7dc644ccceaafa9d\\\"\"" pod="openstack-operators/test-operator-controller-manager-69797bbcbd-mr8bp" podUID="a8584d32-796d-47cd-8dd6-233374660688" Jan 26 17:04:00 crc kubenswrapper[4754]: E0126 17:04:00.555459 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/keystone-operator@sha256:8e340ff11922b38e811261de96982e1aff5f4eb8f225d1d9f5973025a4fe8349\\\"\"" pod="openstack-operators/keystone-operator-controller-manager-b8b6d4659-d8db9" podUID="e3aac642-b3ee-4394-9f1d-bfac315bf162" Jan 26 17:04:00 crc kubenswrapper[4754]: E0126 17:04:00.555530 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/neutron-operator@sha256:816d474f502d730d6a2522a272b0e09a2d579ac63617817655d60c54bda4191e\\\"\"" pod="openstack-operators/neutron-operator-controller-manager-78d58447c5-pgmv6" podUID="2ab80530-a54c-41bf-8e4f-c895dbacb368" Jan 26 17:04:00 crc kubenswrapper[4754]: E0126 17:04:00.555624 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-mqmkw" podUID="b48b5414-4034-4c0e-9d25-5053f84cf246" Jan 26 17:04:00 crc kubenswrapper[4754]: E0126 17:04:00.555627 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:7869203f6f97de780368d507636031090fed3b658d2f7771acbd4481bdfc870b\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-564965969-k7nbl" podUID="138cc11d-6be6-43bb-994e-94ea39bb2e42" Jan 26 17:04:00 crc kubenswrapper[4754]: I0126 17:04:00.838566 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5702a910-71d1-4acd-93aa-9379bc3147ce-cert\") pod \"infra-operator-controller-manager-694cf4f878-p66zm\" (UID: \"5702a910-71d1-4acd-93aa-9379bc3147ce\") " pod="openstack-operators/infra-operator-controller-manager-694cf4f878-p66zm" Jan 26 17:04:00 crc kubenswrapper[4754]: E0126 17:04:00.838713 4754 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Jan 26 17:04:00 crc kubenswrapper[4754]: E0126 17:04:00.838760 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5702a910-71d1-4acd-93aa-9379bc3147ce-cert podName:5702a910-71d1-4acd-93aa-9379bc3147ce nodeName:}" failed. No retries permitted until 2026-01-26 17:04:04.838746712 +0000 UTC m=+1011.362927146 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/5702a910-71d1-4acd-93aa-9379bc3147ce-cert") pod "infra-operator-controller-manager-694cf4f878-p66zm" (UID: "5702a910-71d1-4acd-93aa-9379bc3147ce") : secret "infra-operator-webhook-server-cert" not found Jan 26 17:04:01 crc kubenswrapper[4754]: I0126 17:04:01.344296 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f43cc0a3-41e9-4e09-837f-322a53221560-cert\") pod \"openstack-baremetal-operator-controller-manager-6b68b8b854sclfc\" (UID: \"f43cc0a3-41e9-4e09-837f-322a53221560\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6b68b8b854sclfc" Jan 26 17:04:01 crc kubenswrapper[4754]: E0126 17:04:01.344513 4754 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Jan 26 17:04:01 crc kubenswrapper[4754]: E0126 17:04:01.344861 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f43cc0a3-41e9-4e09-837f-322a53221560-cert podName:f43cc0a3-41e9-4e09-837f-322a53221560 nodeName:}" failed. No retries permitted until 2026-01-26 17:04:05.344828912 +0000 UTC m=+1011.869009376 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/f43cc0a3-41e9-4e09-837f-322a53221560-cert") pod "openstack-baremetal-operator-controller-manager-6b68b8b854sclfc" (UID: "f43cc0a3-41e9-4e09-837f-322a53221560") : secret "openstack-baremetal-operator-webhook-server-cert" not found Jan 26 17:04:01 crc kubenswrapper[4754]: I0126 17:04:01.649364 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/93622668-0766-46f4-a216-83a7a17f36fc-webhook-certs\") pod \"openstack-operator-controller-manager-59fcd4bdb5-gvt4r\" (UID: \"93622668-0766-46f4-a216-83a7a17f36fc\") " pod="openstack-operators/openstack-operator-controller-manager-59fcd4bdb5-gvt4r" Jan 26 17:04:01 crc kubenswrapper[4754]: E0126 17:04:01.649630 4754 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Jan 26 17:04:01 crc kubenswrapper[4754]: E0126 17:04:01.650553 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/93622668-0766-46f4-a216-83a7a17f36fc-webhook-certs podName:93622668-0766-46f4-a216-83a7a17f36fc nodeName:}" failed. No retries permitted until 2026-01-26 17:04:05.650516981 +0000 UTC m=+1012.174697465 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/93622668-0766-46f4-a216-83a7a17f36fc-webhook-certs") pod "openstack-operator-controller-manager-59fcd4bdb5-gvt4r" (UID: "93622668-0766-46f4-a216-83a7a17f36fc") : secret "webhook-server-cert" not found Jan 26 17:04:01 crc kubenswrapper[4754]: I0126 17:04:01.650782 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/93622668-0766-46f4-a216-83a7a17f36fc-metrics-certs\") pod \"openstack-operator-controller-manager-59fcd4bdb5-gvt4r\" (UID: \"93622668-0766-46f4-a216-83a7a17f36fc\") " pod="openstack-operators/openstack-operator-controller-manager-59fcd4bdb5-gvt4r" Jan 26 17:04:01 crc kubenswrapper[4754]: E0126 17:04:01.651092 4754 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Jan 26 17:04:01 crc kubenswrapper[4754]: E0126 17:04:01.651269 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/93622668-0766-46f4-a216-83a7a17f36fc-metrics-certs podName:93622668-0766-46f4-a216-83a7a17f36fc nodeName:}" failed. No retries permitted until 2026-01-26 17:04:05.651242242 +0000 UTC m=+1012.175422746 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/93622668-0766-46f4-a216-83a7a17f36fc-metrics-certs") pod "openstack-operator-controller-manager-59fcd4bdb5-gvt4r" (UID: "93622668-0766-46f4-a216-83a7a17f36fc") : secret "metrics-server-cert" not found Jan 26 17:04:04 crc kubenswrapper[4754]: I0126 17:04:04.901442 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5702a910-71d1-4acd-93aa-9379bc3147ce-cert\") pod \"infra-operator-controller-manager-694cf4f878-p66zm\" (UID: \"5702a910-71d1-4acd-93aa-9379bc3147ce\") " pod="openstack-operators/infra-operator-controller-manager-694cf4f878-p66zm" Jan 26 17:04:04 crc kubenswrapper[4754]: E0126 17:04:04.901941 4754 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Jan 26 17:04:04 crc kubenswrapper[4754]: E0126 17:04:04.901999 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5702a910-71d1-4acd-93aa-9379bc3147ce-cert podName:5702a910-71d1-4acd-93aa-9379bc3147ce nodeName:}" failed. No retries permitted until 2026-01-26 17:04:12.901981262 +0000 UTC m=+1019.426161696 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/5702a910-71d1-4acd-93aa-9379bc3147ce-cert") pod "infra-operator-controller-manager-694cf4f878-p66zm" (UID: "5702a910-71d1-4acd-93aa-9379bc3147ce") : secret "infra-operator-webhook-server-cert" not found Jan 26 17:04:05 crc kubenswrapper[4754]: I0126 17:04:05.408158 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f43cc0a3-41e9-4e09-837f-322a53221560-cert\") pod \"openstack-baremetal-operator-controller-manager-6b68b8b854sclfc\" (UID: \"f43cc0a3-41e9-4e09-837f-322a53221560\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6b68b8b854sclfc" Jan 26 17:04:05 crc kubenswrapper[4754]: E0126 17:04:05.408326 4754 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Jan 26 17:04:05 crc kubenswrapper[4754]: E0126 17:04:05.408390 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f43cc0a3-41e9-4e09-837f-322a53221560-cert podName:f43cc0a3-41e9-4e09-837f-322a53221560 nodeName:}" failed. No retries permitted until 2026-01-26 17:04:13.408372901 +0000 UTC m=+1019.932553335 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/f43cc0a3-41e9-4e09-837f-322a53221560-cert") pod "openstack-baremetal-operator-controller-manager-6b68b8b854sclfc" (UID: "f43cc0a3-41e9-4e09-837f-322a53221560") : secret "openstack-baremetal-operator-webhook-server-cert" not found Jan 26 17:04:05 crc kubenswrapper[4754]: I0126 17:04:05.712837 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/93622668-0766-46f4-a216-83a7a17f36fc-metrics-certs\") pod \"openstack-operator-controller-manager-59fcd4bdb5-gvt4r\" (UID: \"93622668-0766-46f4-a216-83a7a17f36fc\") " pod="openstack-operators/openstack-operator-controller-manager-59fcd4bdb5-gvt4r" Jan 26 17:04:05 crc kubenswrapper[4754]: I0126 17:04:05.712903 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/93622668-0766-46f4-a216-83a7a17f36fc-webhook-certs\") pod \"openstack-operator-controller-manager-59fcd4bdb5-gvt4r\" (UID: \"93622668-0766-46f4-a216-83a7a17f36fc\") " pod="openstack-operators/openstack-operator-controller-manager-59fcd4bdb5-gvt4r" Jan 26 17:04:05 crc kubenswrapper[4754]: E0126 17:04:05.713061 4754 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Jan 26 17:04:05 crc kubenswrapper[4754]: E0126 17:04:05.713122 4754 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Jan 26 17:04:05 crc kubenswrapper[4754]: E0126 17:04:05.713165 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/93622668-0766-46f4-a216-83a7a17f36fc-metrics-certs podName:93622668-0766-46f4-a216-83a7a17f36fc nodeName:}" failed. No retries permitted until 2026-01-26 17:04:13.713144005 +0000 UTC m=+1020.237324459 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/93622668-0766-46f4-a216-83a7a17f36fc-metrics-certs") pod "openstack-operator-controller-manager-59fcd4bdb5-gvt4r" (UID: "93622668-0766-46f4-a216-83a7a17f36fc") : secret "metrics-server-cert" not found Jan 26 17:04:05 crc kubenswrapper[4754]: E0126 17:04:05.713202 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/93622668-0766-46f4-a216-83a7a17f36fc-webhook-certs podName:93622668-0766-46f4-a216-83a7a17f36fc nodeName:}" failed. No retries permitted until 2026-01-26 17:04:13.713179826 +0000 UTC m=+1020.237360330 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/93622668-0766-46f4-a216-83a7a17f36fc-webhook-certs") pod "openstack-operator-controller-manager-59fcd4bdb5-gvt4r" (UID: "93622668-0766-46f4-a216-83a7a17f36fc") : secret "webhook-server-cert" not found Jan 26 17:04:10 crc kubenswrapper[4754]: E0126 17:04:10.251850 4754 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/horizon-operator@sha256:3311e627bcb860d9443592a2c67078417318c9eb77d8ef4d07f9aa7027d46822" Jan 26 17:04:10 crc kubenswrapper[4754]: E0126 17:04:10.252616 4754 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/horizon-operator@sha256:3311e627bcb860d9443592a2c67078417318c9eb77d8ef4d07f9aa7027d46822,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-45qhs,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-operator-controller-manager-77d5c5b54f-8sz77_openstack-operators(fd6b6ccf-b7e9-41fb-b663-a3392d075880): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Jan 26 17:04:10 crc kubenswrapper[4754]: E0126 17:04:10.253881 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/horizon-operator-controller-manager-77d5c5b54f-8sz77" podUID="fd6b6ccf-b7e9-41fb-b663-a3392d075880" Jan 26 17:04:10 crc kubenswrapper[4754]: E0126 17:04:10.623562 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/horizon-operator@sha256:3311e627bcb860d9443592a2c67078417318c9eb77d8ef4d07f9aa7027d46822\\\"\"" pod="openstack-operators/horizon-operator-controller-manager-77d5c5b54f-8sz77" podUID="fd6b6ccf-b7e9-41fb-b663-a3392d075880" Jan 26 17:04:12 crc kubenswrapper[4754]: I0126 17:04:12.639704 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-78c6999f6f-jb5ps" event={"ID":"16e535a9-3fc0-4385-b809-51a2bf719657","Type":"ContainerStarted","Data":"22dfa4b5744ddac272c2127b706fe2e677f16b56aef51c981f82cf7df9b312c4"} Jan 26 17:04:12 crc kubenswrapper[4754]: I0126 17:04:12.640283 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-78c6999f6f-jb5ps" Jan 26 17:04:12 crc kubenswrapper[4754]: I0126 17:04:12.644828 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7f86f8796f-5kbpl" event={"ID":"1a9a2d55-592f-4320-8e2e-49f65ca72dfc","Type":"ContainerStarted","Data":"0fcf654b986d70ca2a278052c356f5a2b5381671efb71a58fb8fe5d53da76d9b"} Jan 26 17:04:12 crc kubenswrapper[4754]: I0126 17:04:12.645523 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-7f86f8796f-5kbpl" Jan 26 17:04:12 crc kubenswrapper[4754]: I0126 17:04:12.649174 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-7bdb645866-27hg7" event={"ID":"d23565a1-cb57-4cbe-88a2-48a01e8056f3","Type":"ContainerStarted","Data":"4b2341fe6aa6fc00e516c10c18e9986cae21755fd5383bcc17a490ee08f0108a"} Jan 26 17:04:12 crc kubenswrapper[4754]: I0126 17:04:12.649651 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-7bdb645866-27hg7" Jan 26 17:04:12 crc kubenswrapper[4754]: I0126 17:04:12.654149 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-78c6999f6f-jb5ps" podStartSLOduration=2.214548595 podStartE2EDuration="15.654136201s" podCreationTimestamp="2026-01-26 17:03:57 +0000 UTC" firstStartedPulling="2026-01-26 17:03:58.791549067 +0000 UTC m=+1005.315729501" lastFinishedPulling="2026-01-26 17:04:12.231136673 +0000 UTC m=+1018.755317107" observedRunningTime="2026-01-26 17:04:12.654046118 +0000 UTC m=+1019.178226562" watchObservedRunningTime="2026-01-26 17:04:12.654136201 +0000 UTC m=+1019.178316635" Jan 26 17:04:12 crc kubenswrapper[4754]: I0126 17:04:12.659503 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-598f7747c9-fvcgg" Jan 26 17:04:12 crc kubenswrapper[4754]: I0126 17:04:12.669738 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-85cd9769bb-dq4qw" event={"ID":"13391684-794c-48b2-8d7f-23a122f8acc6","Type":"ContainerStarted","Data":"0505ce44dd1e42c21600f5d3cdbb71b31e2c2576684862b7b7d217e7e5106562"} Jan 26 17:04:12 crc kubenswrapper[4754]: I0126 17:04:12.670460 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-85cd9769bb-dq4qw" Jan 26 17:04:12 crc kubenswrapper[4754]: I0126 17:04:12.704038 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-7bdb645866-27hg7" podStartSLOduration=2.201715914 podStartE2EDuration="15.704019978s" podCreationTimestamp="2026-01-26 17:03:57 +0000 UTC" firstStartedPulling="2026-01-26 17:03:58.789496669 +0000 UTC m=+1005.313677103" lastFinishedPulling="2026-01-26 17:04:12.291800743 +0000 UTC m=+1018.815981167" observedRunningTime="2026-01-26 17:04:12.701788374 +0000 UTC m=+1019.225968828" watchObservedRunningTime="2026-01-26 17:04:12.704019978 +0000 UTC m=+1019.228200412" Jan 26 17:04:12 crc kubenswrapper[4754]: I0126 17:04:12.705180 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-7f86f8796f-5kbpl" podStartSLOduration=2.878000434 podStartE2EDuration="16.70517274s" podCreationTimestamp="2026-01-26 17:03:56 +0000 UTC" firstStartedPulling="2026-01-26 17:03:58.382525113 +0000 UTC m=+1004.906705547" lastFinishedPulling="2026-01-26 17:04:12.209697419 +0000 UTC m=+1018.733877853" observedRunningTime="2026-01-26 17:04:12.679716622 +0000 UTC m=+1019.203897056" watchObservedRunningTime="2026-01-26 17:04:12.70517274 +0000 UTC m=+1019.229353174" Jan 26 17:04:12 crc kubenswrapper[4754]: I0126 17:04:12.728367 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-85cd9769bb-dq4qw" podStartSLOduration=2.285520506 podStartE2EDuration="15.728349263s" podCreationTimestamp="2026-01-26 17:03:57 +0000 UTC" firstStartedPulling="2026-01-26 17:03:58.766985395 +0000 UTC m=+1005.291165829" lastFinishedPulling="2026-01-26 17:04:12.209814152 +0000 UTC m=+1018.733994586" observedRunningTime="2026-01-26 17:04:12.725515603 +0000 UTC m=+1019.249696037" watchObservedRunningTime="2026-01-26 17:04:12.728349263 +0000 UTC m=+1019.252529697" Jan 26 17:04:12 crc kubenswrapper[4754]: I0126 17:04:12.751073 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-598f7747c9-fvcgg" podStartSLOduration=3.456910947 podStartE2EDuration="16.751056273s" podCreationTimestamp="2026-01-26 17:03:56 +0000 UTC" firstStartedPulling="2026-01-26 17:03:58.365777711 +0000 UTC m=+1004.889958145" lastFinishedPulling="2026-01-26 17:04:11.659923037 +0000 UTC m=+1018.184103471" observedRunningTime="2026-01-26 17:04:12.742480451 +0000 UTC m=+1019.266660885" watchObservedRunningTime="2026-01-26 17:04:12.751056273 +0000 UTC m=+1019.275236707" Jan 26 17:04:12 crc kubenswrapper[4754]: I0126 17:04:12.912622 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5702a910-71d1-4acd-93aa-9379bc3147ce-cert\") pod \"infra-operator-controller-manager-694cf4f878-p66zm\" (UID: \"5702a910-71d1-4acd-93aa-9379bc3147ce\") " pod="openstack-operators/infra-operator-controller-manager-694cf4f878-p66zm" Jan 26 17:04:12 crc kubenswrapper[4754]: E0126 17:04:12.912831 4754 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Jan 26 17:04:12 crc kubenswrapper[4754]: E0126 17:04:12.912917 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5702a910-71d1-4acd-93aa-9379bc3147ce-cert podName:5702a910-71d1-4acd-93aa-9379bc3147ce nodeName:}" failed. No retries permitted until 2026-01-26 17:04:28.912898617 +0000 UTC m=+1035.437079051 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/5702a910-71d1-4acd-93aa-9379bc3147ce-cert") pod "infra-operator-controller-manager-694cf4f878-p66zm" (UID: "5702a910-71d1-4acd-93aa-9379bc3147ce") : secret "infra-operator-webhook-server-cert" not found Jan 26 17:04:13 crc kubenswrapper[4754]: I0126 17:04:13.418730 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f43cc0a3-41e9-4e09-837f-322a53221560-cert\") pod \"openstack-baremetal-operator-controller-manager-6b68b8b854sclfc\" (UID: \"f43cc0a3-41e9-4e09-837f-322a53221560\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6b68b8b854sclfc" Jan 26 17:04:13 crc kubenswrapper[4754]: I0126 17:04:13.440297 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f43cc0a3-41e9-4e09-837f-322a53221560-cert\") pod \"openstack-baremetal-operator-controller-manager-6b68b8b854sclfc\" (UID: \"f43cc0a3-41e9-4e09-837f-322a53221560\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6b68b8b854sclfc" Jan 26 17:04:13 crc kubenswrapper[4754]: I0126 17:04:13.497460 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6b68b8b854sclfc" Jan 26 17:04:13 crc kubenswrapper[4754]: I0126 17:04:13.679964 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-6b9fb5fdcb-fhwr4" event={"ID":"add7fe6b-a864-439d-a3f4-80c9cc80ddc1","Type":"ContainerStarted","Data":"bcde7a187b9127b8a7de786e0c6d3d4a328fb1ca229d293ce6b186f62185ef54"} Jan 26 17:04:13 crc kubenswrapper[4754]: I0126 17:04:13.680084 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-6b9fb5fdcb-fhwr4" Jan 26 17:04:13 crc kubenswrapper[4754]: I0126 17:04:13.684125 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-594c8c9d5d-hgnm9" event={"ID":"f747ae1d-8181-4ef2-b332-b14db483aab6","Type":"ContainerStarted","Data":"9281e47bfe4229821e3c854a00c92edfb07535ba65b52a3f64c5ac1369d5dc85"} Jan 26 17:04:13 crc kubenswrapper[4754]: I0126 17:04:13.684184 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-594c8c9d5d-hgnm9" Jan 26 17:04:13 crc kubenswrapper[4754]: I0126 17:04:13.694729 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-5f4cd88d46-fkz84" event={"ID":"4fdc4d46-264a-4689-8ad5-0ed253f805df","Type":"ContainerStarted","Data":"1bba1c9516828c2dad88b78a07146be963a8c14c5be67ed11b00f56bb46c27f3"} Jan 26 17:04:13 crc kubenswrapper[4754]: I0126 17:04:13.695415 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-5f4cd88d46-fkz84" Jan 26 17:04:13 crc kubenswrapper[4754]: I0126 17:04:13.701761 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-6f75f45d54-kppds" event={"ID":"8353f62d-eb53-46c3-ba4d-2d643f4f960b","Type":"ContainerStarted","Data":"8f9fd5205ae79ca91411829dbcc93c06a867c255079f0fe6f8c56ff5618e258e"} Jan 26 17:04:13 crc kubenswrapper[4754]: I0126 17:04:13.701837 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-6f75f45d54-kppds" Jan 26 17:04:13 crc kubenswrapper[4754]: I0126 17:04:13.705217 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-6b9fb5fdcb-fhwr4" podStartSLOduration=3.3347511020000002 podStartE2EDuration="16.705179117s" podCreationTimestamp="2026-01-26 17:03:57 +0000 UTC" firstStartedPulling="2026-01-26 17:03:58.78704026 +0000 UTC m=+1005.311220694" lastFinishedPulling="2026-01-26 17:04:12.157468285 +0000 UTC m=+1018.681648709" observedRunningTime="2026-01-26 17:04:13.697706136 +0000 UTC m=+1020.221886570" watchObservedRunningTime="2026-01-26 17:04:13.705179117 +0000 UTC m=+1020.229359551" Jan 26 17:04:13 crc kubenswrapper[4754]: I0126 17:04:13.725407 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/93622668-0766-46f4-a216-83a7a17f36fc-webhook-certs\") pod \"openstack-operator-controller-manager-59fcd4bdb5-gvt4r\" (UID: \"93622668-0766-46f4-a216-83a7a17f36fc\") " pod="openstack-operators/openstack-operator-controller-manager-59fcd4bdb5-gvt4r" Jan 26 17:04:13 crc kubenswrapper[4754]: I0126 17:04:13.725874 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/93622668-0766-46f4-a216-83a7a17f36fc-metrics-certs\") pod \"openstack-operator-controller-manager-59fcd4bdb5-gvt4r\" (UID: \"93622668-0766-46f4-a216-83a7a17f36fc\") " pod="openstack-operators/openstack-operator-controller-manager-59fcd4bdb5-gvt4r" Jan 26 17:04:13 crc kubenswrapper[4754]: I0126 17:04:13.727948 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-594c8c9d5d-hgnm9" podStartSLOduration=3.9132757959999998 podStartE2EDuration="17.727937098s" podCreationTimestamp="2026-01-26 17:03:56 +0000 UTC" firstStartedPulling="2026-01-26 17:03:58.38985264 +0000 UTC m=+1004.914033074" lastFinishedPulling="2026-01-26 17:04:12.204513942 +0000 UTC m=+1018.728694376" observedRunningTime="2026-01-26 17:04:13.72656791 +0000 UTC m=+1020.250748344" watchObservedRunningTime="2026-01-26 17:04:13.727937098 +0000 UTC m=+1020.252117532" Jan 26 17:04:13 crc kubenswrapper[4754]: I0126 17:04:13.754048 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-b45d7bf98-bggfb" event={"ID":"5e071de1-60dc-49d8-b965-90c2f99a6e02","Type":"ContainerStarted","Data":"b77387e917a1e93d0258c78d37e380a939fb21f9e4a24b7a22fc7641becc813c"} Jan 26 17:04:13 crc kubenswrapper[4754]: I0126 17:04:13.754591 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-b45d7bf98-bggfb" Jan 26 17:04:13 crc kubenswrapper[4754]: I0126 17:04:13.757730 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/93622668-0766-46f4-a216-83a7a17f36fc-webhook-certs\") pod \"openstack-operator-controller-manager-59fcd4bdb5-gvt4r\" (UID: \"93622668-0766-46f4-a216-83a7a17f36fc\") " pod="openstack-operators/openstack-operator-controller-manager-59fcd4bdb5-gvt4r" Jan 26 17:04:13 crc kubenswrapper[4754]: I0126 17:04:13.764218 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/93622668-0766-46f4-a216-83a7a17f36fc-metrics-certs\") pod \"openstack-operator-controller-manager-59fcd4bdb5-gvt4r\" (UID: \"93622668-0766-46f4-a216-83a7a17f36fc\") " pod="openstack-operators/openstack-operator-controller-manager-59fcd4bdb5-gvt4r" Jan 26 17:04:13 crc kubenswrapper[4754]: I0126 17:04:13.773849 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-ppmm7" Jan 26 17:04:13 crc kubenswrapper[4754]: I0126 17:04:13.781385 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-59fcd4bdb5-gvt4r" Jan 26 17:04:13 crc kubenswrapper[4754]: I0126 17:04:13.796339 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-5f4cd88d46-fkz84" podStartSLOduration=3.394272168 podStartE2EDuration="16.796324966s" podCreationTimestamp="2026-01-26 17:03:57 +0000 UTC" firstStartedPulling="2026-01-26 17:03:58.755318375 +0000 UTC m=+1005.279498809" lastFinishedPulling="2026-01-26 17:04:12.157371173 +0000 UTC m=+1018.681551607" observedRunningTime="2026-01-26 17:04:13.7542609 +0000 UTC m=+1020.278441334" watchObservedRunningTime="2026-01-26 17:04:13.796324966 +0000 UTC m=+1020.320505400" Jan 26 17:04:13 crc kubenswrapper[4754]: I0126 17:04:13.808610 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-78fdd796fd-c5kf9" event={"ID":"ee98073e-2fdb-4b3e-acb8-00c71df55fa7","Type":"ContainerStarted","Data":"94f52af361d02de79e801e699274963e4875e9488b35d24711479c30dd8faec8"} Jan 26 17:04:13 crc kubenswrapper[4754]: I0126 17:04:13.808689 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-78fdd796fd-c5kf9" Jan 26 17:04:13 crc kubenswrapper[4754]: I0126 17:04:13.812210 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-b45d7bf98-bggfb" podStartSLOduration=4.00916304 podStartE2EDuration="17.812195694s" podCreationTimestamp="2026-01-26 17:03:56 +0000 UTC" firstStartedPulling="2026-01-26 17:03:58.406101599 +0000 UTC m=+1004.930282033" lastFinishedPulling="2026-01-26 17:04:12.209134253 +0000 UTC m=+1018.733314687" observedRunningTime="2026-01-26 17:04:13.781271792 +0000 UTC m=+1020.305452226" watchObservedRunningTime="2026-01-26 17:04:13.812195694 +0000 UTC m=+1020.336376128" Jan 26 17:04:13 crc kubenswrapper[4754]: I0126 17:04:13.832107 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-598f7747c9-fvcgg" event={"ID":"28161331-7731-433f-845f-2ebe1daf5fd0","Type":"ContainerStarted","Data":"e2c82477a5a70d5db379b758346fb6ae7eaae39228aab4cf7bee76c9d8d07a69"} Jan 26 17:04:13 crc kubenswrapper[4754]: I0126 17:04:13.835873 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-6f75f45d54-kppds" podStartSLOduration=3.404764084 podStartE2EDuration="16.835857432s" podCreationTimestamp="2026-01-26 17:03:57 +0000 UTC" firstStartedPulling="2026-01-26 17:03:58.778177399 +0000 UTC m=+1005.302357833" lastFinishedPulling="2026-01-26 17:04:12.209270747 +0000 UTC m=+1018.733451181" observedRunningTime="2026-01-26 17:04:13.80636054 +0000 UTC m=+1020.330540974" watchObservedRunningTime="2026-01-26 17:04:13.835857432 +0000 UTC m=+1020.360037866" Jan 26 17:04:13 crc kubenswrapper[4754]: I0126 17:04:13.840878 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-7478f7dbf9-6vmhq" event={"ID":"cb2628e4-680e-489a-8fc9-d39986c74301","Type":"ContainerStarted","Data":"deb513235844db11b3dd696bcab482c6ace36407e1a622da1d0f40fd8e36669c"} Jan 26 17:04:13 crc kubenswrapper[4754]: I0126 17:04:13.841544 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-7478f7dbf9-6vmhq" Jan 26 17:04:13 crc kubenswrapper[4754]: I0126 17:04:13.876117 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-78fdd796fd-c5kf9" podStartSLOduration=4.120110599 podStartE2EDuration="17.876096686s" podCreationTimestamp="2026-01-26 17:03:56 +0000 UTC" firstStartedPulling="2026-01-26 17:03:58.401369655 +0000 UTC m=+1004.925550099" lastFinishedPulling="2026-01-26 17:04:12.157355752 +0000 UTC m=+1018.681536186" observedRunningTime="2026-01-26 17:04:13.864220192 +0000 UTC m=+1020.388400626" watchObservedRunningTime="2026-01-26 17:04:13.876096686 +0000 UTC m=+1020.400277120" Jan 26 17:04:13 crc kubenswrapper[4754]: I0126 17:04:13.929533 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-7478f7dbf9-6vmhq" podStartSLOduration=4.296010668 podStartE2EDuration="17.929511332s" podCreationTimestamp="2026-01-26 17:03:56 +0000 UTC" firstStartedPulling="2026-01-26 17:03:58.026376161 +0000 UTC m=+1004.550556595" lastFinishedPulling="2026-01-26 17:04:11.659876835 +0000 UTC m=+1018.184057259" observedRunningTime="2026-01-26 17:04:13.894861066 +0000 UTC m=+1020.419041500" watchObservedRunningTime="2026-01-26 17:04:13.929511332 +0000 UTC m=+1020.453691766" Jan 26 17:04:14 crc kubenswrapper[4754]: I0126 17:04:14.001385 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-6b68b8b854sclfc"] Jan 26 17:04:14 crc kubenswrapper[4754]: I0126 17:04:14.143179 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-59fcd4bdb5-gvt4r"] Jan 26 17:04:14 crc kubenswrapper[4754]: W0126 17:04:14.174097 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod93622668_0766_46f4_a216_83a7a17f36fc.slice/crio-3580a1b5a6466cbcdc4707d07441115008222de7bdbbb1e348b4387ec8cdf60e WatchSource:0}: Error finding container 3580a1b5a6466cbcdc4707d07441115008222de7bdbbb1e348b4387ec8cdf60e: Status 404 returned error can't find the container with id 3580a1b5a6466cbcdc4707d07441115008222de7bdbbb1e348b4387ec8cdf60e Jan 26 17:04:14 crc kubenswrapper[4754]: I0126 17:04:14.847004 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-59fcd4bdb5-gvt4r" event={"ID":"93622668-0766-46f4-a216-83a7a17f36fc","Type":"ContainerStarted","Data":"b87bbc2f76d7ac53af8e91e4430a207216d073e92dd28a277a6c455bac30a9a2"} Jan 26 17:04:14 crc kubenswrapper[4754]: I0126 17:04:14.847420 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-59fcd4bdb5-gvt4r" Jan 26 17:04:14 crc kubenswrapper[4754]: I0126 17:04:14.847434 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-59fcd4bdb5-gvt4r" event={"ID":"93622668-0766-46f4-a216-83a7a17f36fc","Type":"ContainerStarted","Data":"3580a1b5a6466cbcdc4707d07441115008222de7bdbbb1e348b4387ec8cdf60e"} Jan 26 17:04:14 crc kubenswrapper[4754]: I0126 17:04:14.848591 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6b68b8b854sclfc" event={"ID":"f43cc0a3-41e9-4e09-837f-322a53221560","Type":"ContainerStarted","Data":"8666fcba3f5d1da0818072eb7c35dab70ca2bc09ad07dc32e062f0d60c26c1da"} Jan 26 17:04:17 crc kubenswrapper[4754]: I0126 17:04:17.191070 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-7f86f8796f-5kbpl" Jan 26 17:04:17 crc kubenswrapper[4754]: I0126 17:04:17.208092 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-59fcd4bdb5-gvt4r" podStartSLOduration=20.208073118 podStartE2EDuration="20.208073118s" podCreationTimestamp="2026-01-26 17:03:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 17:04:14.878149431 +0000 UTC m=+1021.402329885" watchObservedRunningTime="2026-01-26 17:04:17.208073118 +0000 UTC m=+1023.732253562" Jan 26 17:04:17 crc kubenswrapper[4754]: I0126 17:04:17.222917 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-7478f7dbf9-6vmhq" Jan 26 17:04:17 crc kubenswrapper[4754]: I0126 17:04:17.224217 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-b45d7bf98-bggfb" Jan 26 17:04:17 crc kubenswrapper[4754]: I0126 17:04:17.243204 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-78fdd796fd-c5kf9" Jan 26 17:04:17 crc kubenswrapper[4754]: I0126 17:04:17.300861 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-594c8c9d5d-hgnm9" Jan 26 17:04:17 crc kubenswrapper[4754]: I0126 17:04:17.541052 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-598f7747c9-fvcgg" Jan 26 17:04:17 crc kubenswrapper[4754]: I0126 17:04:17.663243 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-78c6999f6f-jb5ps" Jan 26 17:04:17 crc kubenswrapper[4754]: I0126 17:04:17.690133 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-6b9fb5fdcb-fhwr4" Jan 26 17:04:17 crc kubenswrapper[4754]: I0126 17:04:17.757525 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-7bdb645866-27hg7" Jan 26 17:04:17 crc kubenswrapper[4754]: I0126 17:04:17.788280 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-5f4cd88d46-fkz84" Jan 26 17:04:17 crc kubenswrapper[4754]: I0126 17:04:17.851367 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-6f75f45d54-kppds" Jan 26 17:04:17 crc kubenswrapper[4754]: I0126 17:04:17.914773 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-85cd9769bb-dq4qw" Jan 26 17:04:23 crc kubenswrapper[4754]: I0126 17:04:23.794121 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-59fcd4bdb5-gvt4r" Jan 26 17:04:25 crc kubenswrapper[4754]: I0126 17:04:25.948059 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-78d58447c5-pgmv6" event={"ID":"2ab80530-a54c-41bf-8e4f-c895dbacb368","Type":"ContainerStarted","Data":"95d013d128ee8674ebc7fcf28c1abbba8058a68286c24662cdb863ff75c71990"} Jan 26 17:04:25 crc kubenswrapper[4754]: I0126 17:04:25.948876 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-78d58447c5-pgmv6" Jan 26 17:04:25 crc kubenswrapper[4754]: I0126 17:04:25.949179 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-mqmkw" event={"ID":"b48b5414-4034-4c0e-9d25-5053f84cf246","Type":"ContainerStarted","Data":"5087c7b71c6a63f1182c6186c7c60be9e75ccc01d0e1258b4a9bf377f46a6915"} Jan 26 17:04:25 crc kubenswrapper[4754]: I0126 17:04:25.951635 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6b68b8b854sclfc" event={"ID":"f43cc0a3-41e9-4e09-837f-322a53221560","Type":"ContainerStarted","Data":"6710baf4c3b3c8858cc9317fc089c9cb660abf9539be2d1499c041ceb537add3"} Jan 26 17:04:25 crc kubenswrapper[4754]: I0126 17:04:25.951723 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6b68b8b854sclfc" Jan 26 17:04:25 crc kubenswrapper[4754]: I0126 17:04:25.953831 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-b8b6d4659-d8db9" event={"ID":"e3aac642-b3ee-4394-9f1d-bfac315bf162","Type":"ContainerStarted","Data":"3f8df5a51923422c9918c27495c50c0dce44cffd13e25176435b2b77b6a42526"} Jan 26 17:04:25 crc kubenswrapper[4754]: I0126 17:04:25.954031 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-b8b6d4659-d8db9" Jan 26 17:04:25 crc kubenswrapper[4754]: I0126 17:04:25.954922 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-564965969-k7nbl" event={"ID":"138cc11d-6be6-43bb-994e-94ea39bb2e42","Type":"ContainerStarted","Data":"97d6179b5e15d29a1f74a0175dc88885200bb14ed6b85ac79552955dc5fd635e"} Jan 26 17:04:25 crc kubenswrapper[4754]: I0126 17:04:25.955101 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-564965969-k7nbl" Jan 26 17:04:25 crc kubenswrapper[4754]: I0126 17:04:25.956209 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-79d5ccc684-gqtkb" event={"ID":"56addd62-aeed-4139-a0db-37292a6acf8a","Type":"ContainerStarted","Data":"a517d78c0978d113e2a5448e20421a2b01605aa4ae956cdeef95478ffef51c7b"} Jan 26 17:04:25 crc kubenswrapper[4754]: I0126 17:04:25.956362 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-79d5ccc684-gqtkb" Jan 26 17:04:25 crc kubenswrapper[4754]: I0126 17:04:25.957495 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-69797bbcbd-mr8bp" event={"ID":"a8584d32-796d-47cd-8dd6-233374660688","Type":"ContainerStarted","Data":"e24ae030d772248bb163e8fc7c9bd5e845addef868bec8b1dfe3f16c01d725bc"} Jan 26 17:04:25 crc kubenswrapper[4754]: I0126 17:04:25.957691 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-69797bbcbd-mr8bp" Jan 26 17:04:25 crc kubenswrapper[4754]: I0126 17:04:25.958847 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-547cbdb99f-fmvdv" event={"ID":"aaee86fe-018f-4055-b8c5-98e3795c53d8","Type":"ContainerStarted","Data":"5d3281ec7c91fb7f8473a663ecf780f970ead0de4a9593d61b8480280c24b18f"} Jan 26 17:04:25 crc kubenswrapper[4754]: I0126 17:04:25.959024 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-547cbdb99f-fmvdv" Jan 26 17:04:25 crc kubenswrapper[4754]: I0126 17:04:25.960093 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-77d5c5b54f-8sz77" event={"ID":"fd6b6ccf-b7e9-41fb-b663-a3392d075880","Type":"ContainerStarted","Data":"1ef29d3933cf334567ac66f373f0ca9650b2627eda68cb448c83a8e923c72957"} Jan 26 17:04:25 crc kubenswrapper[4754]: I0126 17:04:25.960238 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-77d5c5b54f-8sz77" Jan 26 17:04:25 crc kubenswrapper[4754]: I0126 17:04:25.977415 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-78d58447c5-pgmv6" podStartSLOduration=3.041400099 podStartE2EDuration="28.977401386s" podCreationTimestamp="2026-01-26 17:03:57 +0000 UTC" firstStartedPulling="2026-01-26 17:03:58.807356553 +0000 UTC m=+1005.331536987" lastFinishedPulling="2026-01-26 17:04:24.74335784 +0000 UTC m=+1031.267538274" observedRunningTime="2026-01-26 17:04:25.971726996 +0000 UTC m=+1032.495907450" watchObservedRunningTime="2026-01-26 17:04:25.977401386 +0000 UTC m=+1032.501581820" Jan 26 17:04:25 crc kubenswrapper[4754]: I0126 17:04:25.989616 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-mqmkw" podStartSLOduration=3.10667726 podStartE2EDuration="28.98959666s" podCreationTimestamp="2026-01-26 17:03:57 +0000 UTC" firstStartedPulling="2026-01-26 17:03:58.887469872 +0000 UTC m=+1005.411650296" lastFinishedPulling="2026-01-26 17:04:24.770389272 +0000 UTC m=+1031.294569696" observedRunningTime="2026-01-26 17:04:25.984104405 +0000 UTC m=+1032.508284839" watchObservedRunningTime="2026-01-26 17:04:25.98959666 +0000 UTC m=+1032.513777094" Jan 26 17:04:26 crc kubenswrapper[4754]: I0126 17:04:26.012909 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6b68b8b854sclfc" podStartSLOduration=18.234182139 podStartE2EDuration="29.012890816s" podCreationTimestamp="2026-01-26 17:03:57 +0000 UTC" firstStartedPulling="2026-01-26 17:04:14.024796509 +0000 UTC m=+1020.548976933" lastFinishedPulling="2026-01-26 17:04:24.803505176 +0000 UTC m=+1031.327685610" observedRunningTime="2026-01-26 17:04:26.008297348 +0000 UTC m=+1032.532477792" watchObservedRunningTime="2026-01-26 17:04:26.012890816 +0000 UTC m=+1032.537071260" Jan 26 17:04:26 crc kubenswrapper[4754]: I0126 17:04:26.034969 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-69797bbcbd-mr8bp" podStartSLOduration=3.148441557 podStartE2EDuration="29.034949969s" podCreationTimestamp="2026-01-26 17:03:57 +0000 UTC" firstStartedPulling="2026-01-26 17:03:58.883970503 +0000 UTC m=+1005.408150927" lastFinishedPulling="2026-01-26 17:04:24.770478905 +0000 UTC m=+1031.294659339" observedRunningTime="2026-01-26 17:04:26.028633181 +0000 UTC m=+1032.552813615" watchObservedRunningTime="2026-01-26 17:04:26.034949969 +0000 UTC m=+1032.559130403" Jan 26 17:04:26 crc kubenswrapper[4754]: I0126 17:04:26.050675 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-b8b6d4659-d8db9" podStartSLOduration=2.833347083 podStartE2EDuration="29.050649242s" podCreationTimestamp="2026-01-26 17:03:57 +0000 UTC" firstStartedPulling="2026-01-26 17:03:58.837987807 +0000 UTC m=+1005.362168241" lastFinishedPulling="2026-01-26 17:04:25.055289966 +0000 UTC m=+1031.579470400" observedRunningTime="2026-01-26 17:04:26.049579591 +0000 UTC m=+1032.573760025" watchObservedRunningTime="2026-01-26 17:04:26.050649242 +0000 UTC m=+1032.574829676" Jan 26 17:04:26 crc kubenswrapper[4754]: I0126 17:04:26.070878 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-77d5c5b54f-8sz77" podStartSLOduration=3.088408618 podStartE2EDuration="30.070860051s" podCreationTimestamp="2026-01-26 17:03:56 +0000 UTC" firstStartedPulling="2026-01-26 17:03:58.349628457 +0000 UTC m=+1004.873808891" lastFinishedPulling="2026-01-26 17:04:25.33207989 +0000 UTC m=+1031.856260324" observedRunningTime="2026-01-26 17:04:26.069612557 +0000 UTC m=+1032.593793001" watchObservedRunningTime="2026-01-26 17:04:26.070860051 +0000 UTC m=+1032.595040485" Jan 26 17:04:26 crc kubenswrapper[4754]: I0126 17:04:26.086320 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-79d5ccc684-gqtkb" podStartSLOduration=4.092503097 podStartE2EDuration="29.086305017s" podCreationTimestamp="2026-01-26 17:03:57 +0000 UTC" firstStartedPulling="2026-01-26 17:03:58.828003855 +0000 UTC m=+1005.352184289" lastFinishedPulling="2026-01-26 17:04:23.821805755 +0000 UTC m=+1030.345986209" observedRunningTime="2026-01-26 17:04:26.084193897 +0000 UTC m=+1032.608374331" watchObservedRunningTime="2026-01-26 17:04:26.086305017 +0000 UTC m=+1032.610485451" Jan 26 17:04:26 crc kubenswrapper[4754]: I0126 17:04:26.101599 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-547cbdb99f-fmvdv" podStartSLOduration=4.151041017 podStartE2EDuration="29.101579167s" podCreationTimestamp="2026-01-26 17:03:57 +0000 UTC" firstStartedPulling="2026-01-26 17:03:58.871343757 +0000 UTC m=+1005.395524191" lastFinishedPulling="2026-01-26 17:04:23.821881887 +0000 UTC m=+1030.346062341" observedRunningTime="2026-01-26 17:04:26.097830052 +0000 UTC m=+1032.622010506" watchObservedRunningTime="2026-01-26 17:04:26.101579167 +0000 UTC m=+1032.625759601" Jan 26 17:04:26 crc kubenswrapper[4754]: I0126 17:04:26.113574 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-564965969-k7nbl" podStartSLOduration=4.042053814 podStartE2EDuration="29.113560065s" podCreationTimestamp="2026-01-26 17:03:57 +0000 UTC" firstStartedPulling="2026-01-26 17:03:58.871981515 +0000 UTC m=+1005.396161949" lastFinishedPulling="2026-01-26 17:04:23.943487766 +0000 UTC m=+1030.467668200" observedRunningTime="2026-01-26 17:04:26.111035904 +0000 UTC m=+1032.635216348" watchObservedRunningTime="2026-01-26 17:04:26.113560065 +0000 UTC m=+1032.637740499" Jan 26 17:04:28 crc kubenswrapper[4754]: I0126 17:04:28.950145 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5702a910-71d1-4acd-93aa-9379bc3147ce-cert\") pod \"infra-operator-controller-manager-694cf4f878-p66zm\" (UID: \"5702a910-71d1-4acd-93aa-9379bc3147ce\") " pod="openstack-operators/infra-operator-controller-manager-694cf4f878-p66zm" Jan 26 17:04:28 crc kubenswrapper[4754]: I0126 17:04:28.956250 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5702a910-71d1-4acd-93aa-9379bc3147ce-cert\") pod \"infra-operator-controller-manager-694cf4f878-p66zm\" (UID: \"5702a910-71d1-4acd-93aa-9379bc3147ce\") " pod="openstack-operators/infra-operator-controller-manager-694cf4f878-p66zm" Jan 26 17:04:28 crc kubenswrapper[4754]: I0126 17:04:28.979199 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-dlczc" Jan 26 17:04:28 crc kubenswrapper[4754]: I0126 17:04:28.987693 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-694cf4f878-p66zm" Jan 26 17:04:29 crc kubenswrapper[4754]: I0126 17:04:29.379401 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-694cf4f878-p66zm"] Jan 26 17:04:29 crc kubenswrapper[4754]: I0126 17:04:29.989388 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-694cf4f878-p66zm" event={"ID":"5702a910-71d1-4acd-93aa-9379bc3147ce","Type":"ContainerStarted","Data":"faeaa0018a1ee25707f7c584a6123f9b47fc30a4e6430aeac62f77c509efda8a"} Jan 26 17:04:33 crc kubenswrapper[4754]: I0126 17:04:33.503927 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6b68b8b854sclfc" Jan 26 17:04:37 crc kubenswrapper[4754]: I0126 17:04:37.465762 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-77d5c5b54f-8sz77" Jan 26 17:04:37 crc kubenswrapper[4754]: I0126 17:04:37.642008 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-b8b6d4659-d8db9" Jan 26 17:04:37 crc kubenswrapper[4754]: I0126 17:04:37.723744 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-78d58447c5-pgmv6" Jan 26 17:04:37 crc kubenswrapper[4754]: I0126 17:04:37.862160 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-79d5ccc684-gqtkb" Jan 26 17:04:37 crc kubenswrapper[4754]: I0126 17:04:37.877164 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-547cbdb99f-fmvdv" Jan 26 17:04:38 crc kubenswrapper[4754]: I0126 17:04:38.011366 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-69797bbcbd-mr8bp" Jan 26 17:04:38 crc kubenswrapper[4754]: I0126 17:04:38.087844 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-564965969-k7nbl" Jan 26 17:04:41 crc kubenswrapper[4754]: E0126 17:04:41.185292 4754 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = writing blob: storing blob to file \"/var/tmp/container_images_storage488735512/1\": happened during read: context canceled" image="quay.io/openstack-k8s-operators/infra-operator@sha256:72c534cbfa71ed3501bee4937ab2beb8fda27b890ef7a26789824f52710b3846" Jan 26 17:04:41 crc kubenswrapper[4754]: E0126 17:04:41.186975 4754 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/infra-operator@sha256:72c534cbfa71ed3501bee4937ab2beb8fda27b890ef7a26789824f52710b3846,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:true,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{600 -3} {} 600m DecimalSI},memory: {{2147483648 0} {} 2Gi BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{536870912 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:cert,ReadOnly:true,MountPath:/tmp/k8s-webhook-server/serving-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-95tv6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod infra-operator-controller-manager-694cf4f878-p66zm_openstack-operators(5702a910-71d1-4acd-93aa-9379bc3147ce): ErrImagePull: rpc error: code = Canceled desc = writing blob: storing blob to file \"/var/tmp/container_images_storage488735512/1\": happened during read: context canceled" logger="UnhandledError" Jan 26 17:04:41 crc kubenswrapper[4754]: E0126 17:04:41.188577 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = writing blob: storing blob to file \\\"/var/tmp/container_images_storage488735512/1\\\": happened during read: context canceled\"" pod="openstack-operators/infra-operator-controller-manager-694cf4f878-p66zm" podUID="5702a910-71d1-4acd-93aa-9379bc3147ce" Jan 26 17:04:42 crc kubenswrapper[4754]: E0126 17:04:42.073219 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/infra-operator@sha256:72c534cbfa71ed3501bee4937ab2beb8fda27b890ef7a26789824f52710b3846\\\"\"" pod="openstack-operators/infra-operator-controller-manager-694cf4f878-p66zm" podUID="5702a910-71d1-4acd-93aa-9379bc3147ce" Jan 26 17:04:53 crc kubenswrapper[4754]: I0126 17:04:53.773380 4754 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 26 17:04:57 crc kubenswrapper[4754]: I0126 17:04:57.171292 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-694cf4f878-p66zm" event={"ID":"5702a910-71d1-4acd-93aa-9379bc3147ce","Type":"ContainerStarted","Data":"cda5b797ef9d6e6ce8072931c98ee759f9c6ba1ee3ff26444fae42662f64c666"} Jan 26 17:04:57 crc kubenswrapper[4754]: I0126 17:04:57.172786 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-694cf4f878-p66zm" Jan 26 17:04:57 crc kubenswrapper[4754]: I0126 17:04:57.190152 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-694cf4f878-p66zm" podStartSLOduration=34.427828375 podStartE2EDuration="1m1.190138623s" podCreationTimestamp="2026-01-26 17:03:56 +0000 UTC" firstStartedPulling="2026-01-26 17:04:29.388091467 +0000 UTC m=+1035.912271901" lastFinishedPulling="2026-01-26 17:04:56.150401685 +0000 UTC m=+1062.674582149" observedRunningTime="2026-01-26 17:04:57.187214861 +0000 UTC m=+1063.711395295" watchObservedRunningTime="2026-01-26 17:04:57.190138623 +0000 UTC m=+1063.714319057" Jan 26 17:05:08 crc kubenswrapper[4754]: I0126 17:05:08.997136 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-694cf4f878-p66zm" Jan 26 17:05:33 crc kubenswrapper[4754]: I0126 17:05:33.290451 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-t9w6n"] Jan 26 17:05:33 crc kubenswrapper[4754]: I0126 17:05:33.293593 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-t9w6n" Jan 26 17:05:33 crc kubenswrapper[4754]: I0126 17:05:33.295577 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Jan 26 17:05:33 crc kubenswrapper[4754]: I0126 17:05:33.296167 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-lddkh" Jan 26 17:05:33 crc kubenswrapper[4754]: I0126 17:05:33.296390 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Jan 26 17:05:33 crc kubenswrapper[4754]: I0126 17:05:33.298740 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Jan 26 17:05:33 crc kubenswrapper[4754]: I0126 17:05:33.304326 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-t9w6n"] Jan 26 17:05:33 crc kubenswrapper[4754]: I0126 17:05:33.362535 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-cjsr7"] Jan 26 17:05:33 crc kubenswrapper[4754]: I0126 17:05:33.363866 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-cjsr7" Jan 26 17:05:33 crc kubenswrapper[4754]: I0126 17:05:33.366057 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Jan 26 17:05:33 crc kubenswrapper[4754]: I0126 17:05:33.382456 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-cjsr7"] Jan 26 17:05:33 crc kubenswrapper[4754]: I0126 17:05:33.452656 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jbnkc\" (UniqueName: \"kubernetes.io/projected/c097ce5f-d480-45fd-916b-8b73e3835ecb-kube-api-access-jbnkc\") pod \"dnsmasq-dns-675f4bcbfc-t9w6n\" (UID: \"c097ce5f-d480-45fd-916b-8b73e3835ecb\") " pod="openstack/dnsmasq-dns-675f4bcbfc-t9w6n" Jan 26 17:05:33 crc kubenswrapper[4754]: I0126 17:05:33.452723 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/66fcefe3-6133-4510-9887-20d407a64008-config\") pod \"dnsmasq-dns-78dd6ddcc-cjsr7\" (UID: \"66fcefe3-6133-4510-9887-20d407a64008\") " pod="openstack/dnsmasq-dns-78dd6ddcc-cjsr7" Jan 26 17:05:33 crc kubenswrapper[4754]: I0126 17:05:33.452746 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/66fcefe3-6133-4510-9887-20d407a64008-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-cjsr7\" (UID: \"66fcefe3-6133-4510-9887-20d407a64008\") " pod="openstack/dnsmasq-dns-78dd6ddcc-cjsr7" Jan 26 17:05:33 crc kubenswrapper[4754]: I0126 17:05:33.452766 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wpfcl\" (UniqueName: \"kubernetes.io/projected/66fcefe3-6133-4510-9887-20d407a64008-kube-api-access-wpfcl\") pod \"dnsmasq-dns-78dd6ddcc-cjsr7\" (UID: \"66fcefe3-6133-4510-9887-20d407a64008\") " pod="openstack/dnsmasq-dns-78dd6ddcc-cjsr7" Jan 26 17:05:33 crc kubenswrapper[4754]: I0126 17:05:33.452815 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c097ce5f-d480-45fd-916b-8b73e3835ecb-config\") pod \"dnsmasq-dns-675f4bcbfc-t9w6n\" (UID: \"c097ce5f-d480-45fd-916b-8b73e3835ecb\") " pod="openstack/dnsmasq-dns-675f4bcbfc-t9w6n" Jan 26 17:05:33 crc kubenswrapper[4754]: I0126 17:05:33.554023 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jbnkc\" (UniqueName: \"kubernetes.io/projected/c097ce5f-d480-45fd-916b-8b73e3835ecb-kube-api-access-jbnkc\") pod \"dnsmasq-dns-675f4bcbfc-t9w6n\" (UID: \"c097ce5f-d480-45fd-916b-8b73e3835ecb\") " pod="openstack/dnsmasq-dns-675f4bcbfc-t9w6n" Jan 26 17:05:33 crc kubenswrapper[4754]: I0126 17:05:33.554100 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/66fcefe3-6133-4510-9887-20d407a64008-config\") pod \"dnsmasq-dns-78dd6ddcc-cjsr7\" (UID: \"66fcefe3-6133-4510-9887-20d407a64008\") " pod="openstack/dnsmasq-dns-78dd6ddcc-cjsr7" Jan 26 17:05:33 crc kubenswrapper[4754]: I0126 17:05:33.554125 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/66fcefe3-6133-4510-9887-20d407a64008-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-cjsr7\" (UID: \"66fcefe3-6133-4510-9887-20d407a64008\") " pod="openstack/dnsmasq-dns-78dd6ddcc-cjsr7" Jan 26 17:05:33 crc kubenswrapper[4754]: I0126 17:05:33.554149 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wpfcl\" (UniqueName: \"kubernetes.io/projected/66fcefe3-6133-4510-9887-20d407a64008-kube-api-access-wpfcl\") pod \"dnsmasq-dns-78dd6ddcc-cjsr7\" (UID: \"66fcefe3-6133-4510-9887-20d407a64008\") " pod="openstack/dnsmasq-dns-78dd6ddcc-cjsr7" Jan 26 17:05:33 crc kubenswrapper[4754]: I0126 17:05:33.554214 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c097ce5f-d480-45fd-916b-8b73e3835ecb-config\") pod \"dnsmasq-dns-675f4bcbfc-t9w6n\" (UID: \"c097ce5f-d480-45fd-916b-8b73e3835ecb\") " pod="openstack/dnsmasq-dns-675f4bcbfc-t9w6n" Jan 26 17:05:33 crc kubenswrapper[4754]: I0126 17:05:33.555653 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/66fcefe3-6133-4510-9887-20d407a64008-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-cjsr7\" (UID: \"66fcefe3-6133-4510-9887-20d407a64008\") " pod="openstack/dnsmasq-dns-78dd6ddcc-cjsr7" Jan 26 17:05:33 crc kubenswrapper[4754]: I0126 17:05:33.555656 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c097ce5f-d480-45fd-916b-8b73e3835ecb-config\") pod \"dnsmasq-dns-675f4bcbfc-t9w6n\" (UID: \"c097ce5f-d480-45fd-916b-8b73e3835ecb\") " pod="openstack/dnsmasq-dns-675f4bcbfc-t9w6n" Jan 26 17:05:33 crc kubenswrapper[4754]: I0126 17:05:33.555653 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/66fcefe3-6133-4510-9887-20d407a64008-config\") pod \"dnsmasq-dns-78dd6ddcc-cjsr7\" (UID: \"66fcefe3-6133-4510-9887-20d407a64008\") " pod="openstack/dnsmasq-dns-78dd6ddcc-cjsr7" Jan 26 17:05:33 crc kubenswrapper[4754]: I0126 17:05:33.575120 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jbnkc\" (UniqueName: \"kubernetes.io/projected/c097ce5f-d480-45fd-916b-8b73e3835ecb-kube-api-access-jbnkc\") pod \"dnsmasq-dns-675f4bcbfc-t9w6n\" (UID: \"c097ce5f-d480-45fd-916b-8b73e3835ecb\") " pod="openstack/dnsmasq-dns-675f4bcbfc-t9w6n" Jan 26 17:05:33 crc kubenswrapper[4754]: I0126 17:05:33.582658 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wpfcl\" (UniqueName: \"kubernetes.io/projected/66fcefe3-6133-4510-9887-20d407a64008-kube-api-access-wpfcl\") pod \"dnsmasq-dns-78dd6ddcc-cjsr7\" (UID: \"66fcefe3-6133-4510-9887-20d407a64008\") " pod="openstack/dnsmasq-dns-78dd6ddcc-cjsr7" Jan 26 17:05:33 crc kubenswrapper[4754]: I0126 17:05:33.619649 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-t9w6n" Jan 26 17:05:33 crc kubenswrapper[4754]: I0126 17:05:33.678271 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-cjsr7" Jan 26 17:05:34 crc kubenswrapper[4754]: I0126 17:05:34.056914 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-t9w6n"] Jan 26 17:05:34 crc kubenswrapper[4754]: I0126 17:05:34.136297 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-cjsr7"] Jan 26 17:05:34 crc kubenswrapper[4754]: W0126 17:05:34.138166 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod66fcefe3_6133_4510_9887_20d407a64008.slice/crio-a0666f0f41d1b9a748758cfa1335e9c8d12c3409001099fb7adfe4fa6f1804f9 WatchSource:0}: Error finding container a0666f0f41d1b9a748758cfa1335e9c8d12c3409001099fb7adfe4fa6f1804f9: Status 404 returned error can't find the container with id a0666f0f41d1b9a748758cfa1335e9c8d12c3409001099fb7adfe4fa6f1804f9 Jan 26 17:05:34 crc kubenswrapper[4754]: I0126 17:05:34.426214 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-cjsr7" event={"ID":"66fcefe3-6133-4510-9887-20d407a64008","Type":"ContainerStarted","Data":"a0666f0f41d1b9a748758cfa1335e9c8d12c3409001099fb7adfe4fa6f1804f9"} Jan 26 17:05:34 crc kubenswrapper[4754]: I0126 17:05:34.428282 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-t9w6n" event={"ID":"c097ce5f-d480-45fd-916b-8b73e3835ecb","Type":"ContainerStarted","Data":"ba18206021eff5bf10d8b195ce4483bc3f6726afd9f6af6e03fa9557d5722435"} Jan 26 17:05:36 crc kubenswrapper[4754]: I0126 17:05:36.067392 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-t9w6n"] Jan 26 17:05:36 crc kubenswrapper[4754]: I0126 17:05:36.105974 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-xtjb4"] Jan 26 17:05:36 crc kubenswrapper[4754]: I0126 17:05:36.107116 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-xtjb4" Jan 26 17:05:36 crc kubenswrapper[4754]: I0126 17:05:36.117088 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-xtjb4"] Jan 26 17:05:36 crc kubenswrapper[4754]: I0126 17:05:36.298513 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c4637fc0-04b8-4b18-8986-7c0cebe93d45-config\") pod \"dnsmasq-dns-666b6646f7-xtjb4\" (UID: \"c4637fc0-04b8-4b18-8986-7c0cebe93d45\") " pod="openstack/dnsmasq-dns-666b6646f7-xtjb4" Jan 26 17:05:36 crc kubenswrapper[4754]: I0126 17:05:36.298591 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c4637fc0-04b8-4b18-8986-7c0cebe93d45-dns-svc\") pod \"dnsmasq-dns-666b6646f7-xtjb4\" (UID: \"c4637fc0-04b8-4b18-8986-7c0cebe93d45\") " pod="openstack/dnsmasq-dns-666b6646f7-xtjb4" Jan 26 17:05:36 crc kubenswrapper[4754]: I0126 17:05:36.300481 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j9ww2\" (UniqueName: \"kubernetes.io/projected/c4637fc0-04b8-4b18-8986-7c0cebe93d45-kube-api-access-j9ww2\") pod \"dnsmasq-dns-666b6646f7-xtjb4\" (UID: \"c4637fc0-04b8-4b18-8986-7c0cebe93d45\") " pod="openstack/dnsmasq-dns-666b6646f7-xtjb4" Jan 26 17:05:36 crc kubenswrapper[4754]: I0126 17:05:36.373907 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-cjsr7"] Jan 26 17:05:36 crc kubenswrapper[4754]: I0126 17:05:36.397538 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-72h5n"] Jan 26 17:05:36 crc kubenswrapper[4754]: I0126 17:05:36.398928 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-72h5n" Jan 26 17:05:36 crc kubenswrapper[4754]: I0126 17:05:36.402123 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c4637fc0-04b8-4b18-8986-7c0cebe93d45-dns-svc\") pod \"dnsmasq-dns-666b6646f7-xtjb4\" (UID: \"c4637fc0-04b8-4b18-8986-7c0cebe93d45\") " pod="openstack/dnsmasq-dns-666b6646f7-xtjb4" Jan 26 17:05:36 crc kubenswrapper[4754]: I0126 17:05:36.402234 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j9ww2\" (UniqueName: \"kubernetes.io/projected/c4637fc0-04b8-4b18-8986-7c0cebe93d45-kube-api-access-j9ww2\") pod \"dnsmasq-dns-666b6646f7-xtjb4\" (UID: \"c4637fc0-04b8-4b18-8986-7c0cebe93d45\") " pod="openstack/dnsmasq-dns-666b6646f7-xtjb4" Jan 26 17:05:36 crc kubenswrapper[4754]: I0126 17:05:36.402287 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c4637fc0-04b8-4b18-8986-7c0cebe93d45-config\") pod \"dnsmasq-dns-666b6646f7-xtjb4\" (UID: \"c4637fc0-04b8-4b18-8986-7c0cebe93d45\") " pod="openstack/dnsmasq-dns-666b6646f7-xtjb4" Jan 26 17:05:36 crc kubenswrapper[4754]: I0126 17:05:36.404362 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c4637fc0-04b8-4b18-8986-7c0cebe93d45-config\") pod \"dnsmasq-dns-666b6646f7-xtjb4\" (UID: \"c4637fc0-04b8-4b18-8986-7c0cebe93d45\") " pod="openstack/dnsmasq-dns-666b6646f7-xtjb4" Jan 26 17:05:36 crc kubenswrapper[4754]: I0126 17:05:36.406074 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c4637fc0-04b8-4b18-8986-7c0cebe93d45-dns-svc\") pod \"dnsmasq-dns-666b6646f7-xtjb4\" (UID: \"c4637fc0-04b8-4b18-8986-7c0cebe93d45\") " pod="openstack/dnsmasq-dns-666b6646f7-xtjb4" Jan 26 17:05:36 crc kubenswrapper[4754]: I0126 17:05:36.428287 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-72h5n"] Jan 26 17:05:36 crc kubenswrapper[4754]: I0126 17:05:36.439900 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j9ww2\" (UniqueName: \"kubernetes.io/projected/c4637fc0-04b8-4b18-8986-7c0cebe93d45-kube-api-access-j9ww2\") pod \"dnsmasq-dns-666b6646f7-xtjb4\" (UID: \"c4637fc0-04b8-4b18-8986-7c0cebe93d45\") " pod="openstack/dnsmasq-dns-666b6646f7-xtjb4" Jan 26 17:05:36 crc kubenswrapper[4754]: I0126 17:05:36.504239 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a9ccad70-dcd4-473c-aa09-925915f7420f-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-72h5n\" (UID: \"a9ccad70-dcd4-473c-aa09-925915f7420f\") " pod="openstack/dnsmasq-dns-57d769cc4f-72h5n" Jan 26 17:05:36 crc kubenswrapper[4754]: I0126 17:05:36.504633 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-98s2c\" (UniqueName: \"kubernetes.io/projected/a9ccad70-dcd4-473c-aa09-925915f7420f-kube-api-access-98s2c\") pod \"dnsmasq-dns-57d769cc4f-72h5n\" (UID: \"a9ccad70-dcd4-473c-aa09-925915f7420f\") " pod="openstack/dnsmasq-dns-57d769cc4f-72h5n" Jan 26 17:05:36 crc kubenswrapper[4754]: I0126 17:05:36.504704 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a9ccad70-dcd4-473c-aa09-925915f7420f-config\") pod \"dnsmasq-dns-57d769cc4f-72h5n\" (UID: \"a9ccad70-dcd4-473c-aa09-925915f7420f\") " pod="openstack/dnsmasq-dns-57d769cc4f-72h5n" Jan 26 17:05:36 crc kubenswrapper[4754]: I0126 17:05:36.606378 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a9ccad70-dcd4-473c-aa09-925915f7420f-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-72h5n\" (UID: \"a9ccad70-dcd4-473c-aa09-925915f7420f\") " pod="openstack/dnsmasq-dns-57d769cc4f-72h5n" Jan 26 17:05:36 crc kubenswrapper[4754]: I0126 17:05:36.606442 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-98s2c\" (UniqueName: \"kubernetes.io/projected/a9ccad70-dcd4-473c-aa09-925915f7420f-kube-api-access-98s2c\") pod \"dnsmasq-dns-57d769cc4f-72h5n\" (UID: \"a9ccad70-dcd4-473c-aa09-925915f7420f\") " pod="openstack/dnsmasq-dns-57d769cc4f-72h5n" Jan 26 17:05:36 crc kubenswrapper[4754]: I0126 17:05:36.606467 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a9ccad70-dcd4-473c-aa09-925915f7420f-config\") pod \"dnsmasq-dns-57d769cc4f-72h5n\" (UID: \"a9ccad70-dcd4-473c-aa09-925915f7420f\") " pod="openstack/dnsmasq-dns-57d769cc4f-72h5n" Jan 26 17:05:36 crc kubenswrapper[4754]: I0126 17:05:36.607610 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a9ccad70-dcd4-473c-aa09-925915f7420f-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-72h5n\" (UID: \"a9ccad70-dcd4-473c-aa09-925915f7420f\") " pod="openstack/dnsmasq-dns-57d769cc4f-72h5n" Jan 26 17:05:36 crc kubenswrapper[4754]: I0126 17:05:36.608181 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a9ccad70-dcd4-473c-aa09-925915f7420f-config\") pod \"dnsmasq-dns-57d769cc4f-72h5n\" (UID: \"a9ccad70-dcd4-473c-aa09-925915f7420f\") " pod="openstack/dnsmasq-dns-57d769cc4f-72h5n" Jan 26 17:05:36 crc kubenswrapper[4754]: I0126 17:05:36.627064 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-98s2c\" (UniqueName: \"kubernetes.io/projected/a9ccad70-dcd4-473c-aa09-925915f7420f-kube-api-access-98s2c\") pod \"dnsmasq-dns-57d769cc4f-72h5n\" (UID: \"a9ccad70-dcd4-473c-aa09-925915f7420f\") " pod="openstack/dnsmasq-dns-57d769cc4f-72h5n" Jan 26 17:05:36 crc kubenswrapper[4754]: I0126 17:05:36.718413 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-72h5n" Jan 26 17:05:36 crc kubenswrapper[4754]: I0126 17:05:36.730158 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-xtjb4" Jan 26 17:05:37 crc kubenswrapper[4754]: I0126 17:05:37.129474 4754 patch_prober.go:28] interesting pod/machine-config-daemon-x65wv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 26 17:05:37 crc kubenswrapper[4754]: I0126 17:05:37.130096 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 26 17:05:37 crc kubenswrapper[4754]: I0126 17:05:37.253983 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Jan 26 17:05:37 crc kubenswrapper[4754]: I0126 17:05:37.255730 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Jan 26 17:05:37 crc kubenswrapper[4754]: I0126 17:05:37.257732 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Jan 26 17:05:37 crc kubenswrapper[4754]: I0126 17:05:37.257774 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-xwmcb" Jan 26 17:05:37 crc kubenswrapper[4754]: I0126 17:05:37.257793 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Jan 26 17:05:37 crc kubenswrapper[4754]: I0126 17:05:37.257734 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Jan 26 17:05:37 crc kubenswrapper[4754]: I0126 17:05:37.257843 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Jan 26 17:05:37 crc kubenswrapper[4754]: I0126 17:05:37.259027 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Jan 26 17:05:37 crc kubenswrapper[4754]: I0126 17:05:37.261330 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Jan 26 17:05:37 crc kubenswrapper[4754]: I0126 17:05:37.271095 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Jan 26 17:05:37 crc kubenswrapper[4754]: I0126 17:05:37.284291 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-72h5n"] Jan 26 17:05:37 crc kubenswrapper[4754]: W0126 17:05:37.291347 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda9ccad70_dcd4_473c_aa09_925915f7420f.slice/crio-b5fb2baf91e2321685f3b671aa08db5ac6c89d321bb1ffb21c1c8bbf0966fc5c WatchSource:0}: Error finding container b5fb2baf91e2321685f3b671aa08db5ac6c89d321bb1ffb21c1c8bbf0966fc5c: Status 404 returned error can't find the container with id b5fb2baf91e2321685f3b671aa08db5ac6c89d321bb1ffb21c1c8bbf0966fc5c Jan 26 17:05:37 crc kubenswrapper[4754]: I0126 17:05:37.292956 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-xtjb4"] Jan 26 17:05:37 crc kubenswrapper[4754]: W0126 17:05:37.297412 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc4637fc0_04b8_4b18_8986_7c0cebe93d45.slice/crio-99b53c28d42483c2284404e91db83ee11ef7096181ed5c780a6a385ba5bd0fd8 WatchSource:0}: Error finding container 99b53c28d42483c2284404e91db83ee11ef7096181ed5c780a6a385ba5bd0fd8: Status 404 returned error can't find the container with id 99b53c28d42483c2284404e91db83ee11ef7096181ed5c780a6a385ba5bd0fd8 Jan 26 17:05:37 crc kubenswrapper[4754]: I0126 17:05:37.425688 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/29be5081-5097-4cd1-b35b-192df0ce8faf-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"29be5081-5097-4cd1-b35b-192df0ce8faf\") " pod="openstack/rabbitmq-server-0" Jan 26 17:05:37 crc kubenswrapper[4754]: I0126 17:05:37.425751 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/29be5081-5097-4cd1-b35b-192df0ce8faf-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"29be5081-5097-4cd1-b35b-192df0ce8faf\") " pod="openstack/rabbitmq-server-0" Jan 26 17:05:37 crc kubenswrapper[4754]: I0126 17:05:37.425787 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w6twt\" (UniqueName: \"kubernetes.io/projected/29be5081-5097-4cd1-b35b-192df0ce8faf-kube-api-access-w6twt\") pod \"rabbitmq-server-0\" (UID: \"29be5081-5097-4cd1-b35b-192df0ce8faf\") " pod="openstack/rabbitmq-server-0" Jan 26 17:05:37 crc kubenswrapper[4754]: I0126 17:05:37.425926 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/29be5081-5097-4cd1-b35b-192df0ce8faf-pod-info\") pod \"rabbitmq-server-0\" (UID: \"29be5081-5097-4cd1-b35b-192df0ce8faf\") " pod="openstack/rabbitmq-server-0" Jan 26 17:05:37 crc kubenswrapper[4754]: I0126 17:05:37.425973 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"29be5081-5097-4cd1-b35b-192df0ce8faf\") " pod="openstack/rabbitmq-server-0" Jan 26 17:05:37 crc kubenswrapper[4754]: I0126 17:05:37.426004 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/29be5081-5097-4cd1-b35b-192df0ce8faf-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"29be5081-5097-4cd1-b35b-192df0ce8faf\") " pod="openstack/rabbitmq-server-0" Jan 26 17:05:37 crc kubenswrapper[4754]: I0126 17:05:37.426076 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/29be5081-5097-4cd1-b35b-192df0ce8faf-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"29be5081-5097-4cd1-b35b-192df0ce8faf\") " pod="openstack/rabbitmq-server-0" Jan 26 17:05:37 crc kubenswrapper[4754]: I0126 17:05:37.426144 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/29be5081-5097-4cd1-b35b-192df0ce8faf-server-conf\") pod \"rabbitmq-server-0\" (UID: \"29be5081-5097-4cd1-b35b-192df0ce8faf\") " pod="openstack/rabbitmq-server-0" Jan 26 17:05:37 crc kubenswrapper[4754]: I0126 17:05:37.426181 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/29be5081-5097-4cd1-b35b-192df0ce8faf-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"29be5081-5097-4cd1-b35b-192df0ce8faf\") " pod="openstack/rabbitmq-server-0" Jan 26 17:05:37 crc kubenswrapper[4754]: I0126 17:05:37.426245 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/29be5081-5097-4cd1-b35b-192df0ce8faf-config-data\") pod \"rabbitmq-server-0\" (UID: \"29be5081-5097-4cd1-b35b-192df0ce8faf\") " pod="openstack/rabbitmq-server-0" Jan 26 17:05:37 crc kubenswrapper[4754]: I0126 17:05:37.426272 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/29be5081-5097-4cd1-b35b-192df0ce8faf-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"29be5081-5097-4cd1-b35b-192df0ce8faf\") " pod="openstack/rabbitmq-server-0" Jan 26 17:05:37 crc kubenswrapper[4754]: I0126 17:05:37.461784 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-xtjb4" event={"ID":"c4637fc0-04b8-4b18-8986-7c0cebe93d45","Type":"ContainerStarted","Data":"99b53c28d42483c2284404e91db83ee11ef7096181ed5c780a6a385ba5bd0fd8"} Jan 26 17:05:37 crc kubenswrapper[4754]: I0126 17:05:37.463204 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-72h5n" event={"ID":"a9ccad70-dcd4-473c-aa09-925915f7420f","Type":"ContainerStarted","Data":"b5fb2baf91e2321685f3b671aa08db5ac6c89d321bb1ffb21c1c8bbf0966fc5c"} Jan 26 17:05:37 crc kubenswrapper[4754]: I0126 17:05:37.530267 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/29be5081-5097-4cd1-b35b-192df0ce8faf-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"29be5081-5097-4cd1-b35b-192df0ce8faf\") " pod="openstack/rabbitmq-server-0" Jan 26 17:05:37 crc kubenswrapper[4754]: I0126 17:05:37.530360 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/29be5081-5097-4cd1-b35b-192df0ce8faf-config-data\") pod \"rabbitmq-server-0\" (UID: \"29be5081-5097-4cd1-b35b-192df0ce8faf\") " pod="openstack/rabbitmq-server-0" Jan 26 17:05:37 crc kubenswrapper[4754]: I0126 17:05:37.530387 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/29be5081-5097-4cd1-b35b-192df0ce8faf-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"29be5081-5097-4cd1-b35b-192df0ce8faf\") " pod="openstack/rabbitmq-server-0" Jan 26 17:05:37 crc kubenswrapper[4754]: I0126 17:05:37.530464 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/29be5081-5097-4cd1-b35b-192df0ce8faf-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"29be5081-5097-4cd1-b35b-192df0ce8faf\") " pod="openstack/rabbitmq-server-0" Jan 26 17:05:37 crc kubenswrapper[4754]: I0126 17:05:37.530524 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/29be5081-5097-4cd1-b35b-192df0ce8faf-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"29be5081-5097-4cd1-b35b-192df0ce8faf\") " pod="openstack/rabbitmq-server-0" Jan 26 17:05:37 crc kubenswrapper[4754]: I0126 17:05:37.530545 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w6twt\" (UniqueName: \"kubernetes.io/projected/29be5081-5097-4cd1-b35b-192df0ce8faf-kube-api-access-w6twt\") pod \"rabbitmq-server-0\" (UID: \"29be5081-5097-4cd1-b35b-192df0ce8faf\") " pod="openstack/rabbitmq-server-0" Jan 26 17:05:37 crc kubenswrapper[4754]: I0126 17:05:37.530614 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/29be5081-5097-4cd1-b35b-192df0ce8faf-pod-info\") pod \"rabbitmq-server-0\" (UID: \"29be5081-5097-4cd1-b35b-192df0ce8faf\") " pod="openstack/rabbitmq-server-0" Jan 26 17:05:37 crc kubenswrapper[4754]: I0126 17:05:37.530681 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"29be5081-5097-4cd1-b35b-192df0ce8faf\") " pod="openstack/rabbitmq-server-0" Jan 26 17:05:37 crc kubenswrapper[4754]: I0126 17:05:37.530709 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/29be5081-5097-4cd1-b35b-192df0ce8faf-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"29be5081-5097-4cd1-b35b-192df0ce8faf\") " pod="openstack/rabbitmq-server-0" Jan 26 17:05:37 crc kubenswrapper[4754]: I0126 17:05:37.530770 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/29be5081-5097-4cd1-b35b-192df0ce8faf-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"29be5081-5097-4cd1-b35b-192df0ce8faf\") " pod="openstack/rabbitmq-server-0" Jan 26 17:05:37 crc kubenswrapper[4754]: I0126 17:05:37.530845 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/29be5081-5097-4cd1-b35b-192df0ce8faf-server-conf\") pod \"rabbitmq-server-0\" (UID: \"29be5081-5097-4cd1-b35b-192df0ce8faf\") " pod="openstack/rabbitmq-server-0" Jan 26 17:05:37 crc kubenswrapper[4754]: I0126 17:05:37.531165 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/29be5081-5097-4cd1-b35b-192df0ce8faf-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"29be5081-5097-4cd1-b35b-192df0ce8faf\") " pod="openstack/rabbitmq-server-0" Jan 26 17:05:37 crc kubenswrapper[4754]: I0126 17:05:37.532404 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/29be5081-5097-4cd1-b35b-192df0ce8faf-config-data\") pod \"rabbitmq-server-0\" (UID: \"29be5081-5097-4cd1-b35b-192df0ce8faf\") " pod="openstack/rabbitmq-server-0" Jan 26 17:05:37 crc kubenswrapper[4754]: I0126 17:05:37.533551 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/29be5081-5097-4cd1-b35b-192df0ce8faf-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"29be5081-5097-4cd1-b35b-192df0ce8faf\") " pod="openstack/rabbitmq-server-0" Jan 26 17:05:37 crc kubenswrapper[4754]: I0126 17:05:37.534306 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/29be5081-5097-4cd1-b35b-192df0ce8faf-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"29be5081-5097-4cd1-b35b-192df0ce8faf\") " pod="openstack/rabbitmq-server-0" Jan 26 17:05:37 crc kubenswrapper[4754]: I0126 17:05:37.534506 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/29be5081-5097-4cd1-b35b-192df0ce8faf-server-conf\") pod \"rabbitmq-server-0\" (UID: \"29be5081-5097-4cd1-b35b-192df0ce8faf\") " pod="openstack/rabbitmq-server-0" Jan 26 17:05:37 crc kubenswrapper[4754]: I0126 17:05:37.534843 4754 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"29be5081-5097-4cd1-b35b-192df0ce8faf\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/rabbitmq-server-0" Jan 26 17:05:37 crc kubenswrapper[4754]: I0126 17:05:37.536812 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Jan 26 17:05:37 crc kubenswrapper[4754]: I0126 17:05:37.538220 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Jan 26 17:05:37 crc kubenswrapper[4754]: I0126 17:05:37.544780 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Jan 26 17:05:37 crc kubenswrapper[4754]: I0126 17:05:37.545067 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-mv2rb" Jan 26 17:05:37 crc kubenswrapper[4754]: I0126 17:05:37.546091 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Jan 26 17:05:37 crc kubenswrapper[4754]: I0126 17:05:37.546377 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Jan 26 17:05:37 crc kubenswrapper[4754]: I0126 17:05:37.547921 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Jan 26 17:05:37 crc kubenswrapper[4754]: I0126 17:05:37.548317 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Jan 26 17:05:37 crc kubenswrapper[4754]: I0126 17:05:37.548501 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Jan 26 17:05:37 crc kubenswrapper[4754]: I0126 17:05:37.553725 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Jan 26 17:05:37 crc kubenswrapper[4754]: I0126 17:05:37.554747 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/29be5081-5097-4cd1-b35b-192df0ce8faf-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"29be5081-5097-4cd1-b35b-192df0ce8faf\") " pod="openstack/rabbitmq-server-0" Jan 26 17:05:37 crc kubenswrapper[4754]: I0126 17:05:37.554885 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/29be5081-5097-4cd1-b35b-192df0ce8faf-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"29be5081-5097-4cd1-b35b-192df0ce8faf\") " pod="openstack/rabbitmq-server-0" Jan 26 17:05:37 crc kubenswrapper[4754]: I0126 17:05:37.555029 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/29be5081-5097-4cd1-b35b-192df0ce8faf-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"29be5081-5097-4cd1-b35b-192df0ce8faf\") " pod="openstack/rabbitmq-server-0" Jan 26 17:05:37 crc kubenswrapper[4754]: I0126 17:05:37.556545 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/29be5081-5097-4cd1-b35b-192df0ce8faf-pod-info\") pod \"rabbitmq-server-0\" (UID: \"29be5081-5097-4cd1-b35b-192df0ce8faf\") " pod="openstack/rabbitmq-server-0" Jan 26 17:05:37 crc kubenswrapper[4754]: I0126 17:05:37.558190 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w6twt\" (UniqueName: \"kubernetes.io/projected/29be5081-5097-4cd1-b35b-192df0ce8faf-kube-api-access-w6twt\") pod \"rabbitmq-server-0\" (UID: \"29be5081-5097-4cd1-b35b-192df0ce8faf\") " pod="openstack/rabbitmq-server-0" Jan 26 17:05:37 crc kubenswrapper[4754]: I0126 17:05:37.591860 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"29be5081-5097-4cd1-b35b-192df0ce8faf\") " pod="openstack/rabbitmq-server-0" Jan 26 17:05:37 crc kubenswrapper[4754]: I0126 17:05:37.734827 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/48810c5e-3c40-4cdc-8bab-47efa97e76fa-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"48810c5e-3c40-4cdc-8bab-47efa97e76fa\") " pod="openstack/rabbitmq-cell1-server-0" Jan 26 17:05:37 crc kubenswrapper[4754]: I0126 17:05:37.735019 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/48810c5e-3c40-4cdc-8bab-47efa97e76fa-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"48810c5e-3c40-4cdc-8bab-47efa97e76fa\") " pod="openstack/rabbitmq-cell1-server-0" Jan 26 17:05:37 crc kubenswrapper[4754]: I0126 17:05:37.735053 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/48810c5e-3c40-4cdc-8bab-47efa97e76fa-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"48810c5e-3c40-4cdc-8bab-47efa97e76fa\") " pod="openstack/rabbitmq-cell1-server-0" Jan 26 17:05:37 crc kubenswrapper[4754]: I0126 17:05:37.735140 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/48810c5e-3c40-4cdc-8bab-47efa97e76fa-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"48810c5e-3c40-4cdc-8bab-47efa97e76fa\") " pod="openstack/rabbitmq-cell1-server-0" Jan 26 17:05:37 crc kubenswrapper[4754]: I0126 17:05:37.735193 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/48810c5e-3c40-4cdc-8bab-47efa97e76fa-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"48810c5e-3c40-4cdc-8bab-47efa97e76fa\") " pod="openstack/rabbitmq-cell1-server-0" Jan 26 17:05:37 crc kubenswrapper[4754]: I0126 17:05:37.735217 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/48810c5e-3c40-4cdc-8bab-47efa97e76fa-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"48810c5e-3c40-4cdc-8bab-47efa97e76fa\") " pod="openstack/rabbitmq-cell1-server-0" Jan 26 17:05:37 crc kubenswrapper[4754]: I0126 17:05:37.735294 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/48810c5e-3c40-4cdc-8bab-47efa97e76fa-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"48810c5e-3c40-4cdc-8bab-47efa97e76fa\") " pod="openstack/rabbitmq-cell1-server-0" Jan 26 17:05:37 crc kubenswrapper[4754]: I0126 17:05:37.735370 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"48810c5e-3c40-4cdc-8bab-47efa97e76fa\") " pod="openstack/rabbitmq-cell1-server-0" Jan 26 17:05:37 crc kubenswrapper[4754]: I0126 17:05:37.735517 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/48810c5e-3c40-4cdc-8bab-47efa97e76fa-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"48810c5e-3c40-4cdc-8bab-47efa97e76fa\") " pod="openstack/rabbitmq-cell1-server-0" Jan 26 17:05:37 crc kubenswrapper[4754]: I0126 17:05:37.735566 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p6m5j\" (UniqueName: \"kubernetes.io/projected/48810c5e-3c40-4cdc-8bab-47efa97e76fa-kube-api-access-p6m5j\") pod \"rabbitmq-cell1-server-0\" (UID: \"48810c5e-3c40-4cdc-8bab-47efa97e76fa\") " pod="openstack/rabbitmq-cell1-server-0" Jan 26 17:05:37 crc kubenswrapper[4754]: I0126 17:05:37.735592 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/48810c5e-3c40-4cdc-8bab-47efa97e76fa-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"48810c5e-3c40-4cdc-8bab-47efa97e76fa\") " pod="openstack/rabbitmq-cell1-server-0" Jan 26 17:05:37 crc kubenswrapper[4754]: I0126 17:05:37.838741 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/48810c5e-3c40-4cdc-8bab-47efa97e76fa-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"48810c5e-3c40-4cdc-8bab-47efa97e76fa\") " pod="openstack/rabbitmq-cell1-server-0" Jan 26 17:05:37 crc kubenswrapper[4754]: I0126 17:05:37.838817 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/48810c5e-3c40-4cdc-8bab-47efa97e76fa-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"48810c5e-3c40-4cdc-8bab-47efa97e76fa\") " pod="openstack/rabbitmq-cell1-server-0" Jan 26 17:05:37 crc kubenswrapper[4754]: I0126 17:05:37.838854 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/48810c5e-3c40-4cdc-8bab-47efa97e76fa-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"48810c5e-3c40-4cdc-8bab-47efa97e76fa\") " pod="openstack/rabbitmq-cell1-server-0" Jan 26 17:05:37 crc kubenswrapper[4754]: I0126 17:05:37.838894 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/48810c5e-3c40-4cdc-8bab-47efa97e76fa-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"48810c5e-3c40-4cdc-8bab-47efa97e76fa\") " pod="openstack/rabbitmq-cell1-server-0" Jan 26 17:05:37 crc kubenswrapper[4754]: I0126 17:05:37.838927 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/48810c5e-3c40-4cdc-8bab-47efa97e76fa-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"48810c5e-3c40-4cdc-8bab-47efa97e76fa\") " pod="openstack/rabbitmq-cell1-server-0" Jan 26 17:05:37 crc kubenswrapper[4754]: I0126 17:05:37.838948 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/48810c5e-3c40-4cdc-8bab-47efa97e76fa-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"48810c5e-3c40-4cdc-8bab-47efa97e76fa\") " pod="openstack/rabbitmq-cell1-server-0" Jan 26 17:05:37 crc kubenswrapper[4754]: I0126 17:05:37.838963 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/48810c5e-3c40-4cdc-8bab-47efa97e76fa-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"48810c5e-3c40-4cdc-8bab-47efa97e76fa\") " pod="openstack/rabbitmq-cell1-server-0" Jan 26 17:05:37 crc kubenswrapper[4754]: I0126 17:05:37.839011 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"48810c5e-3c40-4cdc-8bab-47efa97e76fa\") " pod="openstack/rabbitmq-cell1-server-0" Jan 26 17:05:37 crc kubenswrapper[4754]: I0126 17:05:37.839035 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/48810c5e-3c40-4cdc-8bab-47efa97e76fa-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"48810c5e-3c40-4cdc-8bab-47efa97e76fa\") " pod="openstack/rabbitmq-cell1-server-0" Jan 26 17:05:37 crc kubenswrapper[4754]: I0126 17:05:37.839058 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/48810c5e-3c40-4cdc-8bab-47efa97e76fa-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"48810c5e-3c40-4cdc-8bab-47efa97e76fa\") " pod="openstack/rabbitmq-cell1-server-0" Jan 26 17:05:37 crc kubenswrapper[4754]: I0126 17:05:37.839090 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p6m5j\" (UniqueName: \"kubernetes.io/projected/48810c5e-3c40-4cdc-8bab-47efa97e76fa-kube-api-access-p6m5j\") pod \"rabbitmq-cell1-server-0\" (UID: \"48810c5e-3c40-4cdc-8bab-47efa97e76fa\") " pod="openstack/rabbitmq-cell1-server-0" Jan 26 17:05:37 crc kubenswrapper[4754]: I0126 17:05:37.840277 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/48810c5e-3c40-4cdc-8bab-47efa97e76fa-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"48810c5e-3c40-4cdc-8bab-47efa97e76fa\") " pod="openstack/rabbitmq-cell1-server-0" Jan 26 17:05:37 crc kubenswrapper[4754]: I0126 17:05:37.840486 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/48810c5e-3c40-4cdc-8bab-47efa97e76fa-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"48810c5e-3c40-4cdc-8bab-47efa97e76fa\") " pod="openstack/rabbitmq-cell1-server-0" Jan 26 17:05:37 crc kubenswrapper[4754]: I0126 17:05:37.840632 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/48810c5e-3c40-4cdc-8bab-47efa97e76fa-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"48810c5e-3c40-4cdc-8bab-47efa97e76fa\") " pod="openstack/rabbitmq-cell1-server-0" Jan 26 17:05:37 crc kubenswrapper[4754]: I0126 17:05:37.840749 4754 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"48810c5e-3c40-4cdc-8bab-47efa97e76fa\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/rabbitmq-cell1-server-0" Jan 26 17:05:37 crc kubenswrapper[4754]: I0126 17:05:37.840884 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/48810c5e-3c40-4cdc-8bab-47efa97e76fa-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"48810c5e-3c40-4cdc-8bab-47efa97e76fa\") " pod="openstack/rabbitmq-cell1-server-0" Jan 26 17:05:37 crc kubenswrapper[4754]: I0126 17:05:37.841004 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/48810c5e-3c40-4cdc-8bab-47efa97e76fa-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"48810c5e-3c40-4cdc-8bab-47efa97e76fa\") " pod="openstack/rabbitmq-cell1-server-0" Jan 26 17:05:37 crc kubenswrapper[4754]: I0126 17:05:37.843589 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/48810c5e-3c40-4cdc-8bab-47efa97e76fa-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"48810c5e-3c40-4cdc-8bab-47efa97e76fa\") " pod="openstack/rabbitmq-cell1-server-0" Jan 26 17:05:37 crc kubenswrapper[4754]: I0126 17:05:37.843595 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/48810c5e-3c40-4cdc-8bab-47efa97e76fa-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"48810c5e-3c40-4cdc-8bab-47efa97e76fa\") " pod="openstack/rabbitmq-cell1-server-0" Jan 26 17:05:37 crc kubenswrapper[4754]: I0126 17:05:37.844352 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/48810c5e-3c40-4cdc-8bab-47efa97e76fa-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"48810c5e-3c40-4cdc-8bab-47efa97e76fa\") " pod="openstack/rabbitmq-cell1-server-0" Jan 26 17:05:37 crc kubenswrapper[4754]: I0126 17:05:37.845369 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/48810c5e-3c40-4cdc-8bab-47efa97e76fa-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"48810c5e-3c40-4cdc-8bab-47efa97e76fa\") " pod="openstack/rabbitmq-cell1-server-0" Jan 26 17:05:37 crc kubenswrapper[4754]: I0126 17:05:37.862485 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"48810c5e-3c40-4cdc-8bab-47efa97e76fa\") " pod="openstack/rabbitmq-cell1-server-0" Jan 26 17:05:37 crc kubenswrapper[4754]: I0126 17:05:37.866628 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p6m5j\" (UniqueName: \"kubernetes.io/projected/48810c5e-3c40-4cdc-8bab-47efa97e76fa-kube-api-access-p6m5j\") pod \"rabbitmq-cell1-server-0\" (UID: \"48810c5e-3c40-4cdc-8bab-47efa97e76fa\") " pod="openstack/rabbitmq-cell1-server-0" Jan 26 17:05:37 crc kubenswrapper[4754]: I0126 17:05:37.890614 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Jan 26 17:05:37 crc kubenswrapper[4754]: I0126 17:05:37.958470 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Jan 26 17:05:38 crc kubenswrapper[4754]: I0126 17:05:38.523536 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Jan 26 17:05:38 crc kubenswrapper[4754]: W0126 17:05:38.541810 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod29be5081_5097_4cd1_b35b_192df0ce8faf.slice/crio-24b4abcf05c2fd2797980e930d42b56f348c915f39e805757ce1375a89a43382 WatchSource:0}: Error finding container 24b4abcf05c2fd2797980e930d42b56f348c915f39e805757ce1375a89a43382: Status 404 returned error can't find the container with id 24b4abcf05c2fd2797980e930d42b56f348c915f39e805757ce1375a89a43382 Jan 26 17:05:38 crc kubenswrapper[4754]: I0126 17:05:38.602550 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Jan 26 17:05:38 crc kubenswrapper[4754]: W0126 17:05:38.628211 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod48810c5e_3c40_4cdc_8bab_47efa97e76fa.slice/crio-176b2343d92554d1e43a227e02c03aca3ee8b66ec8be27ba113fd85492164c81 WatchSource:0}: Error finding container 176b2343d92554d1e43a227e02c03aca3ee8b66ec8be27ba113fd85492164c81: Status 404 returned error can't find the container with id 176b2343d92554d1e43a227e02c03aca3ee8b66ec8be27ba113fd85492164c81 Jan 26 17:05:39 crc kubenswrapper[4754]: I0126 17:05:39.013209 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Jan 26 17:05:39 crc kubenswrapper[4754]: I0126 17:05:39.015074 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Jan 26 17:05:39 crc kubenswrapper[4754]: I0126 17:05:39.018618 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Jan 26 17:05:39 crc kubenswrapper[4754]: I0126 17:05:39.024979 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-nbnrz" Jan 26 17:05:39 crc kubenswrapper[4754]: I0126 17:05:39.025021 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Jan 26 17:05:39 crc kubenswrapper[4754]: I0126 17:05:39.025178 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Jan 26 17:05:39 crc kubenswrapper[4754]: I0126 17:05:39.031745 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Jan 26 17:05:39 crc kubenswrapper[4754]: I0126 17:05:39.034585 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Jan 26 17:05:39 crc kubenswrapper[4754]: I0126 17:05:39.163012 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/0390a195-4349-4dae-8ae2-72b9b16af4f7-kolla-config\") pod \"openstack-galera-0\" (UID: \"0390a195-4349-4dae-8ae2-72b9b16af4f7\") " pod="openstack/openstack-galera-0" Jan 26 17:05:39 crc kubenswrapper[4754]: I0126 17:05:39.163370 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/0390a195-4349-4dae-8ae2-72b9b16af4f7-config-data-default\") pod \"openstack-galera-0\" (UID: \"0390a195-4349-4dae-8ae2-72b9b16af4f7\") " pod="openstack/openstack-galera-0" Jan 26 17:05:39 crc kubenswrapper[4754]: I0126 17:05:39.163432 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/0390a195-4349-4dae-8ae2-72b9b16af4f7-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"0390a195-4349-4dae-8ae2-72b9b16af4f7\") " pod="openstack/openstack-galera-0" Jan 26 17:05:39 crc kubenswrapper[4754]: I0126 17:05:39.163495 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0390a195-4349-4dae-8ae2-72b9b16af4f7-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"0390a195-4349-4dae-8ae2-72b9b16af4f7\") " pod="openstack/openstack-galera-0" Jan 26 17:05:39 crc kubenswrapper[4754]: I0126 17:05:39.163563 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/0390a195-4349-4dae-8ae2-72b9b16af4f7-config-data-generated\") pod \"openstack-galera-0\" (UID: \"0390a195-4349-4dae-8ae2-72b9b16af4f7\") " pod="openstack/openstack-galera-0" Jan 26 17:05:39 crc kubenswrapper[4754]: I0126 17:05:39.163746 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-99fgx\" (UniqueName: \"kubernetes.io/projected/0390a195-4349-4dae-8ae2-72b9b16af4f7-kube-api-access-99fgx\") pod \"openstack-galera-0\" (UID: \"0390a195-4349-4dae-8ae2-72b9b16af4f7\") " pod="openstack/openstack-galera-0" Jan 26 17:05:39 crc kubenswrapper[4754]: I0126 17:05:39.163842 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0390a195-4349-4dae-8ae2-72b9b16af4f7-operator-scripts\") pod \"openstack-galera-0\" (UID: \"0390a195-4349-4dae-8ae2-72b9b16af4f7\") " pod="openstack/openstack-galera-0" Jan 26 17:05:39 crc kubenswrapper[4754]: I0126 17:05:39.163926 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"openstack-galera-0\" (UID: \"0390a195-4349-4dae-8ae2-72b9b16af4f7\") " pod="openstack/openstack-galera-0" Jan 26 17:05:39 crc kubenswrapper[4754]: I0126 17:05:39.266796 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/0390a195-4349-4dae-8ae2-72b9b16af4f7-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"0390a195-4349-4dae-8ae2-72b9b16af4f7\") " pod="openstack/openstack-galera-0" Jan 26 17:05:39 crc kubenswrapper[4754]: I0126 17:05:39.266859 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0390a195-4349-4dae-8ae2-72b9b16af4f7-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"0390a195-4349-4dae-8ae2-72b9b16af4f7\") " pod="openstack/openstack-galera-0" Jan 26 17:05:39 crc kubenswrapper[4754]: I0126 17:05:39.266891 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/0390a195-4349-4dae-8ae2-72b9b16af4f7-config-data-generated\") pod \"openstack-galera-0\" (UID: \"0390a195-4349-4dae-8ae2-72b9b16af4f7\") " pod="openstack/openstack-galera-0" Jan 26 17:05:39 crc kubenswrapper[4754]: I0126 17:05:39.266945 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-99fgx\" (UniqueName: \"kubernetes.io/projected/0390a195-4349-4dae-8ae2-72b9b16af4f7-kube-api-access-99fgx\") pod \"openstack-galera-0\" (UID: \"0390a195-4349-4dae-8ae2-72b9b16af4f7\") " pod="openstack/openstack-galera-0" Jan 26 17:05:39 crc kubenswrapper[4754]: I0126 17:05:39.266974 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0390a195-4349-4dae-8ae2-72b9b16af4f7-operator-scripts\") pod \"openstack-galera-0\" (UID: \"0390a195-4349-4dae-8ae2-72b9b16af4f7\") " pod="openstack/openstack-galera-0" Jan 26 17:05:39 crc kubenswrapper[4754]: I0126 17:05:39.267009 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"openstack-galera-0\" (UID: \"0390a195-4349-4dae-8ae2-72b9b16af4f7\") " pod="openstack/openstack-galera-0" Jan 26 17:05:39 crc kubenswrapper[4754]: I0126 17:05:39.267041 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/0390a195-4349-4dae-8ae2-72b9b16af4f7-kolla-config\") pod \"openstack-galera-0\" (UID: \"0390a195-4349-4dae-8ae2-72b9b16af4f7\") " pod="openstack/openstack-galera-0" Jan 26 17:05:39 crc kubenswrapper[4754]: I0126 17:05:39.267084 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/0390a195-4349-4dae-8ae2-72b9b16af4f7-config-data-default\") pod \"openstack-galera-0\" (UID: \"0390a195-4349-4dae-8ae2-72b9b16af4f7\") " pod="openstack/openstack-galera-0" Jan 26 17:05:39 crc kubenswrapper[4754]: I0126 17:05:39.267960 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/0390a195-4349-4dae-8ae2-72b9b16af4f7-config-data-default\") pod \"openstack-galera-0\" (UID: \"0390a195-4349-4dae-8ae2-72b9b16af4f7\") " pod="openstack/openstack-galera-0" Jan 26 17:05:39 crc kubenswrapper[4754]: I0126 17:05:39.269267 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/0390a195-4349-4dae-8ae2-72b9b16af4f7-config-data-generated\") pod \"openstack-galera-0\" (UID: \"0390a195-4349-4dae-8ae2-72b9b16af4f7\") " pod="openstack/openstack-galera-0" Jan 26 17:05:39 crc kubenswrapper[4754]: I0126 17:05:39.269453 4754 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"openstack-galera-0\" (UID: \"0390a195-4349-4dae-8ae2-72b9b16af4f7\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/openstack-galera-0" Jan 26 17:05:39 crc kubenswrapper[4754]: I0126 17:05:39.270396 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/0390a195-4349-4dae-8ae2-72b9b16af4f7-kolla-config\") pod \"openstack-galera-0\" (UID: \"0390a195-4349-4dae-8ae2-72b9b16af4f7\") " pod="openstack/openstack-galera-0" Jan 26 17:05:39 crc kubenswrapper[4754]: I0126 17:05:39.271557 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0390a195-4349-4dae-8ae2-72b9b16af4f7-operator-scripts\") pod \"openstack-galera-0\" (UID: \"0390a195-4349-4dae-8ae2-72b9b16af4f7\") " pod="openstack/openstack-galera-0" Jan 26 17:05:39 crc kubenswrapper[4754]: I0126 17:05:39.279262 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0390a195-4349-4dae-8ae2-72b9b16af4f7-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"0390a195-4349-4dae-8ae2-72b9b16af4f7\") " pod="openstack/openstack-galera-0" Jan 26 17:05:39 crc kubenswrapper[4754]: I0126 17:05:39.279266 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/0390a195-4349-4dae-8ae2-72b9b16af4f7-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"0390a195-4349-4dae-8ae2-72b9b16af4f7\") " pod="openstack/openstack-galera-0" Jan 26 17:05:39 crc kubenswrapper[4754]: I0126 17:05:39.287612 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-99fgx\" (UniqueName: \"kubernetes.io/projected/0390a195-4349-4dae-8ae2-72b9b16af4f7-kube-api-access-99fgx\") pod \"openstack-galera-0\" (UID: \"0390a195-4349-4dae-8ae2-72b9b16af4f7\") " pod="openstack/openstack-galera-0" Jan 26 17:05:39 crc kubenswrapper[4754]: I0126 17:05:39.302370 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"openstack-galera-0\" (UID: \"0390a195-4349-4dae-8ae2-72b9b16af4f7\") " pod="openstack/openstack-galera-0" Jan 26 17:05:39 crc kubenswrapper[4754]: I0126 17:05:39.347200 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Jan 26 17:05:39 crc kubenswrapper[4754]: I0126 17:05:39.483600 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"29be5081-5097-4cd1-b35b-192df0ce8faf","Type":"ContainerStarted","Data":"24b4abcf05c2fd2797980e930d42b56f348c915f39e805757ce1375a89a43382"} Jan 26 17:05:39 crc kubenswrapper[4754]: I0126 17:05:39.485250 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"48810c5e-3c40-4cdc-8bab-47efa97e76fa","Type":"ContainerStarted","Data":"176b2343d92554d1e43a227e02c03aca3ee8b66ec8be27ba113fd85492164c81"} Jan 26 17:05:40 crc kubenswrapper[4754]: I0126 17:05:40.406144 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Jan 26 17:05:40 crc kubenswrapper[4754]: I0126 17:05:40.410823 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Jan 26 17:05:40 crc kubenswrapper[4754]: I0126 17:05:40.413866 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-8tzf5" Jan 26 17:05:40 crc kubenswrapper[4754]: I0126 17:05:40.413905 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Jan 26 17:05:40 crc kubenswrapper[4754]: I0126 17:05:40.414281 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Jan 26 17:05:40 crc kubenswrapper[4754]: I0126 17:05:40.414307 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Jan 26 17:05:40 crc kubenswrapper[4754]: I0126 17:05:40.414352 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Jan 26 17:05:40 crc kubenswrapper[4754]: I0126 17:05:40.559843 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Jan 26 17:05:40 crc kubenswrapper[4754]: I0126 17:05:40.560643 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Jan 26 17:05:40 crc kubenswrapper[4754]: I0126 17:05:40.562715 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Jan 26 17:05:40 crc kubenswrapper[4754]: I0126 17:05:40.562883 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Jan 26 17:05:40 crc kubenswrapper[4754]: I0126 17:05:40.563050 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-x88fz" Jan 26 17:05:40 crc kubenswrapper[4754]: I0126 17:05:40.588015 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Jan 26 17:05:40 crc kubenswrapper[4754]: I0126 17:05:40.601586 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/242703dd-5a52-4da8-af40-47c3490fd6ea-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"242703dd-5a52-4da8-af40-47c3490fd6ea\") " pod="openstack/openstack-cell1-galera-0" Jan 26 17:05:40 crc kubenswrapper[4754]: I0126 17:05:40.601644 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"openstack-cell1-galera-0\" (UID: \"242703dd-5a52-4da8-af40-47c3490fd6ea\") " pod="openstack/openstack-cell1-galera-0" Jan 26 17:05:40 crc kubenswrapper[4754]: I0126 17:05:40.601685 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/242703dd-5a52-4da8-af40-47c3490fd6ea-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"242703dd-5a52-4da8-af40-47c3490fd6ea\") " pod="openstack/openstack-cell1-galera-0" Jan 26 17:05:40 crc kubenswrapper[4754]: I0126 17:05:40.601701 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/242703dd-5a52-4da8-af40-47c3490fd6ea-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"242703dd-5a52-4da8-af40-47c3490fd6ea\") " pod="openstack/openstack-cell1-galera-0" Jan 26 17:05:40 crc kubenswrapper[4754]: I0126 17:05:40.601727 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wmc6g\" (UniqueName: \"kubernetes.io/projected/242703dd-5a52-4da8-af40-47c3490fd6ea-kube-api-access-wmc6g\") pod \"openstack-cell1-galera-0\" (UID: \"242703dd-5a52-4da8-af40-47c3490fd6ea\") " pod="openstack/openstack-cell1-galera-0" Jan 26 17:05:40 crc kubenswrapper[4754]: I0126 17:05:40.601755 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/242703dd-5a52-4da8-af40-47c3490fd6ea-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"242703dd-5a52-4da8-af40-47c3490fd6ea\") " pod="openstack/openstack-cell1-galera-0" Jan 26 17:05:40 crc kubenswrapper[4754]: I0126 17:05:40.601770 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/242703dd-5a52-4da8-af40-47c3490fd6ea-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"242703dd-5a52-4da8-af40-47c3490fd6ea\") " pod="openstack/openstack-cell1-galera-0" Jan 26 17:05:40 crc kubenswrapper[4754]: I0126 17:05:40.601824 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/242703dd-5a52-4da8-af40-47c3490fd6ea-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"242703dd-5a52-4da8-af40-47c3490fd6ea\") " pod="openstack/openstack-cell1-galera-0" Jan 26 17:05:40 crc kubenswrapper[4754]: I0126 17:05:40.710161 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/242703dd-5a52-4da8-af40-47c3490fd6ea-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"242703dd-5a52-4da8-af40-47c3490fd6ea\") " pod="openstack/openstack-cell1-galera-0" Jan 26 17:05:40 crc kubenswrapper[4754]: I0126 17:05:40.710302 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/242703dd-5a52-4da8-af40-47c3490fd6ea-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"242703dd-5a52-4da8-af40-47c3490fd6ea\") " pod="openstack/openstack-cell1-galera-0" Jan 26 17:05:40 crc kubenswrapper[4754]: I0126 17:05:40.710385 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/242703dd-5a52-4da8-af40-47c3490fd6ea-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"242703dd-5a52-4da8-af40-47c3490fd6ea\") " pod="openstack/openstack-cell1-galera-0" Jan 26 17:05:40 crc kubenswrapper[4754]: I0126 17:05:40.711192 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wmc6g\" (UniqueName: \"kubernetes.io/projected/242703dd-5a52-4da8-af40-47c3490fd6ea-kube-api-access-wmc6g\") pod \"openstack-cell1-galera-0\" (UID: \"242703dd-5a52-4da8-af40-47c3490fd6ea\") " pod="openstack/openstack-cell1-galera-0" Jan 26 17:05:40 crc kubenswrapper[4754]: I0126 17:05:40.711274 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/242703dd-5a52-4da8-af40-47c3490fd6ea-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"242703dd-5a52-4da8-af40-47c3490fd6ea\") " pod="openstack/openstack-cell1-galera-0" Jan 26 17:05:40 crc kubenswrapper[4754]: I0126 17:05:40.711298 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/242703dd-5a52-4da8-af40-47c3490fd6ea-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"242703dd-5a52-4da8-af40-47c3490fd6ea\") " pod="openstack/openstack-cell1-galera-0" Jan 26 17:05:40 crc kubenswrapper[4754]: I0126 17:05:40.711325 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zj9bp\" (UniqueName: \"kubernetes.io/projected/6fbed1d4-2177-40ba-a3c6-03de6fc2484f-kube-api-access-zj9bp\") pod \"memcached-0\" (UID: \"6fbed1d4-2177-40ba-a3c6-03de6fc2484f\") " pod="openstack/memcached-0" Jan 26 17:05:40 crc kubenswrapper[4754]: I0126 17:05:40.711398 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/6fbed1d4-2177-40ba-a3c6-03de6fc2484f-memcached-tls-certs\") pod \"memcached-0\" (UID: \"6fbed1d4-2177-40ba-a3c6-03de6fc2484f\") " pod="openstack/memcached-0" Jan 26 17:05:40 crc kubenswrapper[4754]: I0126 17:05:40.711443 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6fbed1d4-2177-40ba-a3c6-03de6fc2484f-config-data\") pod \"memcached-0\" (UID: \"6fbed1d4-2177-40ba-a3c6-03de6fc2484f\") " pod="openstack/memcached-0" Jan 26 17:05:40 crc kubenswrapper[4754]: I0126 17:05:40.711480 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/242703dd-5a52-4da8-af40-47c3490fd6ea-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"242703dd-5a52-4da8-af40-47c3490fd6ea\") " pod="openstack/openstack-cell1-galera-0" Jan 26 17:05:40 crc kubenswrapper[4754]: I0126 17:05:40.711495 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/6fbed1d4-2177-40ba-a3c6-03de6fc2484f-kolla-config\") pod \"memcached-0\" (UID: \"6fbed1d4-2177-40ba-a3c6-03de6fc2484f\") " pod="openstack/memcached-0" Jan 26 17:05:40 crc kubenswrapper[4754]: I0126 17:05:40.711612 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/242703dd-5a52-4da8-af40-47c3490fd6ea-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"242703dd-5a52-4da8-af40-47c3490fd6ea\") " pod="openstack/openstack-cell1-galera-0" Jan 26 17:05:40 crc kubenswrapper[4754]: I0126 17:05:40.711633 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6fbed1d4-2177-40ba-a3c6-03de6fc2484f-combined-ca-bundle\") pod \"memcached-0\" (UID: \"6fbed1d4-2177-40ba-a3c6-03de6fc2484f\") " pod="openstack/memcached-0" Jan 26 17:05:40 crc kubenswrapper[4754]: I0126 17:05:40.711652 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"openstack-cell1-galera-0\" (UID: \"242703dd-5a52-4da8-af40-47c3490fd6ea\") " pod="openstack/openstack-cell1-galera-0" Jan 26 17:05:40 crc kubenswrapper[4754]: I0126 17:05:40.711894 4754 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"openstack-cell1-galera-0\" (UID: \"242703dd-5a52-4da8-af40-47c3490fd6ea\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/openstack-cell1-galera-0" Jan 26 17:05:40 crc kubenswrapper[4754]: I0126 17:05:40.715961 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/242703dd-5a52-4da8-af40-47c3490fd6ea-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"242703dd-5a52-4da8-af40-47c3490fd6ea\") " pod="openstack/openstack-cell1-galera-0" Jan 26 17:05:40 crc kubenswrapper[4754]: I0126 17:05:40.716266 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/242703dd-5a52-4da8-af40-47c3490fd6ea-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"242703dd-5a52-4da8-af40-47c3490fd6ea\") " pod="openstack/openstack-cell1-galera-0" Jan 26 17:05:40 crc kubenswrapper[4754]: I0126 17:05:40.716741 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/242703dd-5a52-4da8-af40-47c3490fd6ea-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"242703dd-5a52-4da8-af40-47c3490fd6ea\") " pod="openstack/openstack-cell1-galera-0" Jan 26 17:05:40 crc kubenswrapper[4754]: I0126 17:05:40.723033 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/242703dd-5a52-4da8-af40-47c3490fd6ea-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"242703dd-5a52-4da8-af40-47c3490fd6ea\") " pod="openstack/openstack-cell1-galera-0" Jan 26 17:05:40 crc kubenswrapper[4754]: I0126 17:05:40.726376 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/242703dd-5a52-4da8-af40-47c3490fd6ea-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"242703dd-5a52-4da8-af40-47c3490fd6ea\") " pod="openstack/openstack-cell1-galera-0" Jan 26 17:05:40 crc kubenswrapper[4754]: I0126 17:05:40.741315 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wmc6g\" (UniqueName: \"kubernetes.io/projected/242703dd-5a52-4da8-af40-47c3490fd6ea-kube-api-access-wmc6g\") pod \"openstack-cell1-galera-0\" (UID: \"242703dd-5a52-4da8-af40-47c3490fd6ea\") " pod="openstack/openstack-cell1-galera-0" Jan 26 17:05:40 crc kubenswrapper[4754]: I0126 17:05:40.750890 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"openstack-cell1-galera-0\" (UID: \"242703dd-5a52-4da8-af40-47c3490fd6ea\") " pod="openstack/openstack-cell1-galera-0" Jan 26 17:05:40 crc kubenswrapper[4754]: I0126 17:05:40.812606 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zj9bp\" (UniqueName: \"kubernetes.io/projected/6fbed1d4-2177-40ba-a3c6-03de6fc2484f-kube-api-access-zj9bp\") pod \"memcached-0\" (UID: \"6fbed1d4-2177-40ba-a3c6-03de6fc2484f\") " pod="openstack/memcached-0" Jan 26 17:05:40 crc kubenswrapper[4754]: I0126 17:05:40.812696 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/6fbed1d4-2177-40ba-a3c6-03de6fc2484f-memcached-tls-certs\") pod \"memcached-0\" (UID: \"6fbed1d4-2177-40ba-a3c6-03de6fc2484f\") " pod="openstack/memcached-0" Jan 26 17:05:40 crc kubenswrapper[4754]: I0126 17:05:40.812729 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6fbed1d4-2177-40ba-a3c6-03de6fc2484f-config-data\") pod \"memcached-0\" (UID: \"6fbed1d4-2177-40ba-a3c6-03de6fc2484f\") " pod="openstack/memcached-0" Jan 26 17:05:40 crc kubenswrapper[4754]: I0126 17:05:40.812756 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/6fbed1d4-2177-40ba-a3c6-03de6fc2484f-kolla-config\") pod \"memcached-0\" (UID: \"6fbed1d4-2177-40ba-a3c6-03de6fc2484f\") " pod="openstack/memcached-0" Jan 26 17:05:40 crc kubenswrapper[4754]: I0126 17:05:40.812808 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6fbed1d4-2177-40ba-a3c6-03de6fc2484f-combined-ca-bundle\") pod \"memcached-0\" (UID: \"6fbed1d4-2177-40ba-a3c6-03de6fc2484f\") " pod="openstack/memcached-0" Jan 26 17:05:40 crc kubenswrapper[4754]: I0126 17:05:40.816592 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6fbed1d4-2177-40ba-a3c6-03de6fc2484f-config-data\") pod \"memcached-0\" (UID: \"6fbed1d4-2177-40ba-a3c6-03de6fc2484f\") " pod="openstack/memcached-0" Jan 26 17:05:40 crc kubenswrapper[4754]: I0126 17:05:40.816756 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/6fbed1d4-2177-40ba-a3c6-03de6fc2484f-kolla-config\") pod \"memcached-0\" (UID: \"6fbed1d4-2177-40ba-a3c6-03de6fc2484f\") " pod="openstack/memcached-0" Jan 26 17:05:40 crc kubenswrapper[4754]: I0126 17:05:40.820719 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/6fbed1d4-2177-40ba-a3c6-03de6fc2484f-memcached-tls-certs\") pod \"memcached-0\" (UID: \"6fbed1d4-2177-40ba-a3c6-03de6fc2484f\") " pod="openstack/memcached-0" Jan 26 17:05:40 crc kubenswrapper[4754]: I0126 17:05:40.822326 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6fbed1d4-2177-40ba-a3c6-03de6fc2484f-combined-ca-bundle\") pod \"memcached-0\" (UID: \"6fbed1d4-2177-40ba-a3c6-03de6fc2484f\") " pod="openstack/memcached-0" Jan 26 17:05:40 crc kubenswrapper[4754]: I0126 17:05:40.841420 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zj9bp\" (UniqueName: \"kubernetes.io/projected/6fbed1d4-2177-40ba-a3c6-03de6fc2484f-kube-api-access-zj9bp\") pod \"memcached-0\" (UID: \"6fbed1d4-2177-40ba-a3c6-03de6fc2484f\") " pod="openstack/memcached-0" Jan 26 17:05:40 crc kubenswrapper[4754]: I0126 17:05:40.895572 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Jan 26 17:05:41 crc kubenswrapper[4754]: I0126 17:05:41.039886 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Jan 26 17:05:42 crc kubenswrapper[4754]: I0126 17:05:42.551915 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Jan 26 17:05:42 crc kubenswrapper[4754]: I0126 17:05:42.552833 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Jan 26 17:05:42 crc kubenswrapper[4754]: I0126 17:05:42.555637 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-ll776" Jan 26 17:05:42 crc kubenswrapper[4754]: I0126 17:05:42.563523 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Jan 26 17:05:42 crc kubenswrapper[4754]: I0126 17:05:42.745048 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5lddx\" (UniqueName: \"kubernetes.io/projected/c3c55551-6de7-403e-bbf5-46a2ec685c13-kube-api-access-5lddx\") pod \"kube-state-metrics-0\" (UID: \"c3c55551-6de7-403e-bbf5-46a2ec685c13\") " pod="openstack/kube-state-metrics-0" Jan 26 17:05:42 crc kubenswrapper[4754]: I0126 17:05:42.850121 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5lddx\" (UniqueName: \"kubernetes.io/projected/c3c55551-6de7-403e-bbf5-46a2ec685c13-kube-api-access-5lddx\") pod \"kube-state-metrics-0\" (UID: \"c3c55551-6de7-403e-bbf5-46a2ec685c13\") " pod="openstack/kube-state-metrics-0" Jan 26 17:05:42 crc kubenswrapper[4754]: I0126 17:05:42.881362 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5lddx\" (UniqueName: \"kubernetes.io/projected/c3c55551-6de7-403e-bbf5-46a2ec685c13-kube-api-access-5lddx\") pod \"kube-state-metrics-0\" (UID: \"c3c55551-6de7-403e-bbf5-46a2ec685c13\") " pod="openstack/kube-state-metrics-0" Jan 26 17:05:43 crc kubenswrapper[4754]: I0126 17:05:43.168746 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Jan 26 17:05:46 crc kubenswrapper[4754]: I0126 17:05:46.248130 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-8bgjb"] Jan 26 17:05:46 crc kubenswrapper[4754]: I0126 17:05:46.249569 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-8bgjb" Jan 26 17:05:46 crc kubenswrapper[4754]: I0126 17:05:46.253136 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Jan 26 17:05:46 crc kubenswrapper[4754]: I0126 17:05:46.253218 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-dpmt7" Jan 26 17:05:46 crc kubenswrapper[4754]: I0126 17:05:46.253459 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Jan 26 17:05:46 crc kubenswrapper[4754]: I0126 17:05:46.255436 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-8bgjb"] Jan 26 17:05:46 crc kubenswrapper[4754]: I0126 17:05:46.303094 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-7psgg"] Jan 26 17:05:46 crc kubenswrapper[4754]: I0126 17:05:46.304553 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-7psgg" Jan 26 17:05:46 crc kubenswrapper[4754]: I0126 17:05:46.309983 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/ee1b9e37-8fd3-4280-af09-9f2f45366870-etc-ovs\") pod \"ovn-controller-ovs-7psgg\" (UID: \"ee1b9e37-8fd3-4280-af09-9f2f45366870\") " pod="openstack/ovn-controller-ovs-7psgg" Jan 26 17:05:46 crc kubenswrapper[4754]: I0126 17:05:46.310027 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/ee1b9e37-8fd3-4280-af09-9f2f45366870-var-run\") pod \"ovn-controller-ovs-7psgg\" (UID: \"ee1b9e37-8fd3-4280-af09-9f2f45366870\") " pod="openstack/ovn-controller-ovs-7psgg" Jan 26 17:05:46 crc kubenswrapper[4754]: I0126 17:05:46.310065 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/6db16a49-6566-42bf-91ad-c34be46e7800-var-run-ovn\") pod \"ovn-controller-8bgjb\" (UID: \"6db16a49-6566-42bf-91ad-c34be46e7800\") " pod="openstack/ovn-controller-8bgjb" Jan 26 17:05:46 crc kubenswrapper[4754]: I0126 17:05:46.310155 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/6db16a49-6566-42bf-91ad-c34be46e7800-var-run\") pod \"ovn-controller-8bgjb\" (UID: \"6db16a49-6566-42bf-91ad-c34be46e7800\") " pod="openstack/ovn-controller-8bgjb" Jan 26 17:05:46 crc kubenswrapper[4754]: I0126 17:05:46.310199 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6db16a49-6566-42bf-91ad-c34be46e7800-scripts\") pod \"ovn-controller-8bgjb\" (UID: \"6db16a49-6566-42bf-91ad-c34be46e7800\") " pod="openstack/ovn-controller-8bgjb" Jan 26 17:05:46 crc kubenswrapper[4754]: I0126 17:05:46.310227 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hxjk8\" (UniqueName: \"kubernetes.io/projected/ee1b9e37-8fd3-4280-af09-9f2f45366870-kube-api-access-hxjk8\") pod \"ovn-controller-ovs-7psgg\" (UID: \"ee1b9e37-8fd3-4280-af09-9f2f45366870\") " pod="openstack/ovn-controller-ovs-7psgg" Jan 26 17:05:46 crc kubenswrapper[4754]: I0126 17:05:46.310284 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xs8dr\" (UniqueName: \"kubernetes.io/projected/6db16a49-6566-42bf-91ad-c34be46e7800-kube-api-access-xs8dr\") pod \"ovn-controller-8bgjb\" (UID: \"6db16a49-6566-42bf-91ad-c34be46e7800\") " pod="openstack/ovn-controller-8bgjb" Jan 26 17:05:46 crc kubenswrapper[4754]: I0126 17:05:46.310342 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6db16a49-6566-42bf-91ad-c34be46e7800-combined-ca-bundle\") pod \"ovn-controller-8bgjb\" (UID: \"6db16a49-6566-42bf-91ad-c34be46e7800\") " pod="openstack/ovn-controller-8bgjb" Jan 26 17:05:46 crc kubenswrapper[4754]: I0126 17:05:46.310381 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/ee1b9e37-8fd3-4280-af09-9f2f45366870-var-lib\") pod \"ovn-controller-ovs-7psgg\" (UID: \"ee1b9e37-8fd3-4280-af09-9f2f45366870\") " pod="openstack/ovn-controller-ovs-7psgg" Jan 26 17:05:46 crc kubenswrapper[4754]: I0126 17:05:46.310405 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/6db16a49-6566-42bf-91ad-c34be46e7800-var-log-ovn\") pod \"ovn-controller-8bgjb\" (UID: \"6db16a49-6566-42bf-91ad-c34be46e7800\") " pod="openstack/ovn-controller-8bgjb" Jan 26 17:05:46 crc kubenswrapper[4754]: I0126 17:05:46.310441 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/6db16a49-6566-42bf-91ad-c34be46e7800-ovn-controller-tls-certs\") pod \"ovn-controller-8bgjb\" (UID: \"6db16a49-6566-42bf-91ad-c34be46e7800\") " pod="openstack/ovn-controller-8bgjb" Jan 26 17:05:46 crc kubenswrapper[4754]: I0126 17:05:46.310518 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/ee1b9e37-8fd3-4280-af09-9f2f45366870-var-log\") pod \"ovn-controller-ovs-7psgg\" (UID: \"ee1b9e37-8fd3-4280-af09-9f2f45366870\") " pod="openstack/ovn-controller-ovs-7psgg" Jan 26 17:05:46 crc kubenswrapper[4754]: I0126 17:05:46.310538 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ee1b9e37-8fd3-4280-af09-9f2f45366870-scripts\") pod \"ovn-controller-ovs-7psgg\" (UID: \"ee1b9e37-8fd3-4280-af09-9f2f45366870\") " pod="openstack/ovn-controller-ovs-7psgg" Jan 26 17:05:46 crc kubenswrapper[4754]: I0126 17:05:46.354194 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-7psgg"] Jan 26 17:05:46 crc kubenswrapper[4754]: I0126 17:05:46.411324 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/ee1b9e37-8fd3-4280-af09-9f2f45366870-etc-ovs\") pod \"ovn-controller-ovs-7psgg\" (UID: \"ee1b9e37-8fd3-4280-af09-9f2f45366870\") " pod="openstack/ovn-controller-ovs-7psgg" Jan 26 17:05:46 crc kubenswrapper[4754]: I0126 17:05:46.411364 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/ee1b9e37-8fd3-4280-af09-9f2f45366870-var-run\") pod \"ovn-controller-ovs-7psgg\" (UID: \"ee1b9e37-8fd3-4280-af09-9f2f45366870\") " pod="openstack/ovn-controller-ovs-7psgg" Jan 26 17:05:46 crc kubenswrapper[4754]: I0126 17:05:46.411525 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/6db16a49-6566-42bf-91ad-c34be46e7800-var-run-ovn\") pod \"ovn-controller-8bgjb\" (UID: \"6db16a49-6566-42bf-91ad-c34be46e7800\") " pod="openstack/ovn-controller-8bgjb" Jan 26 17:05:46 crc kubenswrapper[4754]: I0126 17:05:46.411608 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/6db16a49-6566-42bf-91ad-c34be46e7800-var-run\") pod \"ovn-controller-8bgjb\" (UID: \"6db16a49-6566-42bf-91ad-c34be46e7800\") " pod="openstack/ovn-controller-8bgjb" Jan 26 17:05:46 crc kubenswrapper[4754]: I0126 17:05:46.411652 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6db16a49-6566-42bf-91ad-c34be46e7800-scripts\") pod \"ovn-controller-8bgjb\" (UID: \"6db16a49-6566-42bf-91ad-c34be46e7800\") " pod="openstack/ovn-controller-8bgjb" Jan 26 17:05:46 crc kubenswrapper[4754]: I0126 17:05:46.411845 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/ee1b9e37-8fd3-4280-af09-9f2f45366870-var-run\") pod \"ovn-controller-ovs-7psgg\" (UID: \"ee1b9e37-8fd3-4280-af09-9f2f45366870\") " pod="openstack/ovn-controller-ovs-7psgg" Jan 26 17:05:46 crc kubenswrapper[4754]: I0126 17:05:46.411963 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/6db16a49-6566-42bf-91ad-c34be46e7800-var-run-ovn\") pod \"ovn-controller-8bgjb\" (UID: \"6db16a49-6566-42bf-91ad-c34be46e7800\") " pod="openstack/ovn-controller-8bgjb" Jan 26 17:05:46 crc kubenswrapper[4754]: I0126 17:05:46.411977 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/ee1b9e37-8fd3-4280-af09-9f2f45366870-etc-ovs\") pod \"ovn-controller-ovs-7psgg\" (UID: \"ee1b9e37-8fd3-4280-af09-9f2f45366870\") " pod="openstack/ovn-controller-ovs-7psgg" Jan 26 17:05:46 crc kubenswrapper[4754]: I0126 17:05:46.412121 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/6db16a49-6566-42bf-91ad-c34be46e7800-var-run\") pod \"ovn-controller-8bgjb\" (UID: \"6db16a49-6566-42bf-91ad-c34be46e7800\") " pod="openstack/ovn-controller-8bgjb" Jan 26 17:05:46 crc kubenswrapper[4754]: I0126 17:05:46.413727 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6db16a49-6566-42bf-91ad-c34be46e7800-scripts\") pod \"ovn-controller-8bgjb\" (UID: \"6db16a49-6566-42bf-91ad-c34be46e7800\") " pod="openstack/ovn-controller-8bgjb" Jan 26 17:05:46 crc kubenswrapper[4754]: I0126 17:05:46.413729 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hxjk8\" (UniqueName: \"kubernetes.io/projected/ee1b9e37-8fd3-4280-af09-9f2f45366870-kube-api-access-hxjk8\") pod \"ovn-controller-ovs-7psgg\" (UID: \"ee1b9e37-8fd3-4280-af09-9f2f45366870\") " pod="openstack/ovn-controller-ovs-7psgg" Jan 26 17:05:46 crc kubenswrapper[4754]: I0126 17:05:46.413804 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xs8dr\" (UniqueName: \"kubernetes.io/projected/6db16a49-6566-42bf-91ad-c34be46e7800-kube-api-access-xs8dr\") pod \"ovn-controller-8bgjb\" (UID: \"6db16a49-6566-42bf-91ad-c34be46e7800\") " pod="openstack/ovn-controller-8bgjb" Jan 26 17:05:46 crc kubenswrapper[4754]: I0126 17:05:46.413850 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6db16a49-6566-42bf-91ad-c34be46e7800-combined-ca-bundle\") pod \"ovn-controller-8bgjb\" (UID: \"6db16a49-6566-42bf-91ad-c34be46e7800\") " pod="openstack/ovn-controller-8bgjb" Jan 26 17:05:46 crc kubenswrapper[4754]: I0126 17:05:46.413881 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/ee1b9e37-8fd3-4280-af09-9f2f45366870-var-lib\") pod \"ovn-controller-ovs-7psgg\" (UID: \"ee1b9e37-8fd3-4280-af09-9f2f45366870\") " pod="openstack/ovn-controller-ovs-7psgg" Jan 26 17:05:46 crc kubenswrapper[4754]: I0126 17:05:46.413905 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/6db16a49-6566-42bf-91ad-c34be46e7800-var-log-ovn\") pod \"ovn-controller-8bgjb\" (UID: \"6db16a49-6566-42bf-91ad-c34be46e7800\") " pod="openstack/ovn-controller-8bgjb" Jan 26 17:05:46 crc kubenswrapper[4754]: I0126 17:05:46.413934 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/6db16a49-6566-42bf-91ad-c34be46e7800-ovn-controller-tls-certs\") pod \"ovn-controller-8bgjb\" (UID: \"6db16a49-6566-42bf-91ad-c34be46e7800\") " pod="openstack/ovn-controller-8bgjb" Jan 26 17:05:46 crc kubenswrapper[4754]: I0126 17:05:46.413976 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/ee1b9e37-8fd3-4280-af09-9f2f45366870-var-log\") pod \"ovn-controller-ovs-7psgg\" (UID: \"ee1b9e37-8fd3-4280-af09-9f2f45366870\") " pod="openstack/ovn-controller-ovs-7psgg" Jan 26 17:05:46 crc kubenswrapper[4754]: I0126 17:05:46.413997 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ee1b9e37-8fd3-4280-af09-9f2f45366870-scripts\") pod \"ovn-controller-ovs-7psgg\" (UID: \"ee1b9e37-8fd3-4280-af09-9f2f45366870\") " pod="openstack/ovn-controller-ovs-7psgg" Jan 26 17:05:46 crc kubenswrapper[4754]: I0126 17:05:46.414168 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/6db16a49-6566-42bf-91ad-c34be46e7800-var-log-ovn\") pod \"ovn-controller-8bgjb\" (UID: \"6db16a49-6566-42bf-91ad-c34be46e7800\") " pod="openstack/ovn-controller-8bgjb" Jan 26 17:05:46 crc kubenswrapper[4754]: I0126 17:05:46.414254 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/ee1b9e37-8fd3-4280-af09-9f2f45366870-var-log\") pod \"ovn-controller-ovs-7psgg\" (UID: \"ee1b9e37-8fd3-4280-af09-9f2f45366870\") " pod="openstack/ovn-controller-ovs-7psgg" Jan 26 17:05:46 crc kubenswrapper[4754]: I0126 17:05:46.414725 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/ee1b9e37-8fd3-4280-af09-9f2f45366870-var-lib\") pod \"ovn-controller-ovs-7psgg\" (UID: \"ee1b9e37-8fd3-4280-af09-9f2f45366870\") " pod="openstack/ovn-controller-ovs-7psgg" Jan 26 17:05:46 crc kubenswrapper[4754]: I0126 17:05:46.415906 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ee1b9e37-8fd3-4280-af09-9f2f45366870-scripts\") pod \"ovn-controller-ovs-7psgg\" (UID: \"ee1b9e37-8fd3-4280-af09-9f2f45366870\") " pod="openstack/ovn-controller-ovs-7psgg" Jan 26 17:05:46 crc kubenswrapper[4754]: I0126 17:05:46.420583 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/6db16a49-6566-42bf-91ad-c34be46e7800-ovn-controller-tls-certs\") pod \"ovn-controller-8bgjb\" (UID: \"6db16a49-6566-42bf-91ad-c34be46e7800\") " pod="openstack/ovn-controller-8bgjb" Jan 26 17:05:46 crc kubenswrapper[4754]: I0126 17:05:46.429187 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6db16a49-6566-42bf-91ad-c34be46e7800-combined-ca-bundle\") pod \"ovn-controller-8bgjb\" (UID: \"6db16a49-6566-42bf-91ad-c34be46e7800\") " pod="openstack/ovn-controller-8bgjb" Jan 26 17:05:46 crc kubenswrapper[4754]: I0126 17:05:46.429694 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xs8dr\" (UniqueName: \"kubernetes.io/projected/6db16a49-6566-42bf-91ad-c34be46e7800-kube-api-access-xs8dr\") pod \"ovn-controller-8bgjb\" (UID: \"6db16a49-6566-42bf-91ad-c34be46e7800\") " pod="openstack/ovn-controller-8bgjb" Jan 26 17:05:46 crc kubenswrapper[4754]: I0126 17:05:46.436927 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hxjk8\" (UniqueName: \"kubernetes.io/projected/ee1b9e37-8fd3-4280-af09-9f2f45366870-kube-api-access-hxjk8\") pod \"ovn-controller-ovs-7psgg\" (UID: \"ee1b9e37-8fd3-4280-af09-9f2f45366870\") " pod="openstack/ovn-controller-ovs-7psgg" Jan 26 17:05:46 crc kubenswrapper[4754]: I0126 17:05:46.569761 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-8bgjb" Jan 26 17:05:46 crc kubenswrapper[4754]: I0126 17:05:46.617224 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-7psgg" Jan 26 17:05:46 crc kubenswrapper[4754]: I0126 17:05:46.723751 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Jan 26 17:05:46 crc kubenswrapper[4754]: I0126 17:05:46.724842 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Jan 26 17:05:46 crc kubenswrapper[4754]: I0126 17:05:46.727058 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Jan 26 17:05:46 crc kubenswrapper[4754]: I0126 17:05:46.727149 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Jan 26 17:05:46 crc kubenswrapper[4754]: I0126 17:05:46.727201 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Jan 26 17:05:46 crc kubenswrapper[4754]: I0126 17:05:46.727533 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-wfpnw" Jan 26 17:05:46 crc kubenswrapper[4754]: I0126 17:05:46.727550 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Jan 26 17:05:46 crc kubenswrapper[4754]: I0126 17:05:46.738086 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Jan 26 17:05:46 crc kubenswrapper[4754]: I0126 17:05:46.920912 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85c77631-974b-42b0-a934-268213691414-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"85c77631-974b-42b0-a934-268213691414\") " pod="openstack/ovsdbserver-nb-0" Jan 26 17:05:46 crc kubenswrapper[4754]: I0126 17:05:46.920971 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"ovsdbserver-nb-0\" (UID: \"85c77631-974b-42b0-a934-268213691414\") " pod="openstack/ovsdbserver-nb-0" Jan 26 17:05:46 crc kubenswrapper[4754]: I0126 17:05:46.920991 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/85c77631-974b-42b0-a934-268213691414-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"85c77631-974b-42b0-a934-268213691414\") " pod="openstack/ovsdbserver-nb-0" Jan 26 17:05:46 crc kubenswrapper[4754]: I0126 17:05:46.921046 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zgrvw\" (UniqueName: \"kubernetes.io/projected/85c77631-974b-42b0-a934-268213691414-kube-api-access-zgrvw\") pod \"ovsdbserver-nb-0\" (UID: \"85c77631-974b-42b0-a934-268213691414\") " pod="openstack/ovsdbserver-nb-0" Jan 26 17:05:46 crc kubenswrapper[4754]: I0126 17:05:46.921089 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/85c77631-974b-42b0-a934-268213691414-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"85c77631-974b-42b0-a934-268213691414\") " pod="openstack/ovsdbserver-nb-0" Jan 26 17:05:46 crc kubenswrapper[4754]: I0126 17:05:46.921116 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/85c77631-974b-42b0-a934-268213691414-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"85c77631-974b-42b0-a934-268213691414\") " pod="openstack/ovsdbserver-nb-0" Jan 26 17:05:46 crc kubenswrapper[4754]: I0126 17:05:46.921213 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/85c77631-974b-42b0-a934-268213691414-config\") pod \"ovsdbserver-nb-0\" (UID: \"85c77631-974b-42b0-a934-268213691414\") " pod="openstack/ovsdbserver-nb-0" Jan 26 17:05:46 crc kubenswrapper[4754]: I0126 17:05:46.921236 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/85c77631-974b-42b0-a934-268213691414-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"85c77631-974b-42b0-a934-268213691414\") " pod="openstack/ovsdbserver-nb-0" Jan 26 17:05:47 crc kubenswrapper[4754]: I0126 17:05:47.022634 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zgrvw\" (UniqueName: \"kubernetes.io/projected/85c77631-974b-42b0-a934-268213691414-kube-api-access-zgrvw\") pod \"ovsdbserver-nb-0\" (UID: \"85c77631-974b-42b0-a934-268213691414\") " pod="openstack/ovsdbserver-nb-0" Jan 26 17:05:47 crc kubenswrapper[4754]: I0126 17:05:47.022792 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/85c77631-974b-42b0-a934-268213691414-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"85c77631-974b-42b0-a934-268213691414\") " pod="openstack/ovsdbserver-nb-0" Jan 26 17:05:47 crc kubenswrapper[4754]: I0126 17:05:47.022866 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/85c77631-974b-42b0-a934-268213691414-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"85c77631-974b-42b0-a934-268213691414\") " pod="openstack/ovsdbserver-nb-0" Jan 26 17:05:47 crc kubenswrapper[4754]: I0126 17:05:47.022903 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/85c77631-974b-42b0-a934-268213691414-config\") pod \"ovsdbserver-nb-0\" (UID: \"85c77631-974b-42b0-a934-268213691414\") " pod="openstack/ovsdbserver-nb-0" Jan 26 17:05:47 crc kubenswrapper[4754]: I0126 17:05:47.022942 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/85c77631-974b-42b0-a934-268213691414-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"85c77631-974b-42b0-a934-268213691414\") " pod="openstack/ovsdbserver-nb-0" Jan 26 17:05:47 crc kubenswrapper[4754]: I0126 17:05:47.023015 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85c77631-974b-42b0-a934-268213691414-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"85c77631-974b-42b0-a934-268213691414\") " pod="openstack/ovsdbserver-nb-0" Jan 26 17:05:47 crc kubenswrapper[4754]: I0126 17:05:47.023065 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"ovsdbserver-nb-0\" (UID: \"85c77631-974b-42b0-a934-268213691414\") " pod="openstack/ovsdbserver-nb-0" Jan 26 17:05:47 crc kubenswrapper[4754]: I0126 17:05:47.023102 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/85c77631-974b-42b0-a934-268213691414-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"85c77631-974b-42b0-a934-268213691414\") " pod="openstack/ovsdbserver-nb-0" Jan 26 17:05:47 crc kubenswrapper[4754]: I0126 17:05:47.023425 4754 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"ovsdbserver-nb-0\" (UID: \"85c77631-974b-42b0-a934-268213691414\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/ovsdbserver-nb-0" Jan 26 17:05:47 crc kubenswrapper[4754]: I0126 17:05:47.023529 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/85c77631-974b-42b0-a934-268213691414-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"85c77631-974b-42b0-a934-268213691414\") " pod="openstack/ovsdbserver-nb-0" Jan 26 17:05:47 crc kubenswrapper[4754]: I0126 17:05:47.024268 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/85c77631-974b-42b0-a934-268213691414-config\") pod \"ovsdbserver-nb-0\" (UID: \"85c77631-974b-42b0-a934-268213691414\") " pod="openstack/ovsdbserver-nb-0" Jan 26 17:05:47 crc kubenswrapper[4754]: I0126 17:05:47.024826 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/85c77631-974b-42b0-a934-268213691414-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"85c77631-974b-42b0-a934-268213691414\") " pod="openstack/ovsdbserver-nb-0" Jan 26 17:05:47 crc kubenswrapper[4754]: I0126 17:05:47.026914 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/85c77631-974b-42b0-a934-268213691414-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"85c77631-974b-42b0-a934-268213691414\") " pod="openstack/ovsdbserver-nb-0" Jan 26 17:05:47 crc kubenswrapper[4754]: I0126 17:05:47.048510 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/85c77631-974b-42b0-a934-268213691414-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"85c77631-974b-42b0-a934-268213691414\") " pod="openstack/ovsdbserver-nb-0" Jan 26 17:05:47 crc kubenswrapper[4754]: I0126 17:05:47.048608 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85c77631-974b-42b0-a934-268213691414-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"85c77631-974b-42b0-a934-268213691414\") " pod="openstack/ovsdbserver-nb-0" Jan 26 17:05:47 crc kubenswrapper[4754]: I0126 17:05:47.053773 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"ovsdbserver-nb-0\" (UID: \"85c77631-974b-42b0-a934-268213691414\") " pod="openstack/ovsdbserver-nb-0" Jan 26 17:05:47 crc kubenswrapper[4754]: I0126 17:05:47.055481 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zgrvw\" (UniqueName: \"kubernetes.io/projected/85c77631-974b-42b0-a934-268213691414-kube-api-access-zgrvw\") pod \"ovsdbserver-nb-0\" (UID: \"85c77631-974b-42b0-a934-268213691414\") " pod="openstack/ovsdbserver-nb-0" Jan 26 17:05:47 crc kubenswrapper[4754]: I0126 17:05:47.346421 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Jan 26 17:05:50 crc kubenswrapper[4754]: I0126 17:05:50.379507 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Jan 26 17:05:50 crc kubenswrapper[4754]: I0126 17:05:50.381537 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Jan 26 17:05:50 crc kubenswrapper[4754]: I0126 17:05:50.386735 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Jan 26 17:05:50 crc kubenswrapper[4754]: I0126 17:05:50.386968 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Jan 26 17:05:50 crc kubenswrapper[4754]: I0126 17:05:50.387165 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-c98nz" Jan 26 17:05:50 crc kubenswrapper[4754]: I0126 17:05:50.387358 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Jan 26 17:05:50 crc kubenswrapper[4754]: I0126 17:05:50.393882 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Jan 26 17:05:50 crc kubenswrapper[4754]: I0126 17:05:50.491109 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/10afcb49-f3f2-4598-a8a8-45729720e109-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"10afcb49-f3f2-4598-a8a8-45729720e109\") " pod="openstack/ovsdbserver-sb-0" Jan 26 17:05:50 crc kubenswrapper[4754]: I0126 17:05:50.491163 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"ovsdbserver-sb-0\" (UID: \"10afcb49-f3f2-4598-a8a8-45729720e109\") " pod="openstack/ovsdbserver-sb-0" Jan 26 17:05:50 crc kubenswrapper[4754]: I0126 17:05:50.491210 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/10afcb49-f3f2-4598-a8a8-45729720e109-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"10afcb49-f3f2-4598-a8a8-45729720e109\") " pod="openstack/ovsdbserver-sb-0" Jan 26 17:05:50 crc kubenswrapper[4754]: I0126 17:05:50.491376 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-96mf5\" (UniqueName: \"kubernetes.io/projected/10afcb49-f3f2-4598-a8a8-45729720e109-kube-api-access-96mf5\") pod \"ovsdbserver-sb-0\" (UID: \"10afcb49-f3f2-4598-a8a8-45729720e109\") " pod="openstack/ovsdbserver-sb-0" Jan 26 17:05:50 crc kubenswrapper[4754]: I0126 17:05:50.491440 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/10afcb49-f3f2-4598-a8a8-45729720e109-config\") pod \"ovsdbserver-sb-0\" (UID: \"10afcb49-f3f2-4598-a8a8-45729720e109\") " pod="openstack/ovsdbserver-sb-0" Jan 26 17:05:50 crc kubenswrapper[4754]: I0126 17:05:50.491496 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10afcb49-f3f2-4598-a8a8-45729720e109-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"10afcb49-f3f2-4598-a8a8-45729720e109\") " pod="openstack/ovsdbserver-sb-0" Jan 26 17:05:50 crc kubenswrapper[4754]: I0126 17:05:50.491570 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/10afcb49-f3f2-4598-a8a8-45729720e109-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"10afcb49-f3f2-4598-a8a8-45729720e109\") " pod="openstack/ovsdbserver-sb-0" Jan 26 17:05:50 crc kubenswrapper[4754]: I0126 17:05:50.491614 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/10afcb49-f3f2-4598-a8a8-45729720e109-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"10afcb49-f3f2-4598-a8a8-45729720e109\") " pod="openstack/ovsdbserver-sb-0" Jan 26 17:05:50 crc kubenswrapper[4754]: I0126 17:05:50.593041 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/10afcb49-f3f2-4598-a8a8-45729720e109-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"10afcb49-f3f2-4598-a8a8-45729720e109\") " pod="openstack/ovsdbserver-sb-0" Jan 26 17:05:50 crc kubenswrapper[4754]: I0126 17:05:50.593108 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"ovsdbserver-sb-0\" (UID: \"10afcb49-f3f2-4598-a8a8-45729720e109\") " pod="openstack/ovsdbserver-sb-0" Jan 26 17:05:50 crc kubenswrapper[4754]: I0126 17:05:50.593192 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/10afcb49-f3f2-4598-a8a8-45729720e109-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"10afcb49-f3f2-4598-a8a8-45729720e109\") " pod="openstack/ovsdbserver-sb-0" Jan 26 17:05:50 crc kubenswrapper[4754]: I0126 17:05:50.593263 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-96mf5\" (UniqueName: \"kubernetes.io/projected/10afcb49-f3f2-4598-a8a8-45729720e109-kube-api-access-96mf5\") pod \"ovsdbserver-sb-0\" (UID: \"10afcb49-f3f2-4598-a8a8-45729720e109\") " pod="openstack/ovsdbserver-sb-0" Jan 26 17:05:50 crc kubenswrapper[4754]: I0126 17:05:50.593295 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/10afcb49-f3f2-4598-a8a8-45729720e109-config\") pod \"ovsdbserver-sb-0\" (UID: \"10afcb49-f3f2-4598-a8a8-45729720e109\") " pod="openstack/ovsdbserver-sb-0" Jan 26 17:05:50 crc kubenswrapper[4754]: I0126 17:05:50.593334 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10afcb49-f3f2-4598-a8a8-45729720e109-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"10afcb49-f3f2-4598-a8a8-45729720e109\") " pod="openstack/ovsdbserver-sb-0" Jan 26 17:05:50 crc kubenswrapper[4754]: I0126 17:05:50.593371 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/10afcb49-f3f2-4598-a8a8-45729720e109-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"10afcb49-f3f2-4598-a8a8-45729720e109\") " pod="openstack/ovsdbserver-sb-0" Jan 26 17:05:50 crc kubenswrapper[4754]: I0126 17:05:50.593422 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/10afcb49-f3f2-4598-a8a8-45729720e109-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"10afcb49-f3f2-4598-a8a8-45729720e109\") " pod="openstack/ovsdbserver-sb-0" Jan 26 17:05:50 crc kubenswrapper[4754]: I0126 17:05:50.593545 4754 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"ovsdbserver-sb-0\" (UID: \"10afcb49-f3f2-4598-a8a8-45729720e109\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/ovsdbserver-sb-0" Jan 26 17:05:50 crc kubenswrapper[4754]: I0126 17:05:50.594071 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/10afcb49-f3f2-4598-a8a8-45729720e109-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"10afcb49-f3f2-4598-a8a8-45729720e109\") " pod="openstack/ovsdbserver-sb-0" Jan 26 17:05:50 crc kubenswrapper[4754]: I0126 17:05:50.594475 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/10afcb49-f3f2-4598-a8a8-45729720e109-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"10afcb49-f3f2-4598-a8a8-45729720e109\") " pod="openstack/ovsdbserver-sb-0" Jan 26 17:05:50 crc kubenswrapper[4754]: I0126 17:05:50.594583 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/10afcb49-f3f2-4598-a8a8-45729720e109-config\") pod \"ovsdbserver-sb-0\" (UID: \"10afcb49-f3f2-4598-a8a8-45729720e109\") " pod="openstack/ovsdbserver-sb-0" Jan 26 17:05:50 crc kubenswrapper[4754]: I0126 17:05:50.598894 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/10afcb49-f3f2-4598-a8a8-45729720e109-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"10afcb49-f3f2-4598-a8a8-45729720e109\") " pod="openstack/ovsdbserver-sb-0" Jan 26 17:05:50 crc kubenswrapper[4754]: I0126 17:05:50.598939 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/10afcb49-f3f2-4598-a8a8-45729720e109-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"10afcb49-f3f2-4598-a8a8-45729720e109\") " pod="openstack/ovsdbserver-sb-0" Jan 26 17:05:50 crc kubenswrapper[4754]: I0126 17:05:50.606208 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10afcb49-f3f2-4598-a8a8-45729720e109-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"10afcb49-f3f2-4598-a8a8-45729720e109\") " pod="openstack/ovsdbserver-sb-0" Jan 26 17:05:50 crc kubenswrapper[4754]: I0126 17:05:50.611777 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-96mf5\" (UniqueName: \"kubernetes.io/projected/10afcb49-f3f2-4598-a8a8-45729720e109-kube-api-access-96mf5\") pod \"ovsdbserver-sb-0\" (UID: \"10afcb49-f3f2-4598-a8a8-45729720e109\") " pod="openstack/ovsdbserver-sb-0" Jan 26 17:05:50 crc kubenswrapper[4754]: I0126 17:05:50.629555 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"ovsdbserver-sb-0\" (UID: \"10afcb49-f3f2-4598-a8a8-45729720e109\") " pod="openstack/ovsdbserver-sb-0" Jan 26 17:05:50 crc kubenswrapper[4754]: I0126 17:05:50.705770 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Jan 26 17:05:57 crc kubenswrapper[4754]: E0126 17:05:57.375620 4754 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Jan 26 17:05:57 crc kubenswrapper[4754]: E0126 17:05:57.376076 4754 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-jbnkc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-675f4bcbfc-t9w6n_openstack(c097ce5f-d480-45fd-916b-8b73e3835ecb): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Jan 26 17:05:57 crc kubenswrapper[4754]: E0126 17:05:57.377393 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-675f4bcbfc-t9w6n" podUID="c097ce5f-d480-45fd-916b-8b73e3835ecb" Jan 26 17:05:58 crc kubenswrapper[4754]: E0126 17:05:58.019547 4754 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Jan 26 17:05:58 crc kubenswrapper[4754]: E0126 17:05:58.020036 4754 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n659h4h664hbh658h587h67ch89h587h8fh679hc6hf9h55fh644h5d5h698h68dh5cdh5ffh669h54ch9h689hb8hd4h5bfhd8h5d7h5fh665h574q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-98s2c,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-57d769cc4f-72h5n_openstack(a9ccad70-dcd4-473c-aa09-925915f7420f): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Jan 26 17:05:58 crc kubenswrapper[4754]: E0126 17:05:58.021220 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-57d769cc4f-72h5n" podUID="a9ccad70-dcd4-473c-aa09-925915f7420f" Jan 26 17:05:58 crc kubenswrapper[4754]: E0126 17:05:58.190653 4754 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Jan 26 17:05:58 crc kubenswrapper[4754]: E0126 17:05:58.190857 4754 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-wpfcl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-78dd6ddcc-cjsr7_openstack(66fcefe3-6133-4510-9887-20d407a64008): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Jan 26 17:05:58 crc kubenswrapper[4754]: E0126 17:05:58.192167 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-78dd6ddcc-cjsr7" podUID="66fcefe3-6133-4510-9887-20d407a64008" Jan 26 17:05:58 crc kubenswrapper[4754]: I0126 17:05:58.545816 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-t9w6n" Jan 26 17:05:58 crc kubenswrapper[4754]: E0126 17:05:58.562844 4754 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified" Jan 26 17:05:58 crc kubenswrapper[4754]: E0126 17:05:58.562997 4754 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:setup-container,Image:quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified,Command:[sh -c cp /tmp/erlang-cookie-secret/.erlang.cookie /var/lib/rabbitmq/.erlang.cookie && chmod 600 /var/lib/rabbitmq/.erlang.cookie ; cp /tmp/rabbitmq-plugins/enabled_plugins /operator/enabled_plugins ; echo '[default]' > /var/lib/rabbitmq/.rabbitmqadmin.conf && sed -e 's/default_user/username/' -e 's/default_pass/password/' /tmp/default_user.conf >> /var/lib/rabbitmq/.rabbitmqadmin.conf && chmod 600 /var/lib/rabbitmq/.rabbitmqadmin.conf ; sleep 30],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Requests:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:plugins-conf,ReadOnly:false,MountPath:/tmp/rabbitmq-plugins/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-erlang-cookie,ReadOnly:false,MountPath:/var/lib/rabbitmq/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:erlang-cookie-secret,ReadOnly:false,MountPath:/tmp/erlang-cookie-secret/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-plugins,ReadOnly:false,MountPath:/operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:persistence,ReadOnly:false,MountPath:/var/lib/rabbitmq/mnesia/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-confd,ReadOnly:false,MountPath:/tmp/default_user.conf,SubPath:default_user.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-w6twt,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-server-0_openstack(29be5081-5097-4cd1-b35b-192df0ce8faf): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Jan 26 17:05:58 crc kubenswrapper[4754]: E0126 17:05:58.567299 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/rabbitmq-server-0" podUID="29be5081-5097-4cd1-b35b-192df0ce8faf" Jan 26 17:05:58 crc kubenswrapper[4754]: E0126 17:05:58.590702 4754 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified" Jan 26 17:05:58 crc kubenswrapper[4754]: E0126 17:05:58.590882 4754 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:setup-container,Image:quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified,Command:[sh -c cp /tmp/erlang-cookie-secret/.erlang.cookie /var/lib/rabbitmq/.erlang.cookie && chmod 600 /var/lib/rabbitmq/.erlang.cookie ; cp /tmp/rabbitmq-plugins/enabled_plugins /operator/enabled_plugins ; echo '[default]' > /var/lib/rabbitmq/.rabbitmqadmin.conf && sed -e 's/default_user/username/' -e 's/default_pass/password/' /tmp/default_user.conf >> /var/lib/rabbitmq/.rabbitmqadmin.conf && chmod 600 /var/lib/rabbitmq/.rabbitmqadmin.conf ; sleep 30],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Requests:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:plugins-conf,ReadOnly:false,MountPath:/tmp/rabbitmq-plugins/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-erlang-cookie,ReadOnly:false,MountPath:/var/lib/rabbitmq/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:erlang-cookie-secret,ReadOnly:false,MountPath:/tmp/erlang-cookie-secret/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-plugins,ReadOnly:false,MountPath:/operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:persistence,ReadOnly:false,MountPath:/var/lib/rabbitmq/mnesia/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-confd,ReadOnly:false,MountPath:/tmp/default_user.conf,SubPath:default_user.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-p6m5j,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cell1-server-0_openstack(48810c5e-3c40-4cdc-8bab-47efa97e76fa): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Jan 26 17:05:58 crc kubenswrapper[4754]: E0126 17:05:58.592319 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/rabbitmq-cell1-server-0" podUID="48810c5e-3c40-4cdc-8bab-47efa97e76fa" Jan 26 17:05:58 crc kubenswrapper[4754]: I0126 17:05:58.630447 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-t9w6n" event={"ID":"c097ce5f-d480-45fd-916b-8b73e3835ecb","Type":"ContainerDied","Data":"ba18206021eff5bf10d8b195ce4483bc3f6726afd9f6af6e03fa9557d5722435"} Jan 26 17:05:58 crc kubenswrapper[4754]: I0126 17:05:58.630564 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-t9w6n" Jan 26 17:05:58 crc kubenswrapper[4754]: E0126 17:05:58.642167 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified\\\"\"" pod="openstack/dnsmasq-dns-57d769cc4f-72h5n" podUID="a9ccad70-dcd4-473c-aa09-925915f7420f" Jan 26 17:05:58 crc kubenswrapper[4754]: E0126 17:05:58.642209 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified\\\"\"" pod="openstack/rabbitmq-server-0" podUID="29be5081-5097-4cd1-b35b-192df0ce8faf" Jan 26 17:05:58 crc kubenswrapper[4754]: E0126 17:05:58.642543 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified\\\"\"" pod="openstack/rabbitmq-cell1-server-0" podUID="48810c5e-3c40-4cdc-8bab-47efa97e76fa" Jan 26 17:05:58 crc kubenswrapper[4754]: E0126 17:05:58.714999 4754 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Jan 26 17:05:58 crc kubenswrapper[4754]: E0126 17:05:58.715152 4754 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n68chd6h679hbfh55fhc6h5ffh5d8h94h56ch589hb4hc5h57bh677hcdh655h8dh667h675h654h66ch567h8fh659h5b4h675h566h55bh54h67dh6dq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-j9ww2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-666b6646f7-xtjb4_openstack(c4637fc0-04b8-4b18-8986-7c0cebe93d45): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Jan 26 17:05:58 crc kubenswrapper[4754]: E0126 17:05:58.716566 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-666b6646f7-xtjb4" podUID="c4637fc0-04b8-4b18-8986-7c0cebe93d45" Jan 26 17:05:58 crc kubenswrapper[4754]: I0126 17:05:58.741604 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jbnkc\" (UniqueName: \"kubernetes.io/projected/c097ce5f-d480-45fd-916b-8b73e3835ecb-kube-api-access-jbnkc\") pod \"c097ce5f-d480-45fd-916b-8b73e3835ecb\" (UID: \"c097ce5f-d480-45fd-916b-8b73e3835ecb\") " Jan 26 17:05:58 crc kubenswrapper[4754]: I0126 17:05:58.741765 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c097ce5f-d480-45fd-916b-8b73e3835ecb-config\") pod \"c097ce5f-d480-45fd-916b-8b73e3835ecb\" (UID: \"c097ce5f-d480-45fd-916b-8b73e3835ecb\") " Jan 26 17:05:58 crc kubenswrapper[4754]: I0126 17:05:58.743084 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c097ce5f-d480-45fd-916b-8b73e3835ecb-config" (OuterVolumeSpecName: "config") pod "c097ce5f-d480-45fd-916b-8b73e3835ecb" (UID: "c097ce5f-d480-45fd-916b-8b73e3835ecb"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:05:58 crc kubenswrapper[4754]: I0126 17:05:58.751002 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c097ce5f-d480-45fd-916b-8b73e3835ecb-kube-api-access-jbnkc" (OuterVolumeSpecName: "kube-api-access-jbnkc") pod "c097ce5f-d480-45fd-916b-8b73e3835ecb" (UID: "c097ce5f-d480-45fd-916b-8b73e3835ecb"). InnerVolumeSpecName "kube-api-access-jbnkc". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:05:58 crc kubenswrapper[4754]: I0126 17:05:58.844317 4754 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c097ce5f-d480-45fd-916b-8b73e3835ecb-config\") on node \"crc\" DevicePath \"\"" Jan 26 17:05:58 crc kubenswrapper[4754]: I0126 17:05:58.844637 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jbnkc\" (UniqueName: \"kubernetes.io/projected/c097ce5f-d480-45fd-916b-8b73e3835ecb-kube-api-access-jbnkc\") on node \"crc\" DevicePath \"\"" Jan 26 17:05:58 crc kubenswrapper[4754]: I0126 17:05:58.977855 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-t9w6n"] Jan 26 17:05:58 crc kubenswrapper[4754]: I0126 17:05:58.985295 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-t9w6n"] Jan 26 17:05:59 crc kubenswrapper[4754]: I0126 17:05:59.129126 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-8bgjb"] Jan 26 17:05:59 crc kubenswrapper[4754]: W0126 17:05:59.143054 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0390a195_4349_4dae_8ae2_72b9b16af4f7.slice/crio-46e09a2d2f7b449b9ee620a9674a7d32f4ff7638d44eed601a905a4a12f7185b WatchSource:0}: Error finding container 46e09a2d2f7b449b9ee620a9674a7d32f4ff7638d44eed601a905a4a12f7185b: Status 404 returned error can't find the container with id 46e09a2d2f7b449b9ee620a9674a7d32f4ff7638d44eed601a905a4a12f7185b Jan 26 17:05:59 crc kubenswrapper[4754]: I0126 17:05:59.143100 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Jan 26 17:05:59 crc kubenswrapper[4754]: W0126 17:05:59.145871 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc3c55551_6de7_403e_bbf5_46a2ec685c13.slice/crio-f43ccdd4b774f3fdbbf1ff52ffd4e67686633eb09604f68fd5c9a4b779e1f724 WatchSource:0}: Error finding container f43ccdd4b774f3fdbbf1ff52ffd4e67686633eb09604f68fd5c9a4b779e1f724: Status 404 returned error can't find the container with id f43ccdd4b774f3fdbbf1ff52ffd4e67686633eb09604f68fd5c9a4b779e1f724 Jan 26 17:05:59 crc kubenswrapper[4754]: I0126 17:05:59.150902 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Jan 26 17:05:59 crc kubenswrapper[4754]: I0126 17:05:59.190063 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-cjsr7" Jan 26 17:05:59 crc kubenswrapper[4754]: I0126 17:05:59.306783 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Jan 26 17:05:59 crc kubenswrapper[4754]: W0126 17:05:59.317065 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod242703dd_5a52_4da8_af40_47c3490fd6ea.slice/crio-16f6265e98fdca33bbaf41a24790b5ae40adf6e1297a94714543bc419426ebf2 WatchSource:0}: Error finding container 16f6265e98fdca33bbaf41a24790b5ae40adf6e1297a94714543bc419426ebf2: Status 404 returned error can't find the container with id 16f6265e98fdca33bbaf41a24790b5ae40adf6e1297a94714543bc419426ebf2 Jan 26 17:05:59 crc kubenswrapper[4754]: I0126 17:05:59.335010 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Jan 26 17:05:59 crc kubenswrapper[4754]: I0126 17:05:59.366262 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/66fcefe3-6133-4510-9887-20d407a64008-dns-svc\") pod \"66fcefe3-6133-4510-9887-20d407a64008\" (UID: \"66fcefe3-6133-4510-9887-20d407a64008\") " Jan 26 17:05:59 crc kubenswrapper[4754]: I0126 17:05:59.366573 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/66fcefe3-6133-4510-9887-20d407a64008-config\") pod \"66fcefe3-6133-4510-9887-20d407a64008\" (UID: \"66fcefe3-6133-4510-9887-20d407a64008\") " Jan 26 17:05:59 crc kubenswrapper[4754]: I0126 17:05:59.366752 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wpfcl\" (UniqueName: \"kubernetes.io/projected/66fcefe3-6133-4510-9887-20d407a64008-kube-api-access-wpfcl\") pod \"66fcefe3-6133-4510-9887-20d407a64008\" (UID: \"66fcefe3-6133-4510-9887-20d407a64008\") " Jan 26 17:05:59 crc kubenswrapper[4754]: I0126 17:05:59.367048 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/66fcefe3-6133-4510-9887-20d407a64008-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "66fcefe3-6133-4510-9887-20d407a64008" (UID: "66fcefe3-6133-4510-9887-20d407a64008"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:05:59 crc kubenswrapper[4754]: I0126 17:05:59.367435 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/66fcefe3-6133-4510-9887-20d407a64008-config" (OuterVolumeSpecName: "config") pod "66fcefe3-6133-4510-9887-20d407a64008" (UID: "66fcefe3-6133-4510-9887-20d407a64008"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:05:59 crc kubenswrapper[4754]: I0126 17:05:59.372487 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/66fcefe3-6133-4510-9887-20d407a64008-kube-api-access-wpfcl" (OuterVolumeSpecName: "kube-api-access-wpfcl") pod "66fcefe3-6133-4510-9887-20d407a64008" (UID: "66fcefe3-6133-4510-9887-20d407a64008"). InnerVolumeSpecName "kube-api-access-wpfcl". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:05:59 crc kubenswrapper[4754]: I0126 17:05:59.411084 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Jan 26 17:05:59 crc kubenswrapper[4754]: W0126 17:05:59.442695 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod85c77631_974b_42b0_a934_268213691414.slice/crio-63e70c72a681c18ebc85320908102bd84714007a57ffba997ad672b2e455ae84 WatchSource:0}: Error finding container 63e70c72a681c18ebc85320908102bd84714007a57ffba997ad672b2e455ae84: Status 404 returned error can't find the container with id 63e70c72a681c18ebc85320908102bd84714007a57ffba997ad672b2e455ae84 Jan 26 17:05:59 crc kubenswrapper[4754]: I0126 17:05:59.469217 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wpfcl\" (UniqueName: \"kubernetes.io/projected/66fcefe3-6133-4510-9887-20d407a64008-kube-api-access-wpfcl\") on node \"crc\" DevicePath \"\"" Jan 26 17:05:59 crc kubenswrapper[4754]: I0126 17:05:59.469257 4754 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/66fcefe3-6133-4510-9887-20d407a64008-dns-svc\") on node \"crc\" DevicePath \"\"" Jan 26 17:05:59 crc kubenswrapper[4754]: I0126 17:05:59.469269 4754 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/66fcefe3-6133-4510-9887-20d407a64008-config\") on node \"crc\" DevicePath \"\"" Jan 26 17:05:59 crc kubenswrapper[4754]: I0126 17:05:59.516108 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-7psgg"] Jan 26 17:05:59 crc kubenswrapper[4754]: I0126 17:05:59.637111 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"85c77631-974b-42b0-a934-268213691414","Type":"ContainerStarted","Data":"63e70c72a681c18ebc85320908102bd84714007a57ffba997ad672b2e455ae84"} Jan 26 17:05:59 crc kubenswrapper[4754]: I0126 17:05:59.639559 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"0390a195-4349-4dae-8ae2-72b9b16af4f7","Type":"ContainerStarted","Data":"46e09a2d2f7b449b9ee620a9674a7d32f4ff7638d44eed601a905a4a12f7185b"} Jan 26 17:05:59 crc kubenswrapper[4754]: I0126 17:05:59.640465 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-cjsr7" event={"ID":"66fcefe3-6133-4510-9887-20d407a64008","Type":"ContainerDied","Data":"a0666f0f41d1b9a748758cfa1335e9c8d12c3409001099fb7adfe4fa6f1804f9"} Jan 26 17:05:59 crc kubenswrapper[4754]: I0126 17:05:59.640651 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-cjsr7" Jan 26 17:05:59 crc kubenswrapper[4754]: I0126 17:05:59.644879 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"242703dd-5a52-4da8-af40-47c3490fd6ea","Type":"ContainerStarted","Data":"16f6265e98fdca33bbaf41a24790b5ae40adf6e1297a94714543bc419426ebf2"} Jan 26 17:05:59 crc kubenswrapper[4754]: I0126 17:05:59.646498 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"c3c55551-6de7-403e-bbf5-46a2ec685c13","Type":"ContainerStarted","Data":"f43ccdd4b774f3fdbbf1ff52ffd4e67686633eb09604f68fd5c9a4b779e1f724"} Jan 26 17:05:59 crc kubenswrapper[4754]: I0126 17:05:59.648812 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"6fbed1d4-2177-40ba-a3c6-03de6fc2484f","Type":"ContainerStarted","Data":"10cccbbf0bc90d90d8efe66262f98222e0284f40b138a67d5d15636306f6122a"} Jan 26 17:05:59 crc kubenswrapper[4754]: I0126 17:05:59.650258 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-7psgg" event={"ID":"ee1b9e37-8fd3-4280-af09-9f2f45366870","Type":"ContainerStarted","Data":"c753e78c6da0b398de729297c18a07a6e7eca5337e246c8bb9487d60e01321c3"} Jan 26 17:05:59 crc kubenswrapper[4754]: I0126 17:05:59.652431 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-8bgjb" event={"ID":"6db16a49-6566-42bf-91ad-c34be46e7800","Type":"ContainerStarted","Data":"6cc21628982e1250ece43fa0f2160c859ab6c5f93e4fbd326eca2c9ca5297c45"} Jan 26 17:05:59 crc kubenswrapper[4754]: E0126 17:05:59.653505 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified\\\"\"" pod="openstack/dnsmasq-dns-666b6646f7-xtjb4" podUID="c4637fc0-04b8-4b18-8986-7c0cebe93d45" Jan 26 17:05:59 crc kubenswrapper[4754]: I0126 17:05:59.715426 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-cjsr7"] Jan 26 17:05:59 crc kubenswrapper[4754]: I0126 17:05:59.721317 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-cjsr7"] Jan 26 17:05:59 crc kubenswrapper[4754]: I0126 17:05:59.776657 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="66fcefe3-6133-4510-9887-20d407a64008" path="/var/lib/kubelet/pods/66fcefe3-6133-4510-9887-20d407a64008/volumes" Jan 26 17:05:59 crc kubenswrapper[4754]: I0126 17:05:59.777126 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c097ce5f-d480-45fd-916b-8b73e3835ecb" path="/var/lib/kubelet/pods/c097ce5f-d480-45fd-916b-8b73e3835ecb/volumes" Jan 26 17:06:00 crc kubenswrapper[4754]: I0126 17:06:00.045258 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Jan 26 17:06:00 crc kubenswrapper[4754]: I0126 17:06:00.659634 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"10afcb49-f3f2-4598-a8a8-45729720e109","Type":"ContainerStarted","Data":"0457f9a1b9bb62bec247dd5ceaf99f64681983565213ed5bfdb564f925a48a0d"} Jan 26 17:06:07 crc kubenswrapper[4754]: I0126 17:06:07.129013 4754 patch_prober.go:28] interesting pod/machine-config-daemon-x65wv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 26 17:06:07 crc kubenswrapper[4754]: I0126 17:06:07.129572 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 26 17:06:07 crc kubenswrapper[4754]: I0126 17:06:07.705913 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"85c77631-974b-42b0-a934-268213691414","Type":"ContainerStarted","Data":"0541865575c716c8a44dddb427b98f9b69c73d191b3c5a3eb0aff0a080ee0235"} Jan 26 17:06:07 crc kubenswrapper[4754]: I0126 17:06:07.707348 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"0390a195-4349-4dae-8ae2-72b9b16af4f7","Type":"ContainerStarted","Data":"28e142c4a27b0b8177c57eb52f24474a8456092848b456c7b9dc7d9de1d258b5"} Jan 26 17:06:07 crc kubenswrapper[4754]: I0126 17:06:07.708435 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"242703dd-5a52-4da8-af40-47c3490fd6ea","Type":"ContainerStarted","Data":"696dbd70b22c45a5eef5c8cce1abd3ed82ecf6d4bdb841c819a76bfec023b290"} Jan 26 17:06:07 crc kubenswrapper[4754]: I0126 17:06:07.709607 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"6fbed1d4-2177-40ba-a3c6-03de6fc2484f","Type":"ContainerStarted","Data":"d735e62c9c4b5951e11aa46269ff253e211783f4f57d3b4008d1244247fb3891"} Jan 26 17:06:07 crc kubenswrapper[4754]: I0126 17:06:07.709723 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Jan 26 17:06:07 crc kubenswrapper[4754]: I0126 17:06:07.710767 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-7psgg" event={"ID":"ee1b9e37-8fd3-4280-af09-9f2f45366870","Type":"ContainerStarted","Data":"6c5de4b674ca5bb30c03050e5b5b13fbda5a4fb98de632a4733e594f15447f40"} Jan 26 17:06:07 crc kubenswrapper[4754]: I0126 17:06:07.711755 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"10afcb49-f3f2-4598-a8a8-45729720e109","Type":"ContainerStarted","Data":"245d6638fea6c9db1c7e48d418c3de25e49ceb0283765e05d076f6806477da0b"} Jan 26 17:06:07 crc kubenswrapper[4754]: I0126 17:06:07.798366 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=20.038227781 podStartE2EDuration="27.798349911s" podCreationTimestamp="2026-01-26 17:05:40 +0000 UTC" firstStartedPulling="2026-01-26 17:05:59.345032475 +0000 UTC m=+1125.869212909" lastFinishedPulling="2026-01-26 17:06:07.105154605 +0000 UTC m=+1133.629335039" observedRunningTime="2026-01-26 17:06:07.795913213 +0000 UTC m=+1134.320093647" watchObservedRunningTime="2026-01-26 17:06:07.798349911 +0000 UTC m=+1134.322530345" Jan 26 17:06:09 crc kubenswrapper[4754]: I0126 17:06:09.726978 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-8bgjb" event={"ID":"6db16a49-6566-42bf-91ad-c34be46e7800","Type":"ContainerStarted","Data":"3d904a8d43b7c0f8dc9cc082f11eaafbda6028d0460aaa8c961dba0994165f54"} Jan 26 17:06:09 crc kubenswrapper[4754]: I0126 17:06:09.727611 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-8bgjb" Jan 26 17:06:09 crc kubenswrapper[4754]: I0126 17:06:09.729730 4754 generic.go:334] "Generic (PLEG): container finished" podID="ee1b9e37-8fd3-4280-af09-9f2f45366870" containerID="6c5de4b674ca5bb30c03050e5b5b13fbda5a4fb98de632a4733e594f15447f40" exitCode=0 Jan 26 17:06:09 crc kubenswrapper[4754]: I0126 17:06:09.729800 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-7psgg" event={"ID":"ee1b9e37-8fd3-4280-af09-9f2f45366870","Type":"ContainerDied","Data":"6c5de4b674ca5bb30c03050e5b5b13fbda5a4fb98de632a4733e594f15447f40"} Jan 26 17:06:09 crc kubenswrapper[4754]: I0126 17:06:09.750246 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-8bgjb" podStartSLOduration=15.701596864 podStartE2EDuration="23.750229435s" podCreationTimestamp="2026-01-26 17:05:46 +0000 UTC" firstStartedPulling="2026-01-26 17:05:59.145925362 +0000 UTC m=+1125.670105796" lastFinishedPulling="2026-01-26 17:06:07.194557933 +0000 UTC m=+1133.718738367" observedRunningTime="2026-01-26 17:06:09.74716912 +0000 UTC m=+1136.271349574" watchObservedRunningTime="2026-01-26 17:06:09.750229435 +0000 UTC m=+1136.274409869" Jan 26 17:06:10 crc kubenswrapper[4754]: I0126 17:06:10.741045 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-7psgg" event={"ID":"ee1b9e37-8fd3-4280-af09-9f2f45366870","Type":"ContainerStarted","Data":"4d24b478f1c6f38c19c56cc1125a91bdad6b34db6eac68f64d78d8c0f2cf17c1"} Jan 26 17:06:10 crc kubenswrapper[4754]: I0126 17:06:10.741824 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-7psgg" event={"ID":"ee1b9e37-8fd3-4280-af09-9f2f45366870","Type":"ContainerStarted","Data":"8bc40cc7942ef69c306c44561a28ec4308ed3a7835e868c7ae86a57505ba2fbc"} Jan 26 17:06:10 crc kubenswrapper[4754]: I0126 17:06:10.742256 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-7psgg" Jan 26 17:06:10 crc kubenswrapper[4754]: I0126 17:06:10.768910 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-7psgg" podStartSLOduration=17.095607853 podStartE2EDuration="24.768888086s" podCreationTimestamp="2026-01-26 17:05:46 +0000 UTC" firstStartedPulling="2026-01-26 17:05:59.520144347 +0000 UTC m=+1126.044324781" lastFinishedPulling="2026-01-26 17:06:07.19342457 +0000 UTC m=+1133.717605014" observedRunningTime="2026-01-26 17:06:10.762755664 +0000 UTC m=+1137.286936098" watchObservedRunningTime="2026-01-26 17:06:10.768888086 +0000 UTC m=+1137.293068530" Jan 26 17:06:11 crc kubenswrapper[4754]: I0126 17:06:11.617557 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-7psgg" Jan 26 17:06:12 crc kubenswrapper[4754]: I0126 17:06:12.757639 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"85c77631-974b-42b0-a934-268213691414","Type":"ContainerStarted","Data":"2e3368fa575de0448d17b99f39a5f1a49aeb660c0a8755f9083a29de47fef15e"} Jan 26 17:06:12 crc kubenswrapper[4754]: I0126 17:06:12.760329 4754 generic.go:334] "Generic (PLEG): container finished" podID="0390a195-4349-4dae-8ae2-72b9b16af4f7" containerID="28e142c4a27b0b8177c57eb52f24474a8456092848b456c7b9dc7d9de1d258b5" exitCode=0 Jan 26 17:06:12 crc kubenswrapper[4754]: I0126 17:06:12.760491 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"0390a195-4349-4dae-8ae2-72b9b16af4f7","Type":"ContainerDied","Data":"28e142c4a27b0b8177c57eb52f24474a8456092848b456c7b9dc7d9de1d258b5"} Jan 26 17:06:12 crc kubenswrapper[4754]: I0126 17:06:12.763751 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"242703dd-5a52-4da8-af40-47c3490fd6ea","Type":"ContainerDied","Data":"696dbd70b22c45a5eef5c8cce1abd3ed82ecf6d4bdb841c819a76bfec023b290"} Jan 26 17:06:12 crc kubenswrapper[4754]: I0126 17:06:12.763644 4754 generic.go:334] "Generic (PLEG): container finished" podID="242703dd-5a52-4da8-af40-47c3490fd6ea" containerID="696dbd70b22c45a5eef5c8cce1abd3ed82ecf6d4bdb841c819a76bfec023b290" exitCode=0 Jan 26 17:06:12 crc kubenswrapper[4754]: I0126 17:06:12.766846 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"10afcb49-f3f2-4598-a8a8-45729720e109","Type":"ContainerStarted","Data":"2d345c149450a0bde5472520708c6a36aa23e85091c2d61b93ae3fb33b9e35ef"} Jan 26 17:06:12 crc kubenswrapper[4754]: I0126 17:06:12.806221 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=15.02599625 podStartE2EDuration="27.806188936s" podCreationTimestamp="2026-01-26 17:05:45 +0000 UTC" firstStartedPulling="2026-01-26 17:05:59.444884904 +0000 UTC m=+1125.969065338" lastFinishedPulling="2026-01-26 17:06:12.22507759 +0000 UTC m=+1138.749258024" observedRunningTime="2026-01-26 17:06:12.781168487 +0000 UTC m=+1139.305348941" watchObservedRunningTime="2026-01-26 17:06:12.806188936 +0000 UTC m=+1139.330369370" Jan 26 17:06:12 crc kubenswrapper[4754]: I0126 17:06:12.832104 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=11.64844862 podStartE2EDuration="23.832088379s" podCreationTimestamp="2026-01-26 17:05:49 +0000 UTC" firstStartedPulling="2026-01-26 17:06:00.052102499 +0000 UTC m=+1126.576282943" lastFinishedPulling="2026-01-26 17:06:12.235742238 +0000 UTC m=+1138.759922702" observedRunningTime="2026-01-26 17:06:12.822522953 +0000 UTC m=+1139.346703387" watchObservedRunningTime="2026-01-26 17:06:12.832088379 +0000 UTC m=+1139.356268813" Jan 26 17:06:13 crc kubenswrapper[4754]: I0126 17:06:13.777618 4754 generic.go:334] "Generic (PLEG): container finished" podID="a9ccad70-dcd4-473c-aa09-925915f7420f" containerID="9179318a0742938a4d3d8ad7707ca375118c0c3139a54b4fa56314d06e96104a" exitCode=0 Jan 26 17:06:13 crc kubenswrapper[4754]: I0126 17:06:13.780292 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-72h5n" event={"ID":"a9ccad70-dcd4-473c-aa09-925915f7420f","Type":"ContainerDied","Data":"9179318a0742938a4d3d8ad7707ca375118c0c3139a54b4fa56314d06e96104a"} Jan 26 17:06:13 crc kubenswrapper[4754]: I0126 17:06:13.780329 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"0390a195-4349-4dae-8ae2-72b9b16af4f7","Type":"ContainerStarted","Data":"a3de55de6c9c7ef2beb28ce15303171a65f2fe8b9bec7970825c05eeeab99e7c"} Jan 26 17:06:13 crc kubenswrapper[4754]: I0126 17:06:13.781154 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"29be5081-5097-4cd1-b35b-192df0ce8faf","Type":"ContainerStarted","Data":"3f67cb54c8ac7414b3b3fe5d6bc7d975fa28e160e0a72a1773a4721747eb4df3"} Jan 26 17:06:13 crc kubenswrapper[4754]: I0126 17:06:13.793787 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"242703dd-5a52-4da8-af40-47c3490fd6ea","Type":"ContainerStarted","Data":"09e73ca803087f77ae6154e0e4b7ad95a6eec0360b71e835d6a0eb5c8f3559c7"} Jan 26 17:06:13 crc kubenswrapper[4754]: I0126 17:06:13.858277 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=28.808130948 podStartE2EDuration="36.858248919s" podCreationTimestamp="2026-01-26 17:05:37 +0000 UTC" firstStartedPulling="2026-01-26 17:05:59.145914002 +0000 UTC m=+1125.670094436" lastFinishedPulling="2026-01-26 17:06:07.196031973 +0000 UTC m=+1133.720212407" observedRunningTime="2026-01-26 17:06:13.850536524 +0000 UTC m=+1140.374716998" watchObservedRunningTime="2026-01-26 17:06:13.858248919 +0000 UTC m=+1140.382429383" Jan 26 17:06:13 crc kubenswrapper[4754]: I0126 17:06:13.906157 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=27.121316188 podStartE2EDuration="34.906130788s" podCreationTimestamp="2026-01-26 17:05:39 +0000 UTC" firstStartedPulling="2026-01-26 17:05:59.320327424 +0000 UTC m=+1125.844507858" lastFinishedPulling="2026-01-26 17:06:07.105142024 +0000 UTC m=+1133.629322458" observedRunningTime="2026-01-26 17:06:13.893404952 +0000 UTC m=+1140.417585386" watchObservedRunningTime="2026-01-26 17:06:13.906130788 +0000 UTC m=+1140.430311222" Jan 26 17:06:14 crc kubenswrapper[4754]: I0126 17:06:14.348136 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Jan 26 17:06:14 crc kubenswrapper[4754]: I0126 17:06:14.395199 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Jan 26 17:06:14 crc kubenswrapper[4754]: I0126 17:06:14.707115 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Jan 26 17:06:14 crc kubenswrapper[4754]: I0126 17:06:14.740758 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Jan 26 17:06:14 crc kubenswrapper[4754]: I0126 17:06:14.802177 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-72h5n" event={"ID":"a9ccad70-dcd4-473c-aa09-925915f7420f","Type":"ContainerStarted","Data":"1d9caa5e5f0c0418fd0df5b7a9010ebb5b310f1dd2c3774dd558998962e06810"} Jan 26 17:06:14 crc kubenswrapper[4754]: I0126 17:06:14.802382 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-57d769cc4f-72h5n" Jan 26 17:06:14 crc kubenswrapper[4754]: I0126 17:06:14.803776 4754 generic.go:334] "Generic (PLEG): container finished" podID="c4637fc0-04b8-4b18-8986-7c0cebe93d45" containerID="9d89c6f811a67ef290616b6f7423814d0698c88e4fa418f456e2c0bded217af0" exitCode=0 Jan 26 17:06:14 crc kubenswrapper[4754]: I0126 17:06:14.803807 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-xtjb4" event={"ID":"c4637fc0-04b8-4b18-8986-7c0cebe93d45","Type":"ContainerDied","Data":"9d89c6f811a67ef290616b6f7423814d0698c88e4fa418f456e2c0bded217af0"} Jan 26 17:06:14 crc kubenswrapper[4754]: I0126 17:06:14.804800 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Jan 26 17:06:14 crc kubenswrapper[4754]: I0126 17:06:14.804840 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Jan 26 17:06:14 crc kubenswrapper[4754]: I0126 17:06:14.822582 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-57d769cc4f-72h5n" podStartSLOduration=3.386781917 podStartE2EDuration="38.822556902s" podCreationTimestamp="2026-01-26 17:05:36 +0000 UTC" firstStartedPulling="2026-01-26 17:05:37.293353538 +0000 UTC m=+1103.817533972" lastFinishedPulling="2026-01-26 17:06:12.729128523 +0000 UTC m=+1139.253308957" observedRunningTime="2026-01-26 17:06:14.81642143 +0000 UTC m=+1141.340601864" watchObservedRunningTime="2026-01-26 17:06:14.822556902 +0000 UTC m=+1141.346737346" Jan 26 17:06:14 crc kubenswrapper[4754]: I0126 17:06:14.854244 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Jan 26 17:06:14 crc kubenswrapper[4754]: I0126 17:06:14.948400 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Jan 26 17:06:15 crc kubenswrapper[4754]: I0126 17:06:15.173750 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-x2tgk"] Jan 26 17:06:15 crc kubenswrapper[4754]: I0126 17:06:15.174739 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-x2tgk" Jan 26 17:06:15 crc kubenswrapper[4754]: I0126 17:06:15.177493 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Jan 26 17:06:15 crc kubenswrapper[4754]: I0126 17:06:15.184397 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-x2tgk"] Jan 26 17:06:15 crc kubenswrapper[4754]: I0126 17:06:15.314494 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-72h5n"] Jan 26 17:06:15 crc kubenswrapper[4754]: I0126 17:06:15.329131 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/28bf8197-f3d2-4ee1-9054-482fa295d92d-ovn-rundir\") pod \"ovn-controller-metrics-x2tgk\" (UID: \"28bf8197-f3d2-4ee1-9054-482fa295d92d\") " pod="openstack/ovn-controller-metrics-x2tgk" Jan 26 17:06:15 crc kubenswrapper[4754]: I0126 17:06:15.329225 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/28bf8197-f3d2-4ee1-9054-482fa295d92d-ovs-rundir\") pod \"ovn-controller-metrics-x2tgk\" (UID: \"28bf8197-f3d2-4ee1-9054-482fa295d92d\") " pod="openstack/ovn-controller-metrics-x2tgk" Jan 26 17:06:15 crc kubenswrapper[4754]: I0126 17:06:15.329283 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/28bf8197-f3d2-4ee1-9054-482fa295d92d-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-x2tgk\" (UID: \"28bf8197-f3d2-4ee1-9054-482fa295d92d\") " pod="openstack/ovn-controller-metrics-x2tgk" Jan 26 17:06:15 crc kubenswrapper[4754]: I0126 17:06:15.329344 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/28bf8197-f3d2-4ee1-9054-482fa295d92d-config\") pod \"ovn-controller-metrics-x2tgk\" (UID: \"28bf8197-f3d2-4ee1-9054-482fa295d92d\") " pod="openstack/ovn-controller-metrics-x2tgk" Jan 26 17:06:15 crc kubenswrapper[4754]: I0126 17:06:15.329373 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/28bf8197-f3d2-4ee1-9054-482fa295d92d-combined-ca-bundle\") pod \"ovn-controller-metrics-x2tgk\" (UID: \"28bf8197-f3d2-4ee1-9054-482fa295d92d\") " pod="openstack/ovn-controller-metrics-x2tgk" Jan 26 17:06:15 crc kubenswrapper[4754]: I0126 17:06:15.329405 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6twpc\" (UniqueName: \"kubernetes.io/projected/28bf8197-f3d2-4ee1-9054-482fa295d92d-kube-api-access-6twpc\") pod \"ovn-controller-metrics-x2tgk\" (UID: \"28bf8197-f3d2-4ee1-9054-482fa295d92d\") " pod="openstack/ovn-controller-metrics-x2tgk" Jan 26 17:06:15 crc kubenswrapper[4754]: I0126 17:06:15.356078 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7f896c8c65-jbhcp"] Jan 26 17:06:15 crc kubenswrapper[4754]: I0126 17:06:15.357475 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7f896c8c65-jbhcp" Jan 26 17:06:15 crc kubenswrapper[4754]: I0126 17:06:15.361594 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Jan 26 17:06:15 crc kubenswrapper[4754]: I0126 17:06:15.379133 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7f896c8c65-jbhcp"] Jan 26 17:06:15 crc kubenswrapper[4754]: I0126 17:06:15.431246 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/28bf8197-f3d2-4ee1-9054-482fa295d92d-ovs-rundir\") pod \"ovn-controller-metrics-x2tgk\" (UID: \"28bf8197-f3d2-4ee1-9054-482fa295d92d\") " pod="openstack/ovn-controller-metrics-x2tgk" Jan 26 17:06:15 crc kubenswrapper[4754]: I0126 17:06:15.431359 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/28bf8197-f3d2-4ee1-9054-482fa295d92d-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-x2tgk\" (UID: \"28bf8197-f3d2-4ee1-9054-482fa295d92d\") " pod="openstack/ovn-controller-metrics-x2tgk" Jan 26 17:06:15 crc kubenswrapper[4754]: I0126 17:06:15.431466 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/28bf8197-f3d2-4ee1-9054-482fa295d92d-config\") pod \"ovn-controller-metrics-x2tgk\" (UID: \"28bf8197-f3d2-4ee1-9054-482fa295d92d\") " pod="openstack/ovn-controller-metrics-x2tgk" Jan 26 17:06:15 crc kubenswrapper[4754]: I0126 17:06:15.431492 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/28bf8197-f3d2-4ee1-9054-482fa295d92d-combined-ca-bundle\") pod \"ovn-controller-metrics-x2tgk\" (UID: \"28bf8197-f3d2-4ee1-9054-482fa295d92d\") " pod="openstack/ovn-controller-metrics-x2tgk" Jan 26 17:06:15 crc kubenswrapper[4754]: I0126 17:06:15.431518 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/28bf8197-f3d2-4ee1-9054-482fa295d92d-ovs-rundir\") pod \"ovn-controller-metrics-x2tgk\" (UID: \"28bf8197-f3d2-4ee1-9054-482fa295d92d\") " pod="openstack/ovn-controller-metrics-x2tgk" Jan 26 17:06:15 crc kubenswrapper[4754]: I0126 17:06:15.431558 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6twpc\" (UniqueName: \"kubernetes.io/projected/28bf8197-f3d2-4ee1-9054-482fa295d92d-kube-api-access-6twpc\") pod \"ovn-controller-metrics-x2tgk\" (UID: \"28bf8197-f3d2-4ee1-9054-482fa295d92d\") " pod="openstack/ovn-controller-metrics-x2tgk" Jan 26 17:06:15 crc kubenswrapper[4754]: I0126 17:06:15.431639 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/28bf8197-f3d2-4ee1-9054-482fa295d92d-ovn-rundir\") pod \"ovn-controller-metrics-x2tgk\" (UID: \"28bf8197-f3d2-4ee1-9054-482fa295d92d\") " pod="openstack/ovn-controller-metrics-x2tgk" Jan 26 17:06:15 crc kubenswrapper[4754]: I0126 17:06:15.431810 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/28bf8197-f3d2-4ee1-9054-482fa295d92d-ovn-rundir\") pod \"ovn-controller-metrics-x2tgk\" (UID: \"28bf8197-f3d2-4ee1-9054-482fa295d92d\") " pod="openstack/ovn-controller-metrics-x2tgk" Jan 26 17:06:15 crc kubenswrapper[4754]: I0126 17:06:15.432322 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/28bf8197-f3d2-4ee1-9054-482fa295d92d-config\") pod \"ovn-controller-metrics-x2tgk\" (UID: \"28bf8197-f3d2-4ee1-9054-482fa295d92d\") " pod="openstack/ovn-controller-metrics-x2tgk" Jan 26 17:06:15 crc kubenswrapper[4754]: I0126 17:06:15.436327 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/28bf8197-f3d2-4ee1-9054-482fa295d92d-combined-ca-bundle\") pod \"ovn-controller-metrics-x2tgk\" (UID: \"28bf8197-f3d2-4ee1-9054-482fa295d92d\") " pod="openstack/ovn-controller-metrics-x2tgk" Jan 26 17:06:15 crc kubenswrapper[4754]: I0126 17:06:15.451912 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/28bf8197-f3d2-4ee1-9054-482fa295d92d-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-x2tgk\" (UID: \"28bf8197-f3d2-4ee1-9054-482fa295d92d\") " pod="openstack/ovn-controller-metrics-x2tgk" Jan 26 17:06:15 crc kubenswrapper[4754]: I0126 17:06:15.457884 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6twpc\" (UniqueName: \"kubernetes.io/projected/28bf8197-f3d2-4ee1-9054-482fa295d92d-kube-api-access-6twpc\") pod \"ovn-controller-metrics-x2tgk\" (UID: \"28bf8197-f3d2-4ee1-9054-482fa295d92d\") " pod="openstack/ovn-controller-metrics-x2tgk" Jan 26 17:06:15 crc kubenswrapper[4754]: I0126 17:06:15.484201 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-xtjb4"] Jan 26 17:06:15 crc kubenswrapper[4754]: I0126 17:06:15.493979 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-x2tgk" Jan 26 17:06:15 crc kubenswrapper[4754]: I0126 17:06:15.535765 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0bf413d0-f8cf-4574-b4bc-dfa3264a41b9-dns-svc\") pod \"dnsmasq-dns-7f896c8c65-jbhcp\" (UID: \"0bf413d0-f8cf-4574-b4bc-dfa3264a41b9\") " pod="openstack/dnsmasq-dns-7f896c8c65-jbhcp" Jan 26 17:06:15 crc kubenswrapper[4754]: I0126 17:06:15.535832 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0bf413d0-f8cf-4574-b4bc-dfa3264a41b9-config\") pod \"dnsmasq-dns-7f896c8c65-jbhcp\" (UID: \"0bf413d0-f8cf-4574-b4bc-dfa3264a41b9\") " pod="openstack/dnsmasq-dns-7f896c8c65-jbhcp" Jan 26 17:06:15 crc kubenswrapper[4754]: I0126 17:06:15.535893 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vzwcn\" (UniqueName: \"kubernetes.io/projected/0bf413d0-f8cf-4574-b4bc-dfa3264a41b9-kube-api-access-vzwcn\") pod \"dnsmasq-dns-7f896c8c65-jbhcp\" (UID: \"0bf413d0-f8cf-4574-b4bc-dfa3264a41b9\") " pod="openstack/dnsmasq-dns-7f896c8c65-jbhcp" Jan 26 17:06:15 crc kubenswrapper[4754]: I0126 17:06:15.535921 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0bf413d0-f8cf-4574-b4bc-dfa3264a41b9-ovsdbserver-sb\") pod \"dnsmasq-dns-7f896c8c65-jbhcp\" (UID: \"0bf413d0-f8cf-4574-b4bc-dfa3264a41b9\") " pod="openstack/dnsmasq-dns-7f896c8c65-jbhcp" Jan 26 17:06:15 crc kubenswrapper[4754]: I0126 17:06:15.555306 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Jan 26 17:06:15 crc kubenswrapper[4754]: I0126 17:06:15.556523 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Jan 26 17:06:15 crc kubenswrapper[4754]: I0126 17:06:15.558551 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-6xlr8" Jan 26 17:06:15 crc kubenswrapper[4754]: I0126 17:06:15.558890 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Jan 26 17:06:15 crc kubenswrapper[4754]: I0126 17:06:15.558973 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Jan 26 17:06:15 crc kubenswrapper[4754]: I0126 17:06:15.559062 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Jan 26 17:06:15 crc kubenswrapper[4754]: I0126 17:06:15.575623 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-r9jrq"] Jan 26 17:06:15 crc kubenswrapper[4754]: I0126 17:06:15.577242 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-r9jrq" Jan 26 17:06:15 crc kubenswrapper[4754]: I0126 17:06:15.579230 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Jan 26 17:06:15 crc kubenswrapper[4754]: I0126 17:06:15.593744 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Jan 26 17:06:15 crc kubenswrapper[4754]: I0126 17:06:15.603404 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-r9jrq"] Jan 26 17:06:15 crc kubenswrapper[4754]: I0126 17:06:15.638577 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0bf413d0-f8cf-4574-b4bc-dfa3264a41b9-dns-svc\") pod \"dnsmasq-dns-7f896c8c65-jbhcp\" (UID: \"0bf413d0-f8cf-4574-b4bc-dfa3264a41b9\") " pod="openstack/dnsmasq-dns-7f896c8c65-jbhcp" Jan 26 17:06:15 crc kubenswrapper[4754]: I0126 17:06:15.638648 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0bf413d0-f8cf-4574-b4bc-dfa3264a41b9-config\") pod \"dnsmasq-dns-7f896c8c65-jbhcp\" (UID: \"0bf413d0-f8cf-4574-b4bc-dfa3264a41b9\") " pod="openstack/dnsmasq-dns-7f896c8c65-jbhcp" Jan 26 17:06:15 crc kubenswrapper[4754]: I0126 17:06:15.638722 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vzwcn\" (UniqueName: \"kubernetes.io/projected/0bf413d0-f8cf-4574-b4bc-dfa3264a41b9-kube-api-access-vzwcn\") pod \"dnsmasq-dns-7f896c8c65-jbhcp\" (UID: \"0bf413d0-f8cf-4574-b4bc-dfa3264a41b9\") " pod="openstack/dnsmasq-dns-7f896c8c65-jbhcp" Jan 26 17:06:15 crc kubenswrapper[4754]: I0126 17:06:15.638749 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0bf413d0-f8cf-4574-b4bc-dfa3264a41b9-ovsdbserver-sb\") pod \"dnsmasq-dns-7f896c8c65-jbhcp\" (UID: \"0bf413d0-f8cf-4574-b4bc-dfa3264a41b9\") " pod="openstack/dnsmasq-dns-7f896c8c65-jbhcp" Jan 26 17:06:15 crc kubenswrapper[4754]: I0126 17:06:15.639566 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0bf413d0-f8cf-4574-b4bc-dfa3264a41b9-ovsdbserver-sb\") pod \"dnsmasq-dns-7f896c8c65-jbhcp\" (UID: \"0bf413d0-f8cf-4574-b4bc-dfa3264a41b9\") " pod="openstack/dnsmasq-dns-7f896c8c65-jbhcp" Jan 26 17:06:15 crc kubenswrapper[4754]: I0126 17:06:15.640354 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0bf413d0-f8cf-4574-b4bc-dfa3264a41b9-dns-svc\") pod \"dnsmasq-dns-7f896c8c65-jbhcp\" (UID: \"0bf413d0-f8cf-4574-b4bc-dfa3264a41b9\") " pod="openstack/dnsmasq-dns-7f896c8c65-jbhcp" Jan 26 17:06:15 crc kubenswrapper[4754]: I0126 17:06:15.640882 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0bf413d0-f8cf-4574-b4bc-dfa3264a41b9-config\") pod \"dnsmasq-dns-7f896c8c65-jbhcp\" (UID: \"0bf413d0-f8cf-4574-b4bc-dfa3264a41b9\") " pod="openstack/dnsmasq-dns-7f896c8c65-jbhcp" Jan 26 17:06:15 crc kubenswrapper[4754]: I0126 17:06:15.659110 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vzwcn\" (UniqueName: \"kubernetes.io/projected/0bf413d0-f8cf-4574-b4bc-dfa3264a41b9-kube-api-access-vzwcn\") pod \"dnsmasq-dns-7f896c8c65-jbhcp\" (UID: \"0bf413d0-f8cf-4574-b4bc-dfa3264a41b9\") " pod="openstack/dnsmasq-dns-7f896c8c65-jbhcp" Jan 26 17:06:15 crc kubenswrapper[4754]: I0126 17:06:15.677733 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7f896c8c65-jbhcp" Jan 26 17:06:15 crc kubenswrapper[4754]: I0126 17:06:15.740307 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0df51b10-10ea-4822-ad9b-e952ba4673f8-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-r9jrq\" (UID: \"0df51b10-10ea-4822-ad9b-e952ba4673f8\") " pod="openstack/dnsmasq-dns-86db49b7ff-r9jrq" Jan 26 17:06:15 crc kubenswrapper[4754]: I0126 17:06:15.740357 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4af75510-1f61-40f7-a292-764facf90f1a-scripts\") pod \"ovn-northd-0\" (UID: \"4af75510-1f61-40f7-a292-764facf90f1a\") " pod="openstack/ovn-northd-0" Jan 26 17:06:15 crc kubenswrapper[4754]: I0126 17:06:15.740430 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4af75510-1f61-40f7-a292-764facf90f1a-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"4af75510-1f61-40f7-a292-764facf90f1a\") " pod="openstack/ovn-northd-0" Jan 26 17:06:15 crc kubenswrapper[4754]: I0126 17:06:15.740481 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0df51b10-10ea-4822-ad9b-e952ba4673f8-config\") pod \"dnsmasq-dns-86db49b7ff-r9jrq\" (UID: \"0df51b10-10ea-4822-ad9b-e952ba4673f8\") " pod="openstack/dnsmasq-dns-86db49b7ff-r9jrq" Jan 26 17:06:15 crc kubenswrapper[4754]: I0126 17:06:15.740557 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/4af75510-1f61-40f7-a292-764facf90f1a-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"4af75510-1f61-40f7-a292-764facf90f1a\") " pod="openstack/ovn-northd-0" Jan 26 17:06:15 crc kubenswrapper[4754]: I0126 17:06:15.740583 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8xd7w\" (UniqueName: \"kubernetes.io/projected/4af75510-1f61-40f7-a292-764facf90f1a-kube-api-access-8xd7w\") pod \"ovn-northd-0\" (UID: \"4af75510-1f61-40f7-a292-764facf90f1a\") " pod="openstack/ovn-northd-0" Jan 26 17:06:15 crc kubenswrapper[4754]: I0126 17:06:15.740637 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0df51b10-10ea-4822-ad9b-e952ba4673f8-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-r9jrq\" (UID: \"0df51b10-10ea-4822-ad9b-e952ba4673f8\") " pod="openstack/dnsmasq-dns-86db49b7ff-r9jrq" Jan 26 17:06:15 crc kubenswrapper[4754]: I0126 17:06:15.740692 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/4af75510-1f61-40f7-a292-764facf90f1a-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"4af75510-1f61-40f7-a292-764facf90f1a\") " pod="openstack/ovn-northd-0" Jan 26 17:06:15 crc kubenswrapper[4754]: I0126 17:06:15.740715 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4af75510-1f61-40f7-a292-764facf90f1a-config\") pod \"ovn-northd-0\" (UID: \"4af75510-1f61-40f7-a292-764facf90f1a\") " pod="openstack/ovn-northd-0" Jan 26 17:06:15 crc kubenswrapper[4754]: I0126 17:06:15.740733 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/4af75510-1f61-40f7-a292-764facf90f1a-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"4af75510-1f61-40f7-a292-764facf90f1a\") " pod="openstack/ovn-northd-0" Jan 26 17:06:15 crc kubenswrapper[4754]: I0126 17:06:15.740782 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2s5m4\" (UniqueName: \"kubernetes.io/projected/0df51b10-10ea-4822-ad9b-e952ba4673f8-kube-api-access-2s5m4\") pod \"dnsmasq-dns-86db49b7ff-r9jrq\" (UID: \"0df51b10-10ea-4822-ad9b-e952ba4673f8\") " pod="openstack/dnsmasq-dns-86db49b7ff-r9jrq" Jan 26 17:06:15 crc kubenswrapper[4754]: I0126 17:06:15.740805 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0df51b10-10ea-4822-ad9b-e952ba4673f8-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-r9jrq\" (UID: \"0df51b10-10ea-4822-ad9b-e952ba4673f8\") " pod="openstack/dnsmasq-dns-86db49b7ff-r9jrq" Jan 26 17:06:15 crc kubenswrapper[4754]: I0126 17:06:15.818392 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-xtjb4" event={"ID":"c4637fc0-04b8-4b18-8986-7c0cebe93d45","Type":"ContainerStarted","Data":"8bbec4769b6e9ae5cc1312f43c17f25772f204c98cb327b116bfb75f9e148995"} Jan 26 17:06:15 crc kubenswrapper[4754]: I0126 17:06:15.818571 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-666b6646f7-xtjb4" podUID="c4637fc0-04b8-4b18-8986-7c0cebe93d45" containerName="dnsmasq-dns" containerID="cri-o://8bbec4769b6e9ae5cc1312f43c17f25772f204c98cb327b116bfb75f9e148995" gracePeriod=10 Jan 26 17:06:15 crc kubenswrapper[4754]: I0126 17:06:15.818852 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-666b6646f7-xtjb4" Jan 26 17:06:15 crc kubenswrapper[4754]: I0126 17:06:15.821326 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"48810c5e-3c40-4cdc-8bab-47efa97e76fa","Type":"ContainerStarted","Data":"397f275bbe0a6c7fe3d67ba536948ee64336fe00addd514ba132a01fca3d4cb6"} Jan 26 17:06:15 crc kubenswrapper[4754]: I0126 17:06:15.844889 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8xd7w\" (UniqueName: \"kubernetes.io/projected/4af75510-1f61-40f7-a292-764facf90f1a-kube-api-access-8xd7w\") pod \"ovn-northd-0\" (UID: \"4af75510-1f61-40f7-a292-764facf90f1a\") " pod="openstack/ovn-northd-0" Jan 26 17:06:15 crc kubenswrapper[4754]: I0126 17:06:15.844951 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0df51b10-10ea-4822-ad9b-e952ba4673f8-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-r9jrq\" (UID: \"0df51b10-10ea-4822-ad9b-e952ba4673f8\") " pod="openstack/dnsmasq-dns-86db49b7ff-r9jrq" Jan 26 17:06:15 crc kubenswrapper[4754]: I0126 17:06:15.844983 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/4af75510-1f61-40f7-a292-764facf90f1a-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"4af75510-1f61-40f7-a292-764facf90f1a\") " pod="openstack/ovn-northd-0" Jan 26 17:06:15 crc kubenswrapper[4754]: I0126 17:06:15.845004 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4af75510-1f61-40f7-a292-764facf90f1a-config\") pod \"ovn-northd-0\" (UID: \"4af75510-1f61-40f7-a292-764facf90f1a\") " pod="openstack/ovn-northd-0" Jan 26 17:06:15 crc kubenswrapper[4754]: I0126 17:06:15.845024 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/4af75510-1f61-40f7-a292-764facf90f1a-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"4af75510-1f61-40f7-a292-764facf90f1a\") " pod="openstack/ovn-northd-0" Jan 26 17:06:15 crc kubenswrapper[4754]: I0126 17:06:15.845055 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2s5m4\" (UniqueName: \"kubernetes.io/projected/0df51b10-10ea-4822-ad9b-e952ba4673f8-kube-api-access-2s5m4\") pod \"dnsmasq-dns-86db49b7ff-r9jrq\" (UID: \"0df51b10-10ea-4822-ad9b-e952ba4673f8\") " pod="openstack/dnsmasq-dns-86db49b7ff-r9jrq" Jan 26 17:06:15 crc kubenswrapper[4754]: I0126 17:06:15.845080 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0df51b10-10ea-4822-ad9b-e952ba4673f8-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-r9jrq\" (UID: \"0df51b10-10ea-4822-ad9b-e952ba4673f8\") " pod="openstack/dnsmasq-dns-86db49b7ff-r9jrq" Jan 26 17:06:15 crc kubenswrapper[4754]: I0126 17:06:15.845107 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4af75510-1f61-40f7-a292-764facf90f1a-scripts\") pod \"ovn-northd-0\" (UID: \"4af75510-1f61-40f7-a292-764facf90f1a\") " pod="openstack/ovn-northd-0" Jan 26 17:06:15 crc kubenswrapper[4754]: I0126 17:06:15.845125 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0df51b10-10ea-4822-ad9b-e952ba4673f8-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-r9jrq\" (UID: \"0df51b10-10ea-4822-ad9b-e952ba4673f8\") " pod="openstack/dnsmasq-dns-86db49b7ff-r9jrq" Jan 26 17:06:15 crc kubenswrapper[4754]: I0126 17:06:15.845157 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4af75510-1f61-40f7-a292-764facf90f1a-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"4af75510-1f61-40f7-a292-764facf90f1a\") " pod="openstack/ovn-northd-0" Jan 26 17:06:15 crc kubenswrapper[4754]: I0126 17:06:15.845188 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0df51b10-10ea-4822-ad9b-e952ba4673f8-config\") pod \"dnsmasq-dns-86db49b7ff-r9jrq\" (UID: \"0df51b10-10ea-4822-ad9b-e952ba4673f8\") " pod="openstack/dnsmasq-dns-86db49b7ff-r9jrq" Jan 26 17:06:15 crc kubenswrapper[4754]: I0126 17:06:15.845247 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/4af75510-1f61-40f7-a292-764facf90f1a-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"4af75510-1f61-40f7-a292-764facf90f1a\") " pod="openstack/ovn-northd-0" Jan 26 17:06:15 crc kubenswrapper[4754]: I0126 17:06:15.846110 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/4af75510-1f61-40f7-a292-764facf90f1a-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"4af75510-1f61-40f7-a292-764facf90f1a\") " pod="openstack/ovn-northd-0" Jan 26 17:06:15 crc kubenswrapper[4754]: I0126 17:06:15.847054 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0df51b10-10ea-4822-ad9b-e952ba4673f8-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-r9jrq\" (UID: \"0df51b10-10ea-4822-ad9b-e952ba4673f8\") " pod="openstack/dnsmasq-dns-86db49b7ff-r9jrq" Jan 26 17:06:15 crc kubenswrapper[4754]: I0126 17:06:15.847392 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4af75510-1f61-40f7-a292-764facf90f1a-config\") pod \"ovn-northd-0\" (UID: \"4af75510-1f61-40f7-a292-764facf90f1a\") " pod="openstack/ovn-northd-0" Jan 26 17:06:15 crc kubenswrapper[4754]: I0126 17:06:15.847802 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4af75510-1f61-40f7-a292-764facf90f1a-scripts\") pod \"ovn-northd-0\" (UID: \"4af75510-1f61-40f7-a292-764facf90f1a\") " pod="openstack/ovn-northd-0" Jan 26 17:06:15 crc kubenswrapper[4754]: I0126 17:06:15.848238 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0df51b10-10ea-4822-ad9b-e952ba4673f8-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-r9jrq\" (UID: \"0df51b10-10ea-4822-ad9b-e952ba4673f8\") " pod="openstack/dnsmasq-dns-86db49b7ff-r9jrq" Jan 26 17:06:15 crc kubenswrapper[4754]: I0126 17:06:15.848504 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0df51b10-10ea-4822-ad9b-e952ba4673f8-config\") pod \"dnsmasq-dns-86db49b7ff-r9jrq\" (UID: \"0df51b10-10ea-4822-ad9b-e952ba4673f8\") " pod="openstack/dnsmasq-dns-86db49b7ff-r9jrq" Jan 26 17:06:15 crc kubenswrapper[4754]: I0126 17:06:15.848575 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0df51b10-10ea-4822-ad9b-e952ba4673f8-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-r9jrq\" (UID: \"0df51b10-10ea-4822-ad9b-e952ba4673f8\") " pod="openstack/dnsmasq-dns-86db49b7ff-r9jrq" Jan 26 17:06:15 crc kubenswrapper[4754]: I0126 17:06:15.851236 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-666b6646f7-xtjb4" podStartSLOduration=-9223371997.003557 podStartE2EDuration="39.851218131s" podCreationTimestamp="2026-01-26 17:05:36 +0000 UTC" firstStartedPulling="2026-01-26 17:05:37.301781604 +0000 UTC m=+1103.825962038" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 17:06:15.842971121 +0000 UTC m=+1142.367151605" watchObservedRunningTime="2026-01-26 17:06:15.851218131 +0000 UTC m=+1142.375398565" Jan 26 17:06:15 crc kubenswrapper[4754]: I0126 17:06:15.856590 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/4af75510-1f61-40f7-a292-764facf90f1a-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"4af75510-1f61-40f7-a292-764facf90f1a\") " pod="openstack/ovn-northd-0" Jan 26 17:06:15 crc kubenswrapper[4754]: I0126 17:06:15.861481 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4af75510-1f61-40f7-a292-764facf90f1a-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"4af75510-1f61-40f7-a292-764facf90f1a\") " pod="openstack/ovn-northd-0" Jan 26 17:06:15 crc kubenswrapper[4754]: I0126 17:06:15.862962 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/4af75510-1f61-40f7-a292-764facf90f1a-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"4af75510-1f61-40f7-a292-764facf90f1a\") " pod="openstack/ovn-northd-0" Jan 26 17:06:15 crc kubenswrapper[4754]: I0126 17:06:15.873211 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8xd7w\" (UniqueName: \"kubernetes.io/projected/4af75510-1f61-40f7-a292-764facf90f1a-kube-api-access-8xd7w\") pod \"ovn-northd-0\" (UID: \"4af75510-1f61-40f7-a292-764facf90f1a\") " pod="openstack/ovn-northd-0" Jan 26 17:06:15 crc kubenswrapper[4754]: I0126 17:06:15.886352 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2s5m4\" (UniqueName: \"kubernetes.io/projected/0df51b10-10ea-4822-ad9b-e952ba4673f8-kube-api-access-2s5m4\") pod \"dnsmasq-dns-86db49b7ff-r9jrq\" (UID: \"0df51b10-10ea-4822-ad9b-e952ba4673f8\") " pod="openstack/dnsmasq-dns-86db49b7ff-r9jrq" Jan 26 17:06:15 crc kubenswrapper[4754]: I0126 17:06:15.894061 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Jan 26 17:06:15 crc kubenswrapper[4754]: I0126 17:06:15.904703 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Jan 26 17:06:15 crc kubenswrapper[4754]: I0126 17:06:15.909333 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-r9jrq" Jan 26 17:06:16 crc kubenswrapper[4754]: I0126 17:06:16.000166 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-x2tgk"] Jan 26 17:06:16 crc kubenswrapper[4754]: I0126 17:06:16.170191 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7f896c8c65-jbhcp"] Jan 26 17:06:16 crc kubenswrapper[4754]: I0126 17:06:16.411619 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Jan 26 17:06:16 crc kubenswrapper[4754]: W0126 17:06:16.475174 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4af75510_1f61_40f7_a292_764facf90f1a.slice/crio-74a50d4f3a78a6d580d645dc84c61c6450f04756be875818ba7cabf9486d14bf WatchSource:0}: Error finding container 74a50d4f3a78a6d580d645dc84c61c6450f04756be875818ba7cabf9486d14bf: Status 404 returned error can't find the container with id 74a50d4f3a78a6d580d645dc84c61c6450f04756be875818ba7cabf9486d14bf Jan 26 17:06:16 crc kubenswrapper[4754]: I0126 17:06:16.476384 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-r9jrq"] Jan 26 17:06:16 crc kubenswrapper[4754]: W0126 17:06:16.481115 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0df51b10_10ea_4822_ad9b_e952ba4673f8.slice/crio-e8f45055324a26f98dc4083fe586bd50ba46807545bfeaf4613b880ea1fa974e WatchSource:0}: Error finding container e8f45055324a26f98dc4083fe586bd50ba46807545bfeaf4613b880ea1fa974e: Status 404 returned error can't find the container with id e8f45055324a26f98dc4083fe586bd50ba46807545bfeaf4613b880ea1fa974e Jan 26 17:06:16 crc kubenswrapper[4754]: I0126 17:06:16.700012 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-xtjb4" Jan 26 17:06:16 crc kubenswrapper[4754]: I0126 17:06:16.758971 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j9ww2\" (UniqueName: \"kubernetes.io/projected/c4637fc0-04b8-4b18-8986-7c0cebe93d45-kube-api-access-j9ww2\") pod \"c4637fc0-04b8-4b18-8986-7c0cebe93d45\" (UID: \"c4637fc0-04b8-4b18-8986-7c0cebe93d45\") " Jan 26 17:06:16 crc kubenswrapper[4754]: I0126 17:06:16.759088 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c4637fc0-04b8-4b18-8986-7c0cebe93d45-config\") pod \"c4637fc0-04b8-4b18-8986-7c0cebe93d45\" (UID: \"c4637fc0-04b8-4b18-8986-7c0cebe93d45\") " Jan 26 17:06:16 crc kubenswrapper[4754]: I0126 17:06:16.759126 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c4637fc0-04b8-4b18-8986-7c0cebe93d45-dns-svc\") pod \"c4637fc0-04b8-4b18-8986-7c0cebe93d45\" (UID: \"c4637fc0-04b8-4b18-8986-7c0cebe93d45\") " Jan 26 17:06:16 crc kubenswrapper[4754]: I0126 17:06:16.765874 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c4637fc0-04b8-4b18-8986-7c0cebe93d45-kube-api-access-j9ww2" (OuterVolumeSpecName: "kube-api-access-j9ww2") pod "c4637fc0-04b8-4b18-8986-7c0cebe93d45" (UID: "c4637fc0-04b8-4b18-8986-7c0cebe93d45"). InnerVolumeSpecName "kube-api-access-j9ww2". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:06:16 crc kubenswrapper[4754]: I0126 17:06:16.840493 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c4637fc0-04b8-4b18-8986-7c0cebe93d45-config" (OuterVolumeSpecName: "config") pod "c4637fc0-04b8-4b18-8986-7c0cebe93d45" (UID: "c4637fc0-04b8-4b18-8986-7c0cebe93d45"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:06:16 crc kubenswrapper[4754]: I0126 17:06:16.848149 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c4637fc0-04b8-4b18-8986-7c0cebe93d45-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "c4637fc0-04b8-4b18-8986-7c0cebe93d45" (UID: "c4637fc0-04b8-4b18-8986-7c0cebe93d45"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:06:16 crc kubenswrapper[4754]: I0126 17:06:16.865276 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j9ww2\" (UniqueName: \"kubernetes.io/projected/c4637fc0-04b8-4b18-8986-7c0cebe93d45-kube-api-access-j9ww2\") on node \"crc\" DevicePath \"\"" Jan 26 17:06:16 crc kubenswrapper[4754]: I0126 17:06:16.865309 4754 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c4637fc0-04b8-4b18-8986-7c0cebe93d45-config\") on node \"crc\" DevicePath \"\"" Jan 26 17:06:16 crc kubenswrapper[4754]: I0126 17:06:16.865321 4754 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c4637fc0-04b8-4b18-8986-7c0cebe93d45-dns-svc\") on node \"crc\" DevicePath \"\"" Jan 26 17:06:16 crc kubenswrapper[4754]: I0126 17:06:16.868958 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"4af75510-1f61-40f7-a292-764facf90f1a","Type":"ContainerStarted","Data":"74a50d4f3a78a6d580d645dc84c61c6450f04756be875818ba7cabf9486d14bf"} Jan 26 17:06:16 crc kubenswrapper[4754]: I0126 17:06:16.925910 4754 generic.go:334] "Generic (PLEG): container finished" podID="0bf413d0-f8cf-4574-b4bc-dfa3264a41b9" containerID="2b608aaa53ab4942a14be828c91bb9b5fbc5de669bb71c595bbfff4931865335" exitCode=0 Jan 26 17:06:16 crc kubenswrapper[4754]: I0126 17:06:16.926725 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7f896c8c65-jbhcp" event={"ID":"0bf413d0-f8cf-4574-b4bc-dfa3264a41b9","Type":"ContainerDied","Data":"2b608aaa53ab4942a14be828c91bb9b5fbc5de669bb71c595bbfff4931865335"} Jan 26 17:06:16 crc kubenswrapper[4754]: I0126 17:06:16.926749 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7f896c8c65-jbhcp" event={"ID":"0bf413d0-f8cf-4574-b4bc-dfa3264a41b9","Type":"ContainerStarted","Data":"b5ffcba758c81e5fab424c6de16c7df0a82d482567a6d60c6132f2be69ac129c"} Jan 26 17:06:16 crc kubenswrapper[4754]: I0126 17:06:16.936222 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-x2tgk" event={"ID":"28bf8197-f3d2-4ee1-9054-482fa295d92d","Type":"ContainerStarted","Data":"c876bbb98e0d5fb41c58dffc80f103be6f4b5a3a66899ddbeb0ae6ed39e2e187"} Jan 26 17:06:16 crc kubenswrapper[4754]: I0126 17:06:16.936543 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-x2tgk" event={"ID":"28bf8197-f3d2-4ee1-9054-482fa295d92d","Type":"ContainerStarted","Data":"877988468ab6d9d908c186a7d0f5caefd0794a52d338ebd0774500699cc0fdd0"} Jan 26 17:06:16 crc kubenswrapper[4754]: I0126 17:06:16.937437 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-r9jrq" event={"ID":"0df51b10-10ea-4822-ad9b-e952ba4673f8","Type":"ContainerStarted","Data":"62dc8ee2c3e18b05a5ad1bdbc9c5d7af30f0f506b69d408c389b431fec3b3fb8"} Jan 26 17:06:16 crc kubenswrapper[4754]: I0126 17:06:16.937467 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-r9jrq" event={"ID":"0df51b10-10ea-4822-ad9b-e952ba4673f8","Type":"ContainerStarted","Data":"e8f45055324a26f98dc4083fe586bd50ba46807545bfeaf4613b880ea1fa974e"} Jan 26 17:06:16 crc kubenswrapper[4754]: I0126 17:06:16.939924 4754 generic.go:334] "Generic (PLEG): container finished" podID="c4637fc0-04b8-4b18-8986-7c0cebe93d45" containerID="8bbec4769b6e9ae5cc1312f43c17f25772f204c98cb327b116bfb75f9e148995" exitCode=0 Jan 26 17:06:16 crc kubenswrapper[4754]: I0126 17:06:16.940043 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-xtjb4" event={"ID":"c4637fc0-04b8-4b18-8986-7c0cebe93d45","Type":"ContainerDied","Data":"8bbec4769b6e9ae5cc1312f43c17f25772f204c98cb327b116bfb75f9e148995"} Jan 26 17:06:16 crc kubenswrapper[4754]: I0126 17:06:16.940127 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-xtjb4" event={"ID":"c4637fc0-04b8-4b18-8986-7c0cebe93d45","Type":"ContainerDied","Data":"99b53c28d42483c2284404e91db83ee11ef7096181ed5c780a6a385ba5bd0fd8"} Jan 26 17:06:16 crc kubenswrapper[4754]: I0126 17:06:16.940150 4754 scope.go:117] "RemoveContainer" containerID="8bbec4769b6e9ae5cc1312f43c17f25772f204c98cb327b116bfb75f9e148995" Jan 26 17:06:16 crc kubenswrapper[4754]: I0126 17:06:16.940378 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-xtjb4" Jan 26 17:06:16 crc kubenswrapper[4754]: I0126 17:06:16.942059 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-57d769cc4f-72h5n" podUID="a9ccad70-dcd4-473c-aa09-925915f7420f" containerName="dnsmasq-dns" containerID="cri-o://1d9caa5e5f0c0418fd0df5b7a9010ebb5b310f1dd2c3774dd558998962e06810" gracePeriod=10 Jan 26 17:06:17 crc kubenswrapper[4754]: I0126 17:06:17.009220 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-x2tgk" podStartSLOduration=2.009199994 podStartE2EDuration="2.009199994s" podCreationTimestamp="2026-01-26 17:06:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 17:06:16.99364521 +0000 UTC m=+1143.517825664" watchObservedRunningTime="2026-01-26 17:06:17.009199994 +0000 UTC m=+1143.533380428" Jan 26 17:06:17 crc kubenswrapper[4754]: I0126 17:06:17.137229 4754 scope.go:117] "RemoveContainer" containerID="9d89c6f811a67ef290616b6f7423814d0698c88e4fa418f456e2c0bded217af0" Jan 26 17:06:17 crc kubenswrapper[4754]: I0126 17:06:17.161260 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-xtjb4"] Jan 26 17:06:17 crc kubenswrapper[4754]: I0126 17:06:17.171004 4754 scope.go:117] "RemoveContainer" containerID="8bbec4769b6e9ae5cc1312f43c17f25772f204c98cb327b116bfb75f9e148995" Jan 26 17:06:17 crc kubenswrapper[4754]: E0126 17:06:17.175649 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8bbec4769b6e9ae5cc1312f43c17f25772f204c98cb327b116bfb75f9e148995\": container with ID starting with 8bbec4769b6e9ae5cc1312f43c17f25772f204c98cb327b116bfb75f9e148995 not found: ID does not exist" containerID="8bbec4769b6e9ae5cc1312f43c17f25772f204c98cb327b116bfb75f9e148995" Jan 26 17:06:17 crc kubenswrapper[4754]: I0126 17:06:17.175728 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8bbec4769b6e9ae5cc1312f43c17f25772f204c98cb327b116bfb75f9e148995"} err="failed to get container status \"8bbec4769b6e9ae5cc1312f43c17f25772f204c98cb327b116bfb75f9e148995\": rpc error: code = NotFound desc = could not find container \"8bbec4769b6e9ae5cc1312f43c17f25772f204c98cb327b116bfb75f9e148995\": container with ID starting with 8bbec4769b6e9ae5cc1312f43c17f25772f204c98cb327b116bfb75f9e148995 not found: ID does not exist" Jan 26 17:06:17 crc kubenswrapper[4754]: I0126 17:06:17.175762 4754 scope.go:117] "RemoveContainer" containerID="9d89c6f811a67ef290616b6f7423814d0698c88e4fa418f456e2c0bded217af0" Jan 26 17:06:17 crc kubenswrapper[4754]: E0126 17:06:17.176119 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9d89c6f811a67ef290616b6f7423814d0698c88e4fa418f456e2c0bded217af0\": container with ID starting with 9d89c6f811a67ef290616b6f7423814d0698c88e4fa418f456e2c0bded217af0 not found: ID does not exist" containerID="9d89c6f811a67ef290616b6f7423814d0698c88e4fa418f456e2c0bded217af0" Jan 26 17:06:17 crc kubenswrapper[4754]: I0126 17:06:17.176148 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9d89c6f811a67ef290616b6f7423814d0698c88e4fa418f456e2c0bded217af0"} err="failed to get container status \"9d89c6f811a67ef290616b6f7423814d0698c88e4fa418f456e2c0bded217af0\": rpc error: code = NotFound desc = could not find container \"9d89c6f811a67ef290616b6f7423814d0698c88e4fa418f456e2c0bded217af0\": container with ID starting with 9d89c6f811a67ef290616b6f7423814d0698c88e4fa418f456e2c0bded217af0 not found: ID does not exist" Jan 26 17:06:17 crc kubenswrapper[4754]: I0126 17:06:17.177858 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-xtjb4"] Jan 26 17:06:17 crc kubenswrapper[4754]: I0126 17:06:17.776859 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c4637fc0-04b8-4b18-8986-7c0cebe93d45" path="/var/lib/kubelet/pods/c4637fc0-04b8-4b18-8986-7c0cebe93d45/volumes" Jan 26 17:06:17 crc kubenswrapper[4754]: I0126 17:06:17.940427 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-72h5n" Jan 26 17:06:17 crc kubenswrapper[4754]: I0126 17:06:17.966885 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7f896c8c65-jbhcp" event={"ID":"0bf413d0-f8cf-4574-b4bc-dfa3264a41b9","Type":"ContainerStarted","Data":"93a7968f52e595c1a52011c2b828412ae7f861a72b4ce6f0b29ff7953e389879"} Jan 26 17:06:17 crc kubenswrapper[4754]: I0126 17:06:17.967030 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7f896c8c65-jbhcp" Jan 26 17:06:17 crc kubenswrapper[4754]: I0126 17:06:17.974451 4754 generic.go:334] "Generic (PLEG): container finished" podID="a9ccad70-dcd4-473c-aa09-925915f7420f" containerID="1d9caa5e5f0c0418fd0df5b7a9010ebb5b310f1dd2c3774dd558998962e06810" exitCode=0 Jan 26 17:06:17 crc kubenswrapper[4754]: I0126 17:06:17.974517 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-72h5n" event={"ID":"a9ccad70-dcd4-473c-aa09-925915f7420f","Type":"ContainerDied","Data":"1d9caa5e5f0c0418fd0df5b7a9010ebb5b310f1dd2c3774dd558998962e06810"} Jan 26 17:06:17 crc kubenswrapper[4754]: I0126 17:06:17.974525 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-72h5n" Jan 26 17:06:17 crc kubenswrapper[4754]: I0126 17:06:17.974553 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-72h5n" event={"ID":"a9ccad70-dcd4-473c-aa09-925915f7420f","Type":"ContainerDied","Data":"b5fb2baf91e2321685f3b671aa08db5ac6c89d321bb1ffb21c1c8bbf0966fc5c"} Jan 26 17:06:17 crc kubenswrapper[4754]: I0126 17:06:17.974574 4754 scope.go:117] "RemoveContainer" containerID="1d9caa5e5f0c0418fd0df5b7a9010ebb5b310f1dd2c3774dd558998962e06810" Jan 26 17:06:17 crc kubenswrapper[4754]: I0126 17:06:17.983318 4754 generic.go:334] "Generic (PLEG): container finished" podID="0df51b10-10ea-4822-ad9b-e952ba4673f8" containerID="62dc8ee2c3e18b05a5ad1bdbc9c5d7af30f0f506b69d408c389b431fec3b3fb8" exitCode=0 Jan 26 17:06:17 crc kubenswrapper[4754]: I0126 17:06:17.984795 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-r9jrq" event={"ID":"0df51b10-10ea-4822-ad9b-e952ba4673f8","Type":"ContainerDied","Data":"62dc8ee2c3e18b05a5ad1bdbc9c5d7af30f0f506b69d408c389b431fec3b3fb8"} Jan 26 17:06:18 crc kubenswrapper[4754]: I0126 17:06:18.010859 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7f896c8c65-jbhcp" podStartSLOduration=3.010821979 podStartE2EDuration="3.010821979s" podCreationTimestamp="2026-01-26 17:06:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 17:06:17.994507563 +0000 UTC m=+1144.518688007" watchObservedRunningTime="2026-01-26 17:06:18.010821979 +0000 UTC m=+1144.535002413" Jan 26 17:06:18 crc kubenswrapper[4754]: I0126 17:06:18.017553 4754 scope.go:117] "RemoveContainer" containerID="9179318a0742938a4d3d8ad7707ca375118c0c3139a54b4fa56314d06e96104a" Jan 26 17:06:18 crc kubenswrapper[4754]: I0126 17:06:18.045555 4754 scope.go:117] "RemoveContainer" containerID="1d9caa5e5f0c0418fd0df5b7a9010ebb5b310f1dd2c3774dd558998962e06810" Jan 26 17:06:18 crc kubenswrapper[4754]: E0126 17:06:18.047002 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1d9caa5e5f0c0418fd0df5b7a9010ebb5b310f1dd2c3774dd558998962e06810\": container with ID starting with 1d9caa5e5f0c0418fd0df5b7a9010ebb5b310f1dd2c3774dd558998962e06810 not found: ID does not exist" containerID="1d9caa5e5f0c0418fd0df5b7a9010ebb5b310f1dd2c3774dd558998962e06810" Jan 26 17:06:18 crc kubenswrapper[4754]: I0126 17:06:18.047072 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1d9caa5e5f0c0418fd0df5b7a9010ebb5b310f1dd2c3774dd558998962e06810"} err="failed to get container status \"1d9caa5e5f0c0418fd0df5b7a9010ebb5b310f1dd2c3774dd558998962e06810\": rpc error: code = NotFound desc = could not find container \"1d9caa5e5f0c0418fd0df5b7a9010ebb5b310f1dd2c3774dd558998962e06810\": container with ID starting with 1d9caa5e5f0c0418fd0df5b7a9010ebb5b310f1dd2c3774dd558998962e06810 not found: ID does not exist" Jan 26 17:06:18 crc kubenswrapper[4754]: I0126 17:06:18.047107 4754 scope.go:117] "RemoveContainer" containerID="9179318a0742938a4d3d8ad7707ca375118c0c3139a54b4fa56314d06e96104a" Jan 26 17:06:18 crc kubenswrapper[4754]: E0126 17:06:18.048106 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9179318a0742938a4d3d8ad7707ca375118c0c3139a54b4fa56314d06e96104a\": container with ID starting with 9179318a0742938a4d3d8ad7707ca375118c0c3139a54b4fa56314d06e96104a not found: ID does not exist" containerID="9179318a0742938a4d3d8ad7707ca375118c0c3139a54b4fa56314d06e96104a" Jan 26 17:06:18 crc kubenswrapper[4754]: I0126 17:06:18.048222 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9179318a0742938a4d3d8ad7707ca375118c0c3139a54b4fa56314d06e96104a"} err="failed to get container status \"9179318a0742938a4d3d8ad7707ca375118c0c3139a54b4fa56314d06e96104a\": rpc error: code = NotFound desc = could not find container \"9179318a0742938a4d3d8ad7707ca375118c0c3139a54b4fa56314d06e96104a\": container with ID starting with 9179318a0742938a4d3d8ad7707ca375118c0c3139a54b4fa56314d06e96104a not found: ID does not exist" Jan 26 17:06:18 crc kubenswrapper[4754]: I0126 17:06:18.054511 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-98s2c\" (UniqueName: \"kubernetes.io/projected/a9ccad70-dcd4-473c-aa09-925915f7420f-kube-api-access-98s2c\") pod \"a9ccad70-dcd4-473c-aa09-925915f7420f\" (UID: \"a9ccad70-dcd4-473c-aa09-925915f7420f\") " Jan 26 17:06:18 crc kubenswrapper[4754]: I0126 17:06:18.055100 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a9ccad70-dcd4-473c-aa09-925915f7420f-dns-svc\") pod \"a9ccad70-dcd4-473c-aa09-925915f7420f\" (UID: \"a9ccad70-dcd4-473c-aa09-925915f7420f\") " Jan 26 17:06:18 crc kubenswrapper[4754]: I0126 17:06:18.055163 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a9ccad70-dcd4-473c-aa09-925915f7420f-config\") pod \"a9ccad70-dcd4-473c-aa09-925915f7420f\" (UID: \"a9ccad70-dcd4-473c-aa09-925915f7420f\") " Jan 26 17:06:18 crc kubenswrapper[4754]: I0126 17:06:18.061727 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a9ccad70-dcd4-473c-aa09-925915f7420f-kube-api-access-98s2c" (OuterVolumeSpecName: "kube-api-access-98s2c") pod "a9ccad70-dcd4-473c-aa09-925915f7420f" (UID: "a9ccad70-dcd4-473c-aa09-925915f7420f"). InnerVolumeSpecName "kube-api-access-98s2c". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:06:18 crc kubenswrapper[4754]: I0126 17:06:18.102797 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a9ccad70-dcd4-473c-aa09-925915f7420f-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "a9ccad70-dcd4-473c-aa09-925915f7420f" (UID: "a9ccad70-dcd4-473c-aa09-925915f7420f"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:06:18 crc kubenswrapper[4754]: I0126 17:06:18.115844 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a9ccad70-dcd4-473c-aa09-925915f7420f-config" (OuterVolumeSpecName: "config") pod "a9ccad70-dcd4-473c-aa09-925915f7420f" (UID: "a9ccad70-dcd4-473c-aa09-925915f7420f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:06:18 crc kubenswrapper[4754]: I0126 17:06:18.156866 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-98s2c\" (UniqueName: \"kubernetes.io/projected/a9ccad70-dcd4-473c-aa09-925915f7420f-kube-api-access-98s2c\") on node \"crc\" DevicePath \"\"" Jan 26 17:06:18 crc kubenswrapper[4754]: I0126 17:06:18.156899 4754 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a9ccad70-dcd4-473c-aa09-925915f7420f-dns-svc\") on node \"crc\" DevicePath \"\"" Jan 26 17:06:18 crc kubenswrapper[4754]: I0126 17:06:18.156910 4754 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a9ccad70-dcd4-473c-aa09-925915f7420f-config\") on node \"crc\" DevicePath \"\"" Jan 26 17:06:18 crc kubenswrapper[4754]: I0126 17:06:18.302386 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-72h5n"] Jan 26 17:06:18 crc kubenswrapper[4754]: I0126 17:06:18.309761 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-72h5n"] Jan 26 17:06:18 crc kubenswrapper[4754]: I0126 17:06:18.994656 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-r9jrq" event={"ID":"0df51b10-10ea-4822-ad9b-e952ba4673f8","Type":"ContainerStarted","Data":"cd0b28dc0eb9a531434655b61d84d44919718ba6e136e2e784afe94300f332f5"} Jan 26 17:06:18 crc kubenswrapper[4754]: I0126 17:06:18.995078 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-86db49b7ff-r9jrq" Jan 26 17:06:18 crc kubenswrapper[4754]: I0126 17:06:18.996622 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"4af75510-1f61-40f7-a292-764facf90f1a","Type":"ContainerStarted","Data":"5a460d74bfe649313c2ed6dbe24cf056bd2097cf120ecbddb82b88aca0b9feb5"} Jan 26 17:06:18 crc kubenswrapper[4754]: I0126 17:06:18.996716 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"4af75510-1f61-40f7-a292-764facf90f1a","Type":"ContainerStarted","Data":"bc5bcda2b136c51dfe4be797c60841d96f20afcdb75c1e109ceb6358bef1ddd7"} Jan 26 17:06:19 crc kubenswrapper[4754]: I0126 17:06:19.015254 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-86db49b7ff-r9jrq" podStartSLOduration=4.01523618 podStartE2EDuration="4.01523618s" podCreationTimestamp="2026-01-26 17:06:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 17:06:19.012465723 +0000 UTC m=+1145.536646167" watchObservedRunningTime="2026-01-26 17:06:19.01523618 +0000 UTC m=+1145.539416614" Jan 26 17:06:19 crc kubenswrapper[4754]: I0126 17:06:19.049866 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=2.647791945 podStartE2EDuration="4.039639792s" podCreationTimestamp="2026-01-26 17:06:15 +0000 UTC" firstStartedPulling="2026-01-26 17:06:16.478076365 +0000 UTC m=+1143.002256799" lastFinishedPulling="2026-01-26 17:06:17.869924212 +0000 UTC m=+1144.394104646" observedRunningTime="2026-01-26 17:06:19.033891542 +0000 UTC m=+1145.558072036" watchObservedRunningTime="2026-01-26 17:06:19.039639792 +0000 UTC m=+1145.563820226" Jan 26 17:06:19 crc kubenswrapper[4754]: I0126 17:06:19.348103 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Jan 26 17:06:19 crc kubenswrapper[4754]: I0126 17:06:19.348259 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Jan 26 17:06:19 crc kubenswrapper[4754]: I0126 17:06:19.436831 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Jan 26 17:06:19 crc kubenswrapper[4754]: I0126 17:06:19.783432 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a9ccad70-dcd4-473c-aa09-925915f7420f" path="/var/lib/kubelet/pods/a9ccad70-dcd4-473c-aa09-925915f7420f/volumes" Jan 26 17:06:20 crc kubenswrapper[4754]: I0126 17:06:20.006274 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Jan 26 17:06:20 crc kubenswrapper[4754]: I0126 17:06:20.086721 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Jan 26 17:06:20 crc kubenswrapper[4754]: I0126 17:06:20.593811 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-f12b-account-create-update-scv4g"] Jan 26 17:06:20 crc kubenswrapper[4754]: E0126 17:06:20.594745 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c4637fc0-04b8-4b18-8986-7c0cebe93d45" containerName="init" Jan 26 17:06:20 crc kubenswrapper[4754]: I0126 17:06:20.594769 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="c4637fc0-04b8-4b18-8986-7c0cebe93d45" containerName="init" Jan 26 17:06:20 crc kubenswrapper[4754]: E0126 17:06:20.594795 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9ccad70-dcd4-473c-aa09-925915f7420f" containerName="init" Jan 26 17:06:20 crc kubenswrapper[4754]: I0126 17:06:20.594804 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9ccad70-dcd4-473c-aa09-925915f7420f" containerName="init" Jan 26 17:06:20 crc kubenswrapper[4754]: E0126 17:06:20.594822 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c4637fc0-04b8-4b18-8986-7c0cebe93d45" containerName="dnsmasq-dns" Jan 26 17:06:20 crc kubenswrapper[4754]: I0126 17:06:20.594841 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="c4637fc0-04b8-4b18-8986-7c0cebe93d45" containerName="dnsmasq-dns" Jan 26 17:06:20 crc kubenswrapper[4754]: E0126 17:06:20.594871 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9ccad70-dcd4-473c-aa09-925915f7420f" containerName="dnsmasq-dns" Jan 26 17:06:20 crc kubenswrapper[4754]: I0126 17:06:20.594879 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9ccad70-dcd4-473c-aa09-925915f7420f" containerName="dnsmasq-dns" Jan 26 17:06:20 crc kubenswrapper[4754]: I0126 17:06:20.595066 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="a9ccad70-dcd4-473c-aa09-925915f7420f" containerName="dnsmasq-dns" Jan 26 17:06:20 crc kubenswrapper[4754]: I0126 17:06:20.595089 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="c4637fc0-04b8-4b18-8986-7c0cebe93d45" containerName="dnsmasq-dns" Jan 26 17:06:20 crc kubenswrapper[4754]: I0126 17:06:20.595757 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-f12b-account-create-update-scv4g" Jan 26 17:06:20 crc kubenswrapper[4754]: I0126 17:06:20.598181 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Jan 26 17:06:20 crc kubenswrapper[4754]: I0126 17:06:20.602216 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-8dnlq"] Jan 26 17:06:20 crc kubenswrapper[4754]: I0126 17:06:20.603311 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-8dnlq" Jan 26 17:06:20 crc kubenswrapper[4754]: I0126 17:06:20.610061 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-f12b-account-create-update-scv4g"] Jan 26 17:06:20 crc kubenswrapper[4754]: I0126 17:06:20.624753 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-8dnlq"] Jan 26 17:06:20 crc kubenswrapper[4754]: I0126 17:06:20.697828 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bw69j\" (UniqueName: \"kubernetes.io/projected/d1acc0d8-ac6a-4b56-bc9b-d926176314c3-kube-api-access-bw69j\") pod \"keystone-f12b-account-create-update-scv4g\" (UID: \"d1acc0d8-ac6a-4b56-bc9b-d926176314c3\") " pod="openstack/keystone-f12b-account-create-update-scv4g" Jan 26 17:06:20 crc kubenswrapper[4754]: I0126 17:06:20.697880 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9295f950-6f88-42d3-a82e-688cda9cce76-operator-scripts\") pod \"keystone-db-create-8dnlq\" (UID: \"9295f950-6f88-42d3-a82e-688cda9cce76\") " pod="openstack/keystone-db-create-8dnlq" Jan 26 17:06:20 crc kubenswrapper[4754]: I0126 17:06:20.697926 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d1acc0d8-ac6a-4b56-bc9b-d926176314c3-operator-scripts\") pod \"keystone-f12b-account-create-update-scv4g\" (UID: \"d1acc0d8-ac6a-4b56-bc9b-d926176314c3\") " pod="openstack/keystone-f12b-account-create-update-scv4g" Jan 26 17:06:20 crc kubenswrapper[4754]: I0126 17:06:20.698172 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4prbm\" (UniqueName: \"kubernetes.io/projected/9295f950-6f88-42d3-a82e-688cda9cce76-kube-api-access-4prbm\") pod \"keystone-db-create-8dnlq\" (UID: \"9295f950-6f88-42d3-a82e-688cda9cce76\") " pod="openstack/keystone-db-create-8dnlq" Jan 26 17:06:20 crc kubenswrapper[4754]: I0126 17:06:20.792821 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-c7cxf"] Jan 26 17:06:20 crc kubenswrapper[4754]: I0126 17:06:20.794269 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-c7cxf" Jan 26 17:06:20 crc kubenswrapper[4754]: I0126 17:06:20.799388 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bw69j\" (UniqueName: \"kubernetes.io/projected/d1acc0d8-ac6a-4b56-bc9b-d926176314c3-kube-api-access-bw69j\") pod \"keystone-f12b-account-create-update-scv4g\" (UID: \"d1acc0d8-ac6a-4b56-bc9b-d926176314c3\") " pod="openstack/keystone-f12b-account-create-update-scv4g" Jan 26 17:06:20 crc kubenswrapper[4754]: I0126 17:06:20.799439 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9295f950-6f88-42d3-a82e-688cda9cce76-operator-scripts\") pod \"keystone-db-create-8dnlq\" (UID: \"9295f950-6f88-42d3-a82e-688cda9cce76\") " pod="openstack/keystone-db-create-8dnlq" Jan 26 17:06:20 crc kubenswrapper[4754]: I0126 17:06:20.799500 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d1acc0d8-ac6a-4b56-bc9b-d926176314c3-operator-scripts\") pod \"keystone-f12b-account-create-update-scv4g\" (UID: \"d1acc0d8-ac6a-4b56-bc9b-d926176314c3\") " pod="openstack/keystone-f12b-account-create-update-scv4g" Jan 26 17:06:20 crc kubenswrapper[4754]: I0126 17:06:20.799593 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4prbm\" (UniqueName: \"kubernetes.io/projected/9295f950-6f88-42d3-a82e-688cda9cce76-kube-api-access-4prbm\") pod \"keystone-db-create-8dnlq\" (UID: \"9295f950-6f88-42d3-a82e-688cda9cce76\") " pod="openstack/keystone-db-create-8dnlq" Jan 26 17:06:20 crc kubenswrapper[4754]: I0126 17:06:20.800546 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d1acc0d8-ac6a-4b56-bc9b-d926176314c3-operator-scripts\") pod \"keystone-f12b-account-create-update-scv4g\" (UID: \"d1acc0d8-ac6a-4b56-bc9b-d926176314c3\") " pod="openstack/keystone-f12b-account-create-update-scv4g" Jan 26 17:06:20 crc kubenswrapper[4754]: I0126 17:06:20.800657 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9295f950-6f88-42d3-a82e-688cda9cce76-operator-scripts\") pod \"keystone-db-create-8dnlq\" (UID: \"9295f950-6f88-42d3-a82e-688cda9cce76\") " pod="openstack/keystone-db-create-8dnlq" Jan 26 17:06:20 crc kubenswrapper[4754]: I0126 17:06:20.801402 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-c7cxf"] Jan 26 17:06:20 crc kubenswrapper[4754]: I0126 17:06:20.825494 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4prbm\" (UniqueName: \"kubernetes.io/projected/9295f950-6f88-42d3-a82e-688cda9cce76-kube-api-access-4prbm\") pod \"keystone-db-create-8dnlq\" (UID: \"9295f950-6f88-42d3-a82e-688cda9cce76\") " pod="openstack/keystone-db-create-8dnlq" Jan 26 17:06:20 crc kubenswrapper[4754]: I0126 17:06:20.829861 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bw69j\" (UniqueName: \"kubernetes.io/projected/d1acc0d8-ac6a-4b56-bc9b-d926176314c3-kube-api-access-bw69j\") pod \"keystone-f12b-account-create-update-scv4g\" (UID: \"d1acc0d8-ac6a-4b56-bc9b-d926176314c3\") " pod="openstack/keystone-f12b-account-create-update-scv4g" Jan 26 17:06:20 crc kubenswrapper[4754]: I0126 17:06:20.901609 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x4c6r\" (UniqueName: \"kubernetes.io/projected/fe272255-848d-4761-a7f9-347103832e7d-kube-api-access-x4c6r\") pod \"placement-db-create-c7cxf\" (UID: \"fe272255-848d-4761-a7f9-347103832e7d\") " pod="openstack/placement-db-create-c7cxf" Jan 26 17:06:20 crc kubenswrapper[4754]: I0126 17:06:20.901722 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fe272255-848d-4761-a7f9-347103832e7d-operator-scripts\") pod \"placement-db-create-c7cxf\" (UID: \"fe272255-848d-4761-a7f9-347103832e7d\") " pod="openstack/placement-db-create-c7cxf" Jan 26 17:06:20 crc kubenswrapper[4754]: I0126 17:06:20.920606 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-4c00-account-create-update-h9g8d"] Jan 26 17:06:20 crc kubenswrapper[4754]: I0126 17:06:20.922077 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-f12b-account-create-update-scv4g" Jan 26 17:06:20 crc kubenswrapper[4754]: I0126 17:06:20.922335 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-4c00-account-create-update-h9g8d" Jan 26 17:06:20 crc kubenswrapper[4754]: I0126 17:06:20.928130 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-4c00-account-create-update-h9g8d"] Jan 26 17:06:20 crc kubenswrapper[4754]: I0126 17:06:20.929954 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Jan 26 17:06:20 crc kubenswrapper[4754]: I0126 17:06:20.930098 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-8dnlq" Jan 26 17:06:21 crc kubenswrapper[4754]: I0126 17:06:21.003797 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fe272255-848d-4761-a7f9-347103832e7d-operator-scripts\") pod \"placement-db-create-c7cxf\" (UID: \"fe272255-848d-4761-a7f9-347103832e7d\") " pod="openstack/placement-db-create-c7cxf" Jan 26 17:06:21 crc kubenswrapper[4754]: I0126 17:06:21.003919 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z5vl4\" (UniqueName: \"kubernetes.io/projected/dce62524-1013-4b38-b563-d9f6d7701523-kube-api-access-z5vl4\") pod \"placement-4c00-account-create-update-h9g8d\" (UID: \"dce62524-1013-4b38-b563-d9f6d7701523\") " pod="openstack/placement-4c00-account-create-update-h9g8d" Jan 26 17:06:21 crc kubenswrapper[4754]: I0126 17:06:21.003966 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dce62524-1013-4b38-b563-d9f6d7701523-operator-scripts\") pod \"placement-4c00-account-create-update-h9g8d\" (UID: \"dce62524-1013-4b38-b563-d9f6d7701523\") " pod="openstack/placement-4c00-account-create-update-h9g8d" Jan 26 17:06:21 crc kubenswrapper[4754]: I0126 17:06:21.004178 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x4c6r\" (UniqueName: \"kubernetes.io/projected/fe272255-848d-4761-a7f9-347103832e7d-kube-api-access-x4c6r\") pod \"placement-db-create-c7cxf\" (UID: \"fe272255-848d-4761-a7f9-347103832e7d\") " pod="openstack/placement-db-create-c7cxf" Jan 26 17:06:21 crc kubenswrapper[4754]: I0126 17:06:21.004853 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fe272255-848d-4761-a7f9-347103832e7d-operator-scripts\") pod \"placement-db-create-c7cxf\" (UID: \"fe272255-848d-4761-a7f9-347103832e7d\") " pod="openstack/placement-db-create-c7cxf" Jan 26 17:06:21 crc kubenswrapper[4754]: I0126 17:06:21.028276 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x4c6r\" (UniqueName: \"kubernetes.io/projected/fe272255-848d-4761-a7f9-347103832e7d-kube-api-access-x4c6r\") pod \"placement-db-create-c7cxf\" (UID: \"fe272255-848d-4761-a7f9-347103832e7d\") " pod="openstack/placement-db-create-c7cxf" Jan 26 17:06:21 crc kubenswrapper[4754]: I0126 17:06:21.040376 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Jan 26 17:06:21 crc kubenswrapper[4754]: I0126 17:06:21.040415 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Jan 26 17:06:21 crc kubenswrapper[4754]: I0126 17:06:21.105658 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z5vl4\" (UniqueName: \"kubernetes.io/projected/dce62524-1013-4b38-b563-d9f6d7701523-kube-api-access-z5vl4\") pod \"placement-4c00-account-create-update-h9g8d\" (UID: \"dce62524-1013-4b38-b563-d9f6d7701523\") " pod="openstack/placement-4c00-account-create-update-h9g8d" Jan 26 17:06:21 crc kubenswrapper[4754]: I0126 17:06:21.105722 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dce62524-1013-4b38-b563-d9f6d7701523-operator-scripts\") pod \"placement-4c00-account-create-update-h9g8d\" (UID: \"dce62524-1013-4b38-b563-d9f6d7701523\") " pod="openstack/placement-4c00-account-create-update-h9g8d" Jan 26 17:06:21 crc kubenswrapper[4754]: I0126 17:06:21.107337 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dce62524-1013-4b38-b563-d9f6d7701523-operator-scripts\") pod \"placement-4c00-account-create-update-h9g8d\" (UID: \"dce62524-1013-4b38-b563-d9f6d7701523\") " pod="openstack/placement-4c00-account-create-update-h9g8d" Jan 26 17:06:21 crc kubenswrapper[4754]: I0126 17:06:21.107477 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Jan 26 17:06:21 crc kubenswrapper[4754]: I0126 17:06:21.111205 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-c7cxf" Jan 26 17:06:21 crc kubenswrapper[4754]: I0126 17:06:21.127431 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z5vl4\" (UniqueName: \"kubernetes.io/projected/dce62524-1013-4b38-b563-d9f6d7701523-kube-api-access-z5vl4\") pod \"placement-4c00-account-create-update-h9g8d\" (UID: \"dce62524-1013-4b38-b563-d9f6d7701523\") " pod="openstack/placement-4c00-account-create-update-h9g8d" Jan 26 17:06:21 crc kubenswrapper[4754]: I0126 17:06:21.303561 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-4c00-account-create-update-h9g8d" Jan 26 17:06:21 crc kubenswrapper[4754]: I0126 17:06:21.712372 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-f12b-account-create-update-scv4g"] Jan 26 17:06:21 crc kubenswrapper[4754]: I0126 17:06:21.810780 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-8dnlq"] Jan 26 17:06:21 crc kubenswrapper[4754]: I0126 17:06:21.884961 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-4c00-account-create-update-h9g8d"] Jan 26 17:06:21 crc kubenswrapper[4754]: W0126 17:06:21.890137 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddce62524_1013_4b38_b563_d9f6d7701523.slice/crio-c152c67eaf2f36d600bf5e9f7a9a2697bce7f6ef1f957453a2ba47daff11d86f WatchSource:0}: Error finding container c152c67eaf2f36d600bf5e9f7a9a2697bce7f6ef1f957453a2ba47daff11d86f: Status 404 returned error can't find the container with id c152c67eaf2f36d600bf5e9f7a9a2697bce7f6ef1f957453a2ba47daff11d86f Jan 26 17:06:21 crc kubenswrapper[4754]: I0126 17:06:21.890531 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-c7cxf"] Jan 26 17:06:22 crc kubenswrapper[4754]: I0126 17:06:22.024210 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-8dnlq" event={"ID":"9295f950-6f88-42d3-a82e-688cda9cce76","Type":"ContainerStarted","Data":"16604df2d74058ba3cd3c3c5e2450b605595cf261ff70a8ebc5b9b6ff15635d3"} Jan 26 17:06:22 crc kubenswrapper[4754]: I0126 17:06:22.025572 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-f12b-account-create-update-scv4g" event={"ID":"d1acc0d8-ac6a-4b56-bc9b-d926176314c3","Type":"ContainerStarted","Data":"04468a86c787080e75988037b5a37583ed4cf476f3870d033abb5e4371717255"} Jan 26 17:06:22 crc kubenswrapper[4754]: I0126 17:06:22.026712 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-4c00-account-create-update-h9g8d" event={"ID":"dce62524-1013-4b38-b563-d9f6d7701523","Type":"ContainerStarted","Data":"c152c67eaf2f36d600bf5e9f7a9a2697bce7f6ef1f957453a2ba47daff11d86f"} Jan 26 17:06:22 crc kubenswrapper[4754]: I0126 17:06:22.028282 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-c7cxf" event={"ID":"fe272255-848d-4761-a7f9-347103832e7d","Type":"ContainerStarted","Data":"e278e71b5578190df55aa6d3359fae2229bc1a7a1e4c18abe9f86260fcac7d70"} Jan 26 17:06:22 crc kubenswrapper[4754]: I0126 17:06:22.106469 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Jan 26 17:06:23 crc kubenswrapper[4754]: I0126 17:06:23.030922 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7f896c8c65-jbhcp"] Jan 26 17:06:23 crc kubenswrapper[4754]: I0126 17:06:23.031551 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7f896c8c65-jbhcp" podUID="0bf413d0-f8cf-4574-b4bc-dfa3264a41b9" containerName="dnsmasq-dns" containerID="cri-o://93a7968f52e595c1a52011c2b828412ae7f861a72b4ce6f0b29ff7953e389879" gracePeriod=10 Jan 26 17:06:23 crc kubenswrapper[4754]: I0126 17:06:23.036844 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7f896c8c65-jbhcp" Jan 26 17:06:23 crc kubenswrapper[4754]: I0126 17:06:23.066952 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-4c00-account-create-update-h9g8d" event={"ID":"dce62524-1013-4b38-b563-d9f6d7701523","Type":"ContainerStarted","Data":"f9f718438d30237bbc332cf9180c7bba7606646662f6c6d7426532bcfad633bf"} Jan 26 17:06:23 crc kubenswrapper[4754]: I0126 17:06:23.087977 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-c7cxf" event={"ID":"fe272255-848d-4761-a7f9-347103832e7d","Type":"ContainerStarted","Data":"354866a3d7b6dedd7082b7847d6eb060b0999db466f2424c5454aebfe3938d6a"} Jan 26 17:06:23 crc kubenswrapper[4754]: I0126 17:06:23.091900 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-698758b865-z5pcq"] Jan 26 17:06:23 crc kubenswrapper[4754]: I0126 17:06:23.096107 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-z5pcq" Jan 26 17:06:23 crc kubenswrapper[4754]: I0126 17:06:23.107355 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-698758b865-z5pcq"] Jan 26 17:06:23 crc kubenswrapper[4754]: I0126 17:06:23.110737 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-8dnlq" event={"ID":"9295f950-6f88-42d3-a82e-688cda9cce76","Type":"ContainerStarted","Data":"b4c825db5f3f0d1ad8483b5b742fdd62df9527e36ac7253d696c7b31652165e7"} Jan 26 17:06:23 crc kubenswrapper[4754]: I0126 17:06:23.122020 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-f12b-account-create-update-scv4g" event={"ID":"d1acc0d8-ac6a-4b56-bc9b-d926176314c3","Type":"ContainerStarted","Data":"88006dee24ddc93951251d2d1931aa8bd96e1a1ba2aef15158a5bd9c64ebd982"} Jan 26 17:06:23 crc kubenswrapper[4754]: I0126 17:06:23.134043 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-4c00-account-create-update-h9g8d" podStartSLOduration=3.1340252570000002 podStartE2EDuration="3.134025257s" podCreationTimestamp="2026-01-26 17:06:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 17:06:23.120145919 +0000 UTC m=+1149.644326353" watchObservedRunningTime="2026-01-26 17:06:23.134025257 +0000 UTC m=+1149.658205691" Jan 26 17:06:23 crc kubenswrapper[4754]: I0126 17:06:23.141590 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b5317650-1469-4237-8645-6031475c5b37-config\") pod \"dnsmasq-dns-698758b865-z5pcq\" (UID: \"b5317650-1469-4237-8645-6031475c5b37\") " pod="openstack/dnsmasq-dns-698758b865-z5pcq" Jan 26 17:06:23 crc kubenswrapper[4754]: I0126 17:06:23.141661 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b5317650-1469-4237-8645-6031475c5b37-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-z5pcq\" (UID: \"b5317650-1469-4237-8645-6031475c5b37\") " pod="openstack/dnsmasq-dns-698758b865-z5pcq" Jan 26 17:06:23 crc kubenswrapper[4754]: I0126 17:06:23.141754 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b5317650-1469-4237-8645-6031475c5b37-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-z5pcq\" (UID: \"b5317650-1469-4237-8645-6031475c5b37\") " pod="openstack/dnsmasq-dns-698758b865-z5pcq" Jan 26 17:06:23 crc kubenswrapper[4754]: I0126 17:06:23.141772 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b5317650-1469-4237-8645-6031475c5b37-dns-svc\") pod \"dnsmasq-dns-698758b865-z5pcq\" (UID: \"b5317650-1469-4237-8645-6031475c5b37\") " pod="openstack/dnsmasq-dns-698758b865-z5pcq" Jan 26 17:06:23 crc kubenswrapper[4754]: I0126 17:06:23.141807 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hxqbk\" (UniqueName: \"kubernetes.io/projected/b5317650-1469-4237-8645-6031475c5b37-kube-api-access-hxqbk\") pod \"dnsmasq-dns-698758b865-z5pcq\" (UID: \"b5317650-1469-4237-8645-6031475c5b37\") " pod="openstack/dnsmasq-dns-698758b865-z5pcq" Jan 26 17:06:23 crc kubenswrapper[4754]: I0126 17:06:23.172399 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-create-c7cxf" podStartSLOduration=3.172373427 podStartE2EDuration="3.172373427s" podCreationTimestamp="2026-01-26 17:06:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 17:06:23.152015359 +0000 UTC m=+1149.676195793" watchObservedRunningTime="2026-01-26 17:06:23.172373427 +0000 UTC m=+1149.696553871" Jan 26 17:06:23 crc kubenswrapper[4754]: I0126 17:06:23.213766 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-create-8dnlq" podStartSLOduration=3.213747004 podStartE2EDuration="3.213747004s" podCreationTimestamp="2026-01-26 17:06:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 17:06:23.205689828 +0000 UTC m=+1149.729870262" watchObservedRunningTime="2026-01-26 17:06:23.213747004 +0000 UTC m=+1149.737927438" Jan 26 17:06:23 crc kubenswrapper[4754]: I0126 17:06:23.248040 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b5317650-1469-4237-8645-6031475c5b37-config\") pod \"dnsmasq-dns-698758b865-z5pcq\" (UID: \"b5317650-1469-4237-8645-6031475c5b37\") " pod="openstack/dnsmasq-dns-698758b865-z5pcq" Jan 26 17:06:23 crc kubenswrapper[4754]: I0126 17:06:23.248189 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b5317650-1469-4237-8645-6031475c5b37-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-z5pcq\" (UID: \"b5317650-1469-4237-8645-6031475c5b37\") " pod="openstack/dnsmasq-dns-698758b865-z5pcq" Jan 26 17:06:23 crc kubenswrapper[4754]: I0126 17:06:23.248927 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b5317650-1469-4237-8645-6031475c5b37-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-z5pcq\" (UID: \"b5317650-1469-4237-8645-6031475c5b37\") " pod="openstack/dnsmasq-dns-698758b865-z5pcq" Jan 26 17:06:23 crc kubenswrapper[4754]: I0126 17:06:23.248969 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b5317650-1469-4237-8645-6031475c5b37-dns-svc\") pod \"dnsmasq-dns-698758b865-z5pcq\" (UID: \"b5317650-1469-4237-8645-6031475c5b37\") " pod="openstack/dnsmasq-dns-698758b865-z5pcq" Jan 26 17:06:23 crc kubenswrapper[4754]: I0126 17:06:23.249084 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hxqbk\" (UniqueName: \"kubernetes.io/projected/b5317650-1469-4237-8645-6031475c5b37-kube-api-access-hxqbk\") pod \"dnsmasq-dns-698758b865-z5pcq\" (UID: \"b5317650-1469-4237-8645-6031475c5b37\") " pod="openstack/dnsmasq-dns-698758b865-z5pcq" Jan 26 17:06:23 crc kubenswrapper[4754]: I0126 17:06:23.249538 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b5317650-1469-4237-8645-6031475c5b37-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-z5pcq\" (UID: \"b5317650-1469-4237-8645-6031475c5b37\") " pod="openstack/dnsmasq-dns-698758b865-z5pcq" Jan 26 17:06:23 crc kubenswrapper[4754]: I0126 17:06:23.250516 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b5317650-1469-4237-8645-6031475c5b37-config\") pod \"dnsmasq-dns-698758b865-z5pcq\" (UID: \"b5317650-1469-4237-8645-6031475c5b37\") " pod="openstack/dnsmasq-dns-698758b865-z5pcq" Jan 26 17:06:23 crc kubenswrapper[4754]: I0126 17:06:23.261457 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b5317650-1469-4237-8645-6031475c5b37-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-z5pcq\" (UID: \"b5317650-1469-4237-8645-6031475c5b37\") " pod="openstack/dnsmasq-dns-698758b865-z5pcq" Jan 26 17:06:23 crc kubenswrapper[4754]: I0126 17:06:23.262033 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b5317650-1469-4237-8645-6031475c5b37-dns-svc\") pod \"dnsmasq-dns-698758b865-z5pcq\" (UID: \"b5317650-1469-4237-8645-6031475c5b37\") " pod="openstack/dnsmasq-dns-698758b865-z5pcq" Jan 26 17:06:23 crc kubenswrapper[4754]: I0126 17:06:23.262344 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-f12b-account-create-update-scv4g" podStartSLOduration=3.262328841 podStartE2EDuration="3.262328841s" podCreationTimestamp="2026-01-26 17:06:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 17:06:23.246087028 +0000 UTC m=+1149.770267462" watchObservedRunningTime="2026-01-26 17:06:23.262328841 +0000 UTC m=+1149.786509275" Jan 26 17:06:23 crc kubenswrapper[4754]: I0126 17:06:23.292821 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hxqbk\" (UniqueName: \"kubernetes.io/projected/b5317650-1469-4237-8645-6031475c5b37-kube-api-access-hxqbk\") pod \"dnsmasq-dns-698758b865-z5pcq\" (UID: \"b5317650-1469-4237-8645-6031475c5b37\") " pod="openstack/dnsmasq-dns-698758b865-z5pcq" Jan 26 17:06:23 crc kubenswrapper[4754]: I0126 17:06:23.433353 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-z5pcq" Jan 26 17:06:24 crc kubenswrapper[4754]: I0126 17:06:24.071104 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7f896c8c65-jbhcp" Jan 26 17:06:24 crc kubenswrapper[4754]: I0126 17:06:24.131061 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"c3c55551-6de7-403e-bbf5-46a2ec685c13","Type":"ContainerStarted","Data":"4774b9533f2557b915630e35e4b458e258217842d027febdef6f54ade9c3e270"} Jan 26 17:06:24 crc kubenswrapper[4754]: I0126 17:06:24.132149 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Jan 26 17:06:24 crc kubenswrapper[4754]: I0126 17:06:24.134925 4754 generic.go:334] "Generic (PLEG): container finished" podID="d1acc0d8-ac6a-4b56-bc9b-d926176314c3" containerID="88006dee24ddc93951251d2d1931aa8bd96e1a1ba2aef15158a5bd9c64ebd982" exitCode=0 Jan 26 17:06:24 crc kubenswrapper[4754]: I0126 17:06:24.134991 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-f12b-account-create-update-scv4g" event={"ID":"d1acc0d8-ac6a-4b56-bc9b-d926176314c3","Type":"ContainerDied","Data":"88006dee24ddc93951251d2d1931aa8bd96e1a1ba2aef15158a5bd9c64ebd982"} Jan 26 17:06:24 crc kubenswrapper[4754]: I0126 17:06:24.137406 4754 generic.go:334] "Generic (PLEG): container finished" podID="dce62524-1013-4b38-b563-d9f6d7701523" containerID="f9f718438d30237bbc332cf9180c7bba7606646662f6c6d7426532bcfad633bf" exitCode=0 Jan 26 17:06:24 crc kubenswrapper[4754]: I0126 17:06:24.137471 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-4c00-account-create-update-h9g8d" event={"ID":"dce62524-1013-4b38-b563-d9f6d7701523","Type":"ContainerDied","Data":"f9f718438d30237bbc332cf9180c7bba7606646662f6c6d7426532bcfad633bf"} Jan 26 17:06:24 crc kubenswrapper[4754]: I0126 17:06:24.139644 4754 generic.go:334] "Generic (PLEG): container finished" podID="fe272255-848d-4761-a7f9-347103832e7d" containerID="354866a3d7b6dedd7082b7847d6eb060b0999db466f2424c5454aebfe3938d6a" exitCode=0 Jan 26 17:06:24 crc kubenswrapper[4754]: I0126 17:06:24.139694 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-c7cxf" event={"ID":"fe272255-848d-4761-a7f9-347103832e7d","Type":"ContainerDied","Data":"354866a3d7b6dedd7082b7847d6eb060b0999db466f2424c5454aebfe3938d6a"} Jan 26 17:06:24 crc kubenswrapper[4754]: I0126 17:06:24.142011 4754 generic.go:334] "Generic (PLEG): container finished" podID="0bf413d0-f8cf-4574-b4bc-dfa3264a41b9" containerID="93a7968f52e595c1a52011c2b828412ae7f861a72b4ce6f0b29ff7953e389879" exitCode=0 Jan 26 17:06:24 crc kubenswrapper[4754]: I0126 17:06:24.142097 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7f896c8c65-jbhcp" event={"ID":"0bf413d0-f8cf-4574-b4bc-dfa3264a41b9","Type":"ContainerDied","Data":"93a7968f52e595c1a52011c2b828412ae7f861a72b4ce6f0b29ff7953e389879"} Jan 26 17:06:24 crc kubenswrapper[4754]: I0126 17:06:24.142127 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7f896c8c65-jbhcp" event={"ID":"0bf413d0-f8cf-4574-b4bc-dfa3264a41b9","Type":"ContainerDied","Data":"b5ffcba758c81e5fab424c6de16c7df0a82d482567a6d60c6132f2be69ac129c"} Jan 26 17:06:24 crc kubenswrapper[4754]: I0126 17:06:24.142137 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7f896c8c65-jbhcp" Jan 26 17:06:24 crc kubenswrapper[4754]: I0126 17:06:24.142154 4754 scope.go:117] "RemoveContainer" containerID="93a7968f52e595c1a52011c2b828412ae7f861a72b4ce6f0b29ff7953e389879" Jan 26 17:06:24 crc kubenswrapper[4754]: I0126 17:06:24.143878 4754 generic.go:334] "Generic (PLEG): container finished" podID="9295f950-6f88-42d3-a82e-688cda9cce76" containerID="b4c825db5f3f0d1ad8483b5b742fdd62df9527e36ac7253d696c7b31652165e7" exitCode=0 Jan 26 17:06:24 crc kubenswrapper[4754]: I0126 17:06:24.143915 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-8dnlq" event={"ID":"9295f950-6f88-42d3-a82e-688cda9cce76","Type":"ContainerDied","Data":"b4c825db5f3f0d1ad8483b5b742fdd62df9527e36ac7253d696c7b31652165e7"} Jan 26 17:06:24 crc kubenswrapper[4754]: I0126 17:06:24.153269 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=17.477936298 podStartE2EDuration="42.153250672s" podCreationTimestamp="2026-01-26 17:05:42 +0000 UTC" firstStartedPulling="2026-01-26 17:05:59.147037603 +0000 UTC m=+1125.671218037" lastFinishedPulling="2026-01-26 17:06:23.822351977 +0000 UTC m=+1150.346532411" observedRunningTime="2026-01-26 17:06:24.149760765 +0000 UTC m=+1150.673941219" watchObservedRunningTime="2026-01-26 17:06:24.153250672 +0000 UTC m=+1150.677431106" Jan 26 17:06:24 crc kubenswrapper[4754]: I0126 17:06:24.163274 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vzwcn\" (UniqueName: \"kubernetes.io/projected/0bf413d0-f8cf-4574-b4bc-dfa3264a41b9-kube-api-access-vzwcn\") pod \"0bf413d0-f8cf-4574-b4bc-dfa3264a41b9\" (UID: \"0bf413d0-f8cf-4574-b4bc-dfa3264a41b9\") " Jan 26 17:06:24 crc kubenswrapper[4754]: I0126 17:06:24.163339 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0bf413d0-f8cf-4574-b4bc-dfa3264a41b9-dns-svc\") pod \"0bf413d0-f8cf-4574-b4bc-dfa3264a41b9\" (UID: \"0bf413d0-f8cf-4574-b4bc-dfa3264a41b9\") " Jan 26 17:06:24 crc kubenswrapper[4754]: I0126 17:06:24.163380 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0bf413d0-f8cf-4574-b4bc-dfa3264a41b9-config\") pod \"0bf413d0-f8cf-4574-b4bc-dfa3264a41b9\" (UID: \"0bf413d0-f8cf-4574-b4bc-dfa3264a41b9\") " Jan 26 17:06:24 crc kubenswrapper[4754]: I0126 17:06:24.163520 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0bf413d0-f8cf-4574-b4bc-dfa3264a41b9-ovsdbserver-sb\") pod \"0bf413d0-f8cf-4574-b4bc-dfa3264a41b9\" (UID: \"0bf413d0-f8cf-4574-b4bc-dfa3264a41b9\") " Jan 26 17:06:24 crc kubenswrapper[4754]: I0126 17:06:24.165795 4754 scope.go:117] "RemoveContainer" containerID="2b608aaa53ab4942a14be828c91bb9b5fbc5de669bb71c595bbfff4931865335" Jan 26 17:06:24 crc kubenswrapper[4754]: I0126 17:06:24.243134 4754 scope.go:117] "RemoveContainer" containerID="93a7968f52e595c1a52011c2b828412ae7f861a72b4ce6f0b29ff7953e389879" Jan 26 17:06:24 crc kubenswrapper[4754]: I0126 17:06:24.245896 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0bf413d0-f8cf-4574-b4bc-dfa3264a41b9-kube-api-access-vzwcn" (OuterVolumeSpecName: "kube-api-access-vzwcn") pod "0bf413d0-f8cf-4574-b4bc-dfa3264a41b9" (UID: "0bf413d0-f8cf-4574-b4bc-dfa3264a41b9"). InnerVolumeSpecName "kube-api-access-vzwcn". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:06:24 crc kubenswrapper[4754]: E0126 17:06:24.269468 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"93a7968f52e595c1a52011c2b828412ae7f861a72b4ce6f0b29ff7953e389879\": container with ID starting with 93a7968f52e595c1a52011c2b828412ae7f861a72b4ce6f0b29ff7953e389879 not found: ID does not exist" containerID="93a7968f52e595c1a52011c2b828412ae7f861a72b4ce6f0b29ff7953e389879" Jan 26 17:06:24 crc kubenswrapper[4754]: I0126 17:06:24.269524 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"93a7968f52e595c1a52011c2b828412ae7f861a72b4ce6f0b29ff7953e389879"} err="failed to get container status \"93a7968f52e595c1a52011c2b828412ae7f861a72b4ce6f0b29ff7953e389879\": rpc error: code = NotFound desc = could not find container \"93a7968f52e595c1a52011c2b828412ae7f861a72b4ce6f0b29ff7953e389879\": container with ID starting with 93a7968f52e595c1a52011c2b828412ae7f861a72b4ce6f0b29ff7953e389879 not found: ID does not exist" Jan 26 17:06:24 crc kubenswrapper[4754]: I0126 17:06:24.269549 4754 scope.go:117] "RemoveContainer" containerID="2b608aaa53ab4942a14be828c91bb9b5fbc5de669bb71c595bbfff4931865335" Jan 26 17:06:24 crc kubenswrapper[4754]: I0126 17:06:24.271028 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vzwcn\" (UniqueName: \"kubernetes.io/projected/0bf413d0-f8cf-4574-b4bc-dfa3264a41b9-kube-api-access-vzwcn\") on node \"crc\" DevicePath \"\"" Jan 26 17:06:24 crc kubenswrapper[4754]: I0126 17:06:24.293197 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0bf413d0-f8cf-4574-b4bc-dfa3264a41b9-config" (OuterVolumeSpecName: "config") pod "0bf413d0-f8cf-4574-b4bc-dfa3264a41b9" (UID: "0bf413d0-f8cf-4574-b4bc-dfa3264a41b9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:06:24 crc kubenswrapper[4754]: E0126 17:06:24.293339 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2b608aaa53ab4942a14be828c91bb9b5fbc5de669bb71c595bbfff4931865335\": container with ID starting with 2b608aaa53ab4942a14be828c91bb9b5fbc5de669bb71c595bbfff4931865335 not found: ID does not exist" containerID="2b608aaa53ab4942a14be828c91bb9b5fbc5de669bb71c595bbfff4931865335" Jan 26 17:06:24 crc kubenswrapper[4754]: I0126 17:06:24.293479 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2b608aaa53ab4942a14be828c91bb9b5fbc5de669bb71c595bbfff4931865335"} err="failed to get container status \"2b608aaa53ab4942a14be828c91bb9b5fbc5de669bb71c595bbfff4931865335\": rpc error: code = NotFound desc = could not find container \"2b608aaa53ab4942a14be828c91bb9b5fbc5de669bb71c595bbfff4931865335\": container with ID starting with 2b608aaa53ab4942a14be828c91bb9b5fbc5de669bb71c595bbfff4931865335 not found: ID does not exist" Jan 26 17:06:24 crc kubenswrapper[4754]: I0126 17:06:24.310929 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Jan 26 17:06:24 crc kubenswrapper[4754]: I0126 17:06:24.311308 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0bf413d0-f8cf-4574-b4bc-dfa3264a41b9-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "0bf413d0-f8cf-4574-b4bc-dfa3264a41b9" (UID: "0bf413d0-f8cf-4574-b4bc-dfa3264a41b9"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:06:24 crc kubenswrapper[4754]: E0126 17:06:24.311595 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0bf413d0-f8cf-4574-b4bc-dfa3264a41b9" containerName="dnsmasq-dns" Jan 26 17:06:24 crc kubenswrapper[4754]: I0126 17:06:24.311697 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="0bf413d0-f8cf-4574-b4bc-dfa3264a41b9" containerName="dnsmasq-dns" Jan 26 17:06:24 crc kubenswrapper[4754]: E0126 17:06:24.311796 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0bf413d0-f8cf-4574-b4bc-dfa3264a41b9" containerName="init" Jan 26 17:06:24 crc kubenswrapper[4754]: I0126 17:06:24.311872 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="0bf413d0-f8cf-4574-b4bc-dfa3264a41b9" containerName="init" Jan 26 17:06:24 crc kubenswrapper[4754]: I0126 17:06:24.312188 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="0bf413d0-f8cf-4574-b4bc-dfa3264a41b9" containerName="dnsmasq-dns" Jan 26 17:06:24 crc kubenswrapper[4754]: I0126 17:06:24.317861 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Jan 26 17:06:24 crc kubenswrapper[4754]: I0126 17:06:24.325596 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-bfq27" Jan 26 17:06:24 crc kubenswrapper[4754]: I0126 17:06:24.326908 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Jan 26 17:06:24 crc kubenswrapper[4754]: I0126 17:06:24.326999 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Jan 26 17:06:24 crc kubenswrapper[4754]: I0126 17:06:24.327150 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Jan 26 17:06:24 crc kubenswrapper[4754]: I0126 17:06:24.341961 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-698758b865-z5pcq"] Jan 26 17:06:24 crc kubenswrapper[4754]: I0126 17:06:24.366108 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0bf413d0-f8cf-4574-b4bc-dfa3264a41b9-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "0bf413d0-f8cf-4574-b4bc-dfa3264a41b9" (UID: "0bf413d0-f8cf-4574-b4bc-dfa3264a41b9"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:06:24 crc kubenswrapper[4754]: I0126 17:06:24.374653 4754 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0bf413d0-f8cf-4574-b4bc-dfa3264a41b9-config\") on node \"crc\" DevicePath \"\"" Jan 26 17:06:24 crc kubenswrapper[4754]: I0126 17:06:24.374714 4754 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0bf413d0-f8cf-4574-b4bc-dfa3264a41b9-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Jan 26 17:06:24 crc kubenswrapper[4754]: I0126 17:06:24.374722 4754 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0bf413d0-f8cf-4574-b4bc-dfa3264a41b9-dns-svc\") on node \"crc\" DevicePath \"\"" Jan 26 17:06:24 crc kubenswrapper[4754]: I0126 17:06:24.435100 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Jan 26 17:06:24 crc kubenswrapper[4754]: I0126 17:06:24.475648 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/19a0730c-46d0-4029-a86b-812ba1664dcc-lock\") pod \"swift-storage-0\" (UID: \"19a0730c-46d0-4029-a86b-812ba1664dcc\") " pod="openstack/swift-storage-0" Jan 26 17:06:24 crc kubenswrapper[4754]: I0126 17:06:24.475758 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"swift-storage-0\" (UID: \"19a0730c-46d0-4029-a86b-812ba1664dcc\") " pod="openstack/swift-storage-0" Jan 26 17:06:24 crc kubenswrapper[4754]: I0126 17:06:24.475780 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19a0730c-46d0-4029-a86b-812ba1664dcc-combined-ca-bundle\") pod \"swift-storage-0\" (UID: \"19a0730c-46d0-4029-a86b-812ba1664dcc\") " pod="openstack/swift-storage-0" Jan 26 17:06:24 crc kubenswrapper[4754]: I0126 17:06:24.475806 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sfd4m\" (UniqueName: \"kubernetes.io/projected/19a0730c-46d0-4029-a86b-812ba1664dcc-kube-api-access-sfd4m\") pod \"swift-storage-0\" (UID: \"19a0730c-46d0-4029-a86b-812ba1664dcc\") " pod="openstack/swift-storage-0" Jan 26 17:06:24 crc kubenswrapper[4754]: I0126 17:06:24.475829 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/19a0730c-46d0-4029-a86b-812ba1664dcc-etc-swift\") pod \"swift-storage-0\" (UID: \"19a0730c-46d0-4029-a86b-812ba1664dcc\") " pod="openstack/swift-storage-0" Jan 26 17:06:24 crc kubenswrapper[4754]: I0126 17:06:24.475843 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/19a0730c-46d0-4029-a86b-812ba1664dcc-cache\") pod \"swift-storage-0\" (UID: \"19a0730c-46d0-4029-a86b-812ba1664dcc\") " pod="openstack/swift-storage-0" Jan 26 17:06:24 crc kubenswrapper[4754]: I0126 17:06:24.547635 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7f896c8c65-jbhcp"] Jan 26 17:06:24 crc kubenswrapper[4754]: I0126 17:06:24.555058 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7f896c8c65-jbhcp"] Jan 26 17:06:24 crc kubenswrapper[4754]: I0126 17:06:24.577558 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/19a0730c-46d0-4029-a86b-812ba1664dcc-lock\") pod \"swift-storage-0\" (UID: \"19a0730c-46d0-4029-a86b-812ba1664dcc\") " pod="openstack/swift-storage-0" Jan 26 17:06:24 crc kubenswrapper[4754]: I0126 17:06:24.577620 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"swift-storage-0\" (UID: \"19a0730c-46d0-4029-a86b-812ba1664dcc\") " pod="openstack/swift-storage-0" Jan 26 17:06:24 crc kubenswrapper[4754]: I0126 17:06:24.577641 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19a0730c-46d0-4029-a86b-812ba1664dcc-combined-ca-bundle\") pod \"swift-storage-0\" (UID: \"19a0730c-46d0-4029-a86b-812ba1664dcc\") " pod="openstack/swift-storage-0" Jan 26 17:06:24 crc kubenswrapper[4754]: I0126 17:06:24.577695 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sfd4m\" (UniqueName: \"kubernetes.io/projected/19a0730c-46d0-4029-a86b-812ba1664dcc-kube-api-access-sfd4m\") pod \"swift-storage-0\" (UID: \"19a0730c-46d0-4029-a86b-812ba1664dcc\") " pod="openstack/swift-storage-0" Jan 26 17:06:24 crc kubenswrapper[4754]: I0126 17:06:24.577728 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/19a0730c-46d0-4029-a86b-812ba1664dcc-etc-swift\") pod \"swift-storage-0\" (UID: \"19a0730c-46d0-4029-a86b-812ba1664dcc\") " pod="openstack/swift-storage-0" Jan 26 17:06:24 crc kubenswrapper[4754]: I0126 17:06:24.577747 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/19a0730c-46d0-4029-a86b-812ba1664dcc-cache\") pod \"swift-storage-0\" (UID: \"19a0730c-46d0-4029-a86b-812ba1664dcc\") " pod="openstack/swift-storage-0" Jan 26 17:06:24 crc kubenswrapper[4754]: I0126 17:06:24.578822 4754 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"swift-storage-0\" (UID: \"19a0730c-46d0-4029-a86b-812ba1664dcc\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/swift-storage-0" Jan 26 17:06:24 crc kubenswrapper[4754]: I0126 17:06:24.579110 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/19a0730c-46d0-4029-a86b-812ba1664dcc-lock\") pod \"swift-storage-0\" (UID: \"19a0730c-46d0-4029-a86b-812ba1664dcc\") " pod="openstack/swift-storage-0" Jan 26 17:06:24 crc kubenswrapper[4754]: E0126 17:06:24.579308 4754 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Jan 26 17:06:24 crc kubenswrapper[4754]: E0126 17:06:24.579384 4754 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Jan 26 17:06:24 crc kubenswrapper[4754]: E0126 17:06:24.579485 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/19a0730c-46d0-4029-a86b-812ba1664dcc-etc-swift podName:19a0730c-46d0-4029-a86b-812ba1664dcc nodeName:}" failed. No retries permitted until 2026-01-26 17:06:25.079463451 +0000 UTC m=+1151.603643885 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/19a0730c-46d0-4029-a86b-812ba1664dcc-etc-swift") pod "swift-storage-0" (UID: "19a0730c-46d0-4029-a86b-812ba1664dcc") : configmap "swift-ring-files" not found Jan 26 17:06:24 crc kubenswrapper[4754]: I0126 17:06:24.580019 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/19a0730c-46d0-4029-a86b-812ba1664dcc-cache\") pod \"swift-storage-0\" (UID: \"19a0730c-46d0-4029-a86b-812ba1664dcc\") " pod="openstack/swift-storage-0" Jan 26 17:06:24 crc kubenswrapper[4754]: I0126 17:06:24.583828 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19a0730c-46d0-4029-a86b-812ba1664dcc-combined-ca-bundle\") pod \"swift-storage-0\" (UID: \"19a0730c-46d0-4029-a86b-812ba1664dcc\") " pod="openstack/swift-storage-0" Jan 26 17:06:24 crc kubenswrapper[4754]: I0126 17:06:24.598388 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sfd4m\" (UniqueName: \"kubernetes.io/projected/19a0730c-46d0-4029-a86b-812ba1664dcc-kube-api-access-sfd4m\") pod \"swift-storage-0\" (UID: \"19a0730c-46d0-4029-a86b-812ba1664dcc\") " pod="openstack/swift-storage-0" Jan 26 17:06:24 crc kubenswrapper[4754]: I0126 17:06:24.600286 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"swift-storage-0\" (UID: \"19a0730c-46d0-4029-a86b-812ba1664dcc\") " pod="openstack/swift-storage-0" Jan 26 17:06:24 crc kubenswrapper[4754]: I0126 17:06:24.735163 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-sl9kb"] Jan 26 17:06:24 crc kubenswrapper[4754]: I0126 17:06:24.736204 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-sl9kb" Jan 26 17:06:24 crc kubenswrapper[4754]: I0126 17:06:24.743108 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Jan 26 17:06:24 crc kubenswrapper[4754]: I0126 17:06:24.743608 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Jan 26 17:06:24 crc kubenswrapper[4754]: I0126 17:06:24.745439 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Jan 26 17:06:24 crc kubenswrapper[4754]: I0126 17:06:24.745782 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-sl9kb"] Jan 26 17:06:24 crc kubenswrapper[4754]: I0126 17:06:24.883020 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mqx8z\" (UniqueName: \"kubernetes.io/projected/cb88274a-904f-4827-9518-81b79a0e6a42-kube-api-access-mqx8z\") pod \"swift-ring-rebalance-sl9kb\" (UID: \"cb88274a-904f-4827-9518-81b79a0e6a42\") " pod="openstack/swift-ring-rebalance-sl9kb" Jan 26 17:06:24 crc kubenswrapper[4754]: I0126 17:06:24.883128 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cb88274a-904f-4827-9518-81b79a0e6a42-scripts\") pod \"swift-ring-rebalance-sl9kb\" (UID: \"cb88274a-904f-4827-9518-81b79a0e6a42\") " pod="openstack/swift-ring-rebalance-sl9kb" Jan 26 17:06:24 crc kubenswrapper[4754]: I0126 17:06:24.883175 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/cb88274a-904f-4827-9518-81b79a0e6a42-swiftconf\") pod \"swift-ring-rebalance-sl9kb\" (UID: \"cb88274a-904f-4827-9518-81b79a0e6a42\") " pod="openstack/swift-ring-rebalance-sl9kb" Jan 26 17:06:24 crc kubenswrapper[4754]: I0126 17:06:24.883305 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/cb88274a-904f-4827-9518-81b79a0e6a42-etc-swift\") pod \"swift-ring-rebalance-sl9kb\" (UID: \"cb88274a-904f-4827-9518-81b79a0e6a42\") " pod="openstack/swift-ring-rebalance-sl9kb" Jan 26 17:06:24 crc kubenswrapper[4754]: I0126 17:06:24.883356 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/cb88274a-904f-4827-9518-81b79a0e6a42-dispersionconf\") pod \"swift-ring-rebalance-sl9kb\" (UID: \"cb88274a-904f-4827-9518-81b79a0e6a42\") " pod="openstack/swift-ring-rebalance-sl9kb" Jan 26 17:06:24 crc kubenswrapper[4754]: I0126 17:06:24.883565 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/cb88274a-904f-4827-9518-81b79a0e6a42-ring-data-devices\") pod \"swift-ring-rebalance-sl9kb\" (UID: \"cb88274a-904f-4827-9518-81b79a0e6a42\") " pod="openstack/swift-ring-rebalance-sl9kb" Jan 26 17:06:24 crc kubenswrapper[4754]: I0126 17:06:24.883805 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb88274a-904f-4827-9518-81b79a0e6a42-combined-ca-bundle\") pod \"swift-ring-rebalance-sl9kb\" (UID: \"cb88274a-904f-4827-9518-81b79a0e6a42\") " pod="openstack/swift-ring-rebalance-sl9kb" Jan 26 17:06:24 crc kubenswrapper[4754]: I0126 17:06:24.985542 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb88274a-904f-4827-9518-81b79a0e6a42-combined-ca-bundle\") pod \"swift-ring-rebalance-sl9kb\" (UID: \"cb88274a-904f-4827-9518-81b79a0e6a42\") " pod="openstack/swift-ring-rebalance-sl9kb" Jan 26 17:06:24 crc kubenswrapper[4754]: I0126 17:06:24.985620 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mqx8z\" (UniqueName: \"kubernetes.io/projected/cb88274a-904f-4827-9518-81b79a0e6a42-kube-api-access-mqx8z\") pod \"swift-ring-rebalance-sl9kb\" (UID: \"cb88274a-904f-4827-9518-81b79a0e6a42\") " pod="openstack/swift-ring-rebalance-sl9kb" Jan 26 17:06:24 crc kubenswrapper[4754]: I0126 17:06:24.985688 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cb88274a-904f-4827-9518-81b79a0e6a42-scripts\") pod \"swift-ring-rebalance-sl9kb\" (UID: \"cb88274a-904f-4827-9518-81b79a0e6a42\") " pod="openstack/swift-ring-rebalance-sl9kb" Jan 26 17:06:24 crc kubenswrapper[4754]: I0126 17:06:24.985721 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/cb88274a-904f-4827-9518-81b79a0e6a42-swiftconf\") pod \"swift-ring-rebalance-sl9kb\" (UID: \"cb88274a-904f-4827-9518-81b79a0e6a42\") " pod="openstack/swift-ring-rebalance-sl9kb" Jan 26 17:06:24 crc kubenswrapper[4754]: I0126 17:06:24.985771 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/cb88274a-904f-4827-9518-81b79a0e6a42-etc-swift\") pod \"swift-ring-rebalance-sl9kb\" (UID: \"cb88274a-904f-4827-9518-81b79a0e6a42\") " pod="openstack/swift-ring-rebalance-sl9kb" Jan 26 17:06:24 crc kubenswrapper[4754]: I0126 17:06:24.985787 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/cb88274a-904f-4827-9518-81b79a0e6a42-dispersionconf\") pod \"swift-ring-rebalance-sl9kb\" (UID: \"cb88274a-904f-4827-9518-81b79a0e6a42\") " pod="openstack/swift-ring-rebalance-sl9kb" Jan 26 17:06:24 crc kubenswrapper[4754]: I0126 17:06:24.985823 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/cb88274a-904f-4827-9518-81b79a0e6a42-ring-data-devices\") pod \"swift-ring-rebalance-sl9kb\" (UID: \"cb88274a-904f-4827-9518-81b79a0e6a42\") " pod="openstack/swift-ring-rebalance-sl9kb" Jan 26 17:06:24 crc kubenswrapper[4754]: I0126 17:06:24.986211 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/cb88274a-904f-4827-9518-81b79a0e6a42-etc-swift\") pod \"swift-ring-rebalance-sl9kb\" (UID: \"cb88274a-904f-4827-9518-81b79a0e6a42\") " pod="openstack/swift-ring-rebalance-sl9kb" Jan 26 17:06:24 crc kubenswrapper[4754]: I0126 17:06:24.986546 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/cb88274a-904f-4827-9518-81b79a0e6a42-ring-data-devices\") pod \"swift-ring-rebalance-sl9kb\" (UID: \"cb88274a-904f-4827-9518-81b79a0e6a42\") " pod="openstack/swift-ring-rebalance-sl9kb" Jan 26 17:06:24 crc kubenswrapper[4754]: I0126 17:06:24.986704 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cb88274a-904f-4827-9518-81b79a0e6a42-scripts\") pod \"swift-ring-rebalance-sl9kb\" (UID: \"cb88274a-904f-4827-9518-81b79a0e6a42\") " pod="openstack/swift-ring-rebalance-sl9kb" Jan 26 17:06:24 crc kubenswrapper[4754]: I0126 17:06:24.990486 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/cb88274a-904f-4827-9518-81b79a0e6a42-swiftconf\") pod \"swift-ring-rebalance-sl9kb\" (UID: \"cb88274a-904f-4827-9518-81b79a0e6a42\") " pod="openstack/swift-ring-rebalance-sl9kb" Jan 26 17:06:24 crc kubenswrapper[4754]: I0126 17:06:24.990930 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb88274a-904f-4827-9518-81b79a0e6a42-combined-ca-bundle\") pod \"swift-ring-rebalance-sl9kb\" (UID: \"cb88274a-904f-4827-9518-81b79a0e6a42\") " pod="openstack/swift-ring-rebalance-sl9kb" Jan 26 17:06:24 crc kubenswrapper[4754]: I0126 17:06:24.991999 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/cb88274a-904f-4827-9518-81b79a0e6a42-dispersionconf\") pod \"swift-ring-rebalance-sl9kb\" (UID: \"cb88274a-904f-4827-9518-81b79a0e6a42\") " pod="openstack/swift-ring-rebalance-sl9kb" Jan 26 17:06:25 crc kubenswrapper[4754]: I0126 17:06:25.001372 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mqx8z\" (UniqueName: \"kubernetes.io/projected/cb88274a-904f-4827-9518-81b79a0e6a42-kube-api-access-mqx8z\") pod \"swift-ring-rebalance-sl9kb\" (UID: \"cb88274a-904f-4827-9518-81b79a0e6a42\") " pod="openstack/swift-ring-rebalance-sl9kb" Jan 26 17:06:25 crc kubenswrapper[4754]: I0126 17:06:25.057333 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-sl9kb" Jan 26 17:06:25 crc kubenswrapper[4754]: I0126 17:06:25.087544 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/19a0730c-46d0-4029-a86b-812ba1664dcc-etc-swift\") pod \"swift-storage-0\" (UID: \"19a0730c-46d0-4029-a86b-812ba1664dcc\") " pod="openstack/swift-storage-0" Jan 26 17:06:25 crc kubenswrapper[4754]: E0126 17:06:25.087801 4754 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Jan 26 17:06:25 crc kubenswrapper[4754]: E0126 17:06:25.087823 4754 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Jan 26 17:06:25 crc kubenswrapper[4754]: E0126 17:06:25.087872 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/19a0730c-46d0-4029-a86b-812ba1664dcc-etc-swift podName:19a0730c-46d0-4029-a86b-812ba1664dcc nodeName:}" failed. No retries permitted until 2026-01-26 17:06:26.087855824 +0000 UTC m=+1152.612036258 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/19a0730c-46d0-4029-a86b-812ba1664dcc-etc-swift") pod "swift-storage-0" (UID: "19a0730c-46d0-4029-a86b-812ba1664dcc") : configmap "swift-ring-files" not found Jan 26 17:06:25 crc kubenswrapper[4754]: I0126 17:06:25.163772 4754 generic.go:334] "Generic (PLEG): container finished" podID="b5317650-1469-4237-8645-6031475c5b37" containerID="eb843a44f878f648e2b6b755b6528ad1dcc1f5f1fd6ead5670f2b8320ba9d8e4" exitCode=0 Jan 26 17:06:25 crc kubenswrapper[4754]: I0126 17:06:25.164198 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-z5pcq" event={"ID":"b5317650-1469-4237-8645-6031475c5b37","Type":"ContainerDied","Data":"eb843a44f878f648e2b6b755b6528ad1dcc1f5f1fd6ead5670f2b8320ba9d8e4"} Jan 26 17:06:25 crc kubenswrapper[4754]: I0126 17:06:25.164255 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-z5pcq" event={"ID":"b5317650-1469-4237-8645-6031475c5b37","Type":"ContainerStarted","Data":"55eb578a90c8ba0834c5d56e7fdc32dd86c9ce14da617f0b35505ec34e6168cb"} Jan 26 17:06:25 crc kubenswrapper[4754]: I0126 17:06:25.687517 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-c7cxf" Jan 26 17:06:25 crc kubenswrapper[4754]: I0126 17:06:25.694909 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-8dnlq" Jan 26 17:06:25 crc kubenswrapper[4754]: I0126 17:06:25.697896 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-4c00-account-create-update-h9g8d" Jan 26 17:06:25 crc kubenswrapper[4754]: I0126 17:06:25.793519 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0bf413d0-f8cf-4574-b4bc-dfa3264a41b9" path="/var/lib/kubelet/pods/0bf413d0-f8cf-4574-b4bc-dfa3264a41b9/volumes" Jan 26 17:06:25 crc kubenswrapper[4754]: I0126 17:06:25.804982 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fe272255-848d-4761-a7f9-347103832e7d-operator-scripts\") pod \"fe272255-848d-4761-a7f9-347103832e7d\" (UID: \"fe272255-848d-4761-a7f9-347103832e7d\") " Jan 26 17:06:25 crc kubenswrapper[4754]: I0126 17:06:25.805034 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dce62524-1013-4b38-b563-d9f6d7701523-operator-scripts\") pod \"dce62524-1013-4b38-b563-d9f6d7701523\" (UID: \"dce62524-1013-4b38-b563-d9f6d7701523\") " Jan 26 17:06:25 crc kubenswrapper[4754]: I0126 17:06:25.805055 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9295f950-6f88-42d3-a82e-688cda9cce76-operator-scripts\") pod \"9295f950-6f88-42d3-a82e-688cda9cce76\" (UID: \"9295f950-6f88-42d3-a82e-688cda9cce76\") " Jan 26 17:06:25 crc kubenswrapper[4754]: I0126 17:06:25.805255 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4prbm\" (UniqueName: \"kubernetes.io/projected/9295f950-6f88-42d3-a82e-688cda9cce76-kube-api-access-4prbm\") pod \"9295f950-6f88-42d3-a82e-688cda9cce76\" (UID: \"9295f950-6f88-42d3-a82e-688cda9cce76\") " Jan 26 17:06:25 crc kubenswrapper[4754]: I0126 17:06:25.805283 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4c6r\" (UniqueName: \"kubernetes.io/projected/fe272255-848d-4761-a7f9-347103832e7d-kube-api-access-x4c6r\") pod \"fe272255-848d-4761-a7f9-347103832e7d\" (UID: \"fe272255-848d-4761-a7f9-347103832e7d\") " Jan 26 17:06:25 crc kubenswrapper[4754]: I0126 17:06:25.805487 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z5vl4\" (UniqueName: \"kubernetes.io/projected/dce62524-1013-4b38-b563-d9f6d7701523-kube-api-access-z5vl4\") pod \"dce62524-1013-4b38-b563-d9f6d7701523\" (UID: \"dce62524-1013-4b38-b563-d9f6d7701523\") " Jan 26 17:06:25 crc kubenswrapper[4754]: I0126 17:06:25.805937 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fe272255-848d-4761-a7f9-347103832e7d-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "fe272255-848d-4761-a7f9-347103832e7d" (UID: "fe272255-848d-4761-a7f9-347103832e7d"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:06:25 crc kubenswrapper[4754]: I0126 17:06:25.805941 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9295f950-6f88-42d3-a82e-688cda9cce76-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "9295f950-6f88-42d3-a82e-688cda9cce76" (UID: "9295f950-6f88-42d3-a82e-688cda9cce76"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:06:25 crc kubenswrapper[4754]: I0126 17:06:25.806029 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dce62524-1013-4b38-b563-d9f6d7701523-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "dce62524-1013-4b38-b563-d9f6d7701523" (UID: "dce62524-1013-4b38-b563-d9f6d7701523"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:06:25 crc kubenswrapper[4754]: I0126 17:06:25.806576 4754 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fe272255-848d-4761-a7f9-347103832e7d-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 26 17:06:25 crc kubenswrapper[4754]: I0126 17:06:25.806605 4754 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dce62524-1013-4b38-b563-d9f6d7701523-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 26 17:06:25 crc kubenswrapper[4754]: I0126 17:06:25.806618 4754 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9295f950-6f88-42d3-a82e-688cda9cce76-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 26 17:06:25 crc kubenswrapper[4754]: I0126 17:06:25.812929 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9295f950-6f88-42d3-a82e-688cda9cce76-kube-api-access-4prbm" (OuterVolumeSpecName: "kube-api-access-4prbm") pod "9295f950-6f88-42d3-a82e-688cda9cce76" (UID: "9295f950-6f88-42d3-a82e-688cda9cce76"). InnerVolumeSpecName "kube-api-access-4prbm". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:06:25 crc kubenswrapper[4754]: I0126 17:06:25.813592 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dce62524-1013-4b38-b563-d9f6d7701523-kube-api-access-z5vl4" (OuterVolumeSpecName: "kube-api-access-z5vl4") pod "dce62524-1013-4b38-b563-d9f6d7701523" (UID: "dce62524-1013-4b38-b563-d9f6d7701523"). InnerVolumeSpecName "kube-api-access-z5vl4". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:06:25 crc kubenswrapper[4754]: I0126 17:06:25.815624 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fe272255-848d-4761-a7f9-347103832e7d-kube-api-access-x4c6r" (OuterVolumeSpecName: "kube-api-access-x4c6r") pod "fe272255-848d-4761-a7f9-347103832e7d" (UID: "fe272255-848d-4761-a7f9-347103832e7d"). InnerVolumeSpecName "kube-api-access-x4c6r". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:06:25 crc kubenswrapper[4754]: I0126 17:06:25.818018 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-sl9kb"] Jan 26 17:06:25 crc kubenswrapper[4754]: I0126 17:06:25.824810 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-f12b-account-create-update-scv4g" Jan 26 17:06:25 crc kubenswrapper[4754]: I0126 17:06:25.907630 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d1acc0d8-ac6a-4b56-bc9b-d926176314c3-operator-scripts\") pod \"d1acc0d8-ac6a-4b56-bc9b-d926176314c3\" (UID: \"d1acc0d8-ac6a-4b56-bc9b-d926176314c3\") " Jan 26 17:06:25 crc kubenswrapper[4754]: I0126 17:06:25.907718 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bw69j\" (UniqueName: \"kubernetes.io/projected/d1acc0d8-ac6a-4b56-bc9b-d926176314c3-kube-api-access-bw69j\") pod \"d1acc0d8-ac6a-4b56-bc9b-d926176314c3\" (UID: \"d1acc0d8-ac6a-4b56-bc9b-d926176314c3\") " Jan 26 17:06:25 crc kubenswrapper[4754]: I0126 17:06:25.908190 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z5vl4\" (UniqueName: \"kubernetes.io/projected/dce62524-1013-4b38-b563-d9f6d7701523-kube-api-access-z5vl4\") on node \"crc\" DevicePath \"\"" Jan 26 17:06:25 crc kubenswrapper[4754]: I0126 17:06:25.908215 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4prbm\" (UniqueName: \"kubernetes.io/projected/9295f950-6f88-42d3-a82e-688cda9cce76-kube-api-access-4prbm\") on node \"crc\" DevicePath \"\"" Jan 26 17:06:25 crc kubenswrapper[4754]: I0126 17:06:25.908228 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4c6r\" (UniqueName: \"kubernetes.io/projected/fe272255-848d-4761-a7f9-347103832e7d-kube-api-access-x4c6r\") on node \"crc\" DevicePath \"\"" Jan 26 17:06:25 crc kubenswrapper[4754]: I0126 17:06:25.908319 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d1acc0d8-ac6a-4b56-bc9b-d926176314c3-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "d1acc0d8-ac6a-4b56-bc9b-d926176314c3" (UID: "d1acc0d8-ac6a-4b56-bc9b-d926176314c3"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:06:25 crc kubenswrapper[4754]: I0126 17:06:25.911333 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-86db49b7ff-r9jrq" Jan 26 17:06:25 crc kubenswrapper[4754]: I0126 17:06:25.912367 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d1acc0d8-ac6a-4b56-bc9b-d926176314c3-kube-api-access-bw69j" (OuterVolumeSpecName: "kube-api-access-bw69j") pod "d1acc0d8-ac6a-4b56-bc9b-d926176314c3" (UID: "d1acc0d8-ac6a-4b56-bc9b-d926176314c3"). InnerVolumeSpecName "kube-api-access-bw69j". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:06:26 crc kubenswrapper[4754]: I0126 17:06:26.009791 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bw69j\" (UniqueName: \"kubernetes.io/projected/d1acc0d8-ac6a-4b56-bc9b-d926176314c3-kube-api-access-bw69j\") on node \"crc\" DevicePath \"\"" Jan 26 17:06:26 crc kubenswrapper[4754]: I0126 17:06:26.009827 4754 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d1acc0d8-ac6a-4b56-bc9b-d926176314c3-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 26 17:06:26 crc kubenswrapper[4754]: I0126 17:06:26.053183 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-plgbz"] Jan 26 17:06:26 crc kubenswrapper[4754]: E0126 17:06:26.053585 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dce62524-1013-4b38-b563-d9f6d7701523" containerName="mariadb-account-create-update" Jan 26 17:06:26 crc kubenswrapper[4754]: I0126 17:06:26.053606 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="dce62524-1013-4b38-b563-d9f6d7701523" containerName="mariadb-account-create-update" Jan 26 17:06:26 crc kubenswrapper[4754]: E0126 17:06:26.053633 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fe272255-848d-4761-a7f9-347103832e7d" containerName="mariadb-database-create" Jan 26 17:06:26 crc kubenswrapper[4754]: I0126 17:06:26.053642 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="fe272255-848d-4761-a7f9-347103832e7d" containerName="mariadb-database-create" Jan 26 17:06:26 crc kubenswrapper[4754]: E0126 17:06:26.053689 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9295f950-6f88-42d3-a82e-688cda9cce76" containerName="mariadb-database-create" Jan 26 17:06:26 crc kubenswrapper[4754]: I0126 17:06:26.053699 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="9295f950-6f88-42d3-a82e-688cda9cce76" containerName="mariadb-database-create" Jan 26 17:06:26 crc kubenswrapper[4754]: E0126 17:06:26.053714 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1acc0d8-ac6a-4b56-bc9b-d926176314c3" containerName="mariadb-account-create-update" Jan 26 17:06:26 crc kubenswrapper[4754]: I0126 17:06:26.053721 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1acc0d8-ac6a-4b56-bc9b-d926176314c3" containerName="mariadb-account-create-update" Jan 26 17:06:26 crc kubenswrapper[4754]: I0126 17:06:26.053909 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="fe272255-848d-4761-a7f9-347103832e7d" containerName="mariadb-database-create" Jan 26 17:06:26 crc kubenswrapper[4754]: I0126 17:06:26.053930 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="dce62524-1013-4b38-b563-d9f6d7701523" containerName="mariadb-account-create-update" Jan 26 17:06:26 crc kubenswrapper[4754]: I0126 17:06:26.053941 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="d1acc0d8-ac6a-4b56-bc9b-d926176314c3" containerName="mariadb-account-create-update" Jan 26 17:06:26 crc kubenswrapper[4754]: I0126 17:06:26.053955 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="9295f950-6f88-42d3-a82e-688cda9cce76" containerName="mariadb-database-create" Jan 26 17:06:26 crc kubenswrapper[4754]: I0126 17:06:26.054591 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-plgbz" Jan 26 17:06:26 crc kubenswrapper[4754]: I0126 17:06:26.063335 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-plgbz"] Jan 26 17:06:26 crc kubenswrapper[4754]: I0126 17:06:26.111030 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/19a0730c-46d0-4029-a86b-812ba1664dcc-etc-swift\") pod \"swift-storage-0\" (UID: \"19a0730c-46d0-4029-a86b-812ba1664dcc\") " pod="openstack/swift-storage-0" Jan 26 17:06:26 crc kubenswrapper[4754]: E0126 17:06:26.111253 4754 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Jan 26 17:06:26 crc kubenswrapper[4754]: E0126 17:06:26.111285 4754 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Jan 26 17:06:26 crc kubenswrapper[4754]: E0126 17:06:26.111348 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/19a0730c-46d0-4029-a86b-812ba1664dcc-etc-swift podName:19a0730c-46d0-4029-a86b-812ba1664dcc nodeName:}" failed. No retries permitted until 2026-01-26 17:06:28.111328979 +0000 UTC m=+1154.635509413 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/19a0730c-46d0-4029-a86b-812ba1664dcc-etc-swift") pod "swift-storage-0" (UID: "19a0730c-46d0-4029-a86b-812ba1664dcc") : configmap "swift-ring-files" not found Jan 26 17:06:26 crc kubenswrapper[4754]: I0126 17:06:26.145543 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-fc80-account-create-update-jh5xv"] Jan 26 17:06:26 crc kubenswrapper[4754]: I0126 17:06:26.146945 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-fc80-account-create-update-jh5xv" Jan 26 17:06:26 crc kubenswrapper[4754]: I0126 17:06:26.148941 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Jan 26 17:06:26 crc kubenswrapper[4754]: I0126 17:06:26.156445 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-fc80-account-create-update-jh5xv"] Jan 26 17:06:26 crc kubenswrapper[4754]: I0126 17:06:26.174341 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-8dnlq" event={"ID":"9295f950-6f88-42d3-a82e-688cda9cce76","Type":"ContainerDied","Data":"16604df2d74058ba3cd3c3c5e2450b605595cf261ff70a8ebc5b9b6ff15635d3"} Jan 26 17:06:26 crc kubenswrapper[4754]: I0126 17:06:26.174397 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="16604df2d74058ba3cd3c3c5e2450b605595cf261ff70a8ebc5b9b6ff15635d3" Jan 26 17:06:26 crc kubenswrapper[4754]: I0126 17:06:26.174362 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-8dnlq" Jan 26 17:06:26 crc kubenswrapper[4754]: I0126 17:06:26.175641 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-sl9kb" event={"ID":"cb88274a-904f-4827-9518-81b79a0e6a42","Type":"ContainerStarted","Data":"7c406fb5a434d2616d1651f2bf56181f757b5fc67b31825b896ace5a92d83a88"} Jan 26 17:06:26 crc kubenswrapper[4754]: I0126 17:06:26.177408 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-f12b-account-create-update-scv4g" event={"ID":"d1acc0d8-ac6a-4b56-bc9b-d926176314c3","Type":"ContainerDied","Data":"04468a86c787080e75988037b5a37583ed4cf476f3870d033abb5e4371717255"} Jan 26 17:06:26 crc kubenswrapper[4754]: I0126 17:06:26.177433 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-f12b-account-create-update-scv4g" Jan 26 17:06:26 crc kubenswrapper[4754]: I0126 17:06:26.177455 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="04468a86c787080e75988037b5a37583ed4cf476f3870d033abb5e4371717255" Jan 26 17:06:26 crc kubenswrapper[4754]: I0126 17:06:26.179190 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-4c00-account-create-update-h9g8d" event={"ID":"dce62524-1013-4b38-b563-d9f6d7701523","Type":"ContainerDied","Data":"c152c67eaf2f36d600bf5e9f7a9a2697bce7f6ef1f957453a2ba47daff11d86f"} Jan 26 17:06:26 crc kubenswrapper[4754]: I0126 17:06:26.179222 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c152c67eaf2f36d600bf5e9f7a9a2697bce7f6ef1f957453a2ba47daff11d86f" Jan 26 17:06:26 crc kubenswrapper[4754]: I0126 17:06:26.179256 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-4c00-account-create-update-h9g8d" Jan 26 17:06:26 crc kubenswrapper[4754]: I0126 17:06:26.182941 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-z5pcq" event={"ID":"b5317650-1469-4237-8645-6031475c5b37","Type":"ContainerStarted","Data":"195a1f054f708b9ea9b0b61078c6a54456031c18ec70286ce347d62b42231752"} Jan 26 17:06:26 crc kubenswrapper[4754]: I0126 17:06:26.183203 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-698758b865-z5pcq" Jan 26 17:06:26 crc kubenswrapper[4754]: I0126 17:06:26.184766 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-c7cxf" Jan 26 17:06:26 crc kubenswrapper[4754]: I0126 17:06:26.184782 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-c7cxf" event={"ID":"fe272255-848d-4761-a7f9-347103832e7d","Type":"ContainerDied","Data":"e278e71b5578190df55aa6d3359fae2229bc1a7a1e4c18abe9f86260fcac7d70"} Jan 26 17:06:26 crc kubenswrapper[4754]: I0126 17:06:26.184812 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e278e71b5578190df55aa6d3359fae2229bc1a7a1e4c18abe9f86260fcac7d70" Jan 26 17:06:26 crc kubenswrapper[4754]: I0126 17:06:26.206120 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-698758b865-z5pcq" podStartSLOduration=3.206101767 podStartE2EDuration="3.206101767s" podCreationTimestamp="2026-01-26 17:06:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 17:06:26.205571302 +0000 UTC m=+1152.729751746" watchObservedRunningTime="2026-01-26 17:06:26.206101767 +0000 UTC m=+1152.730282201" Jan 26 17:06:26 crc kubenswrapper[4754]: I0126 17:06:26.212347 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7cbcg\" (UniqueName: \"kubernetes.io/projected/07ac2529-0a5d-4a79-95cf-2b5040e67f46-kube-api-access-7cbcg\") pod \"glance-db-create-plgbz\" (UID: \"07ac2529-0a5d-4a79-95cf-2b5040e67f46\") " pod="openstack/glance-db-create-plgbz" Jan 26 17:06:26 crc kubenswrapper[4754]: I0126 17:06:26.212661 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j8zwl\" (UniqueName: \"kubernetes.io/projected/e2c6020a-c674-43b0-9698-b6b31e0ab5b8-kube-api-access-j8zwl\") pod \"glance-fc80-account-create-update-jh5xv\" (UID: \"e2c6020a-c674-43b0-9698-b6b31e0ab5b8\") " pod="openstack/glance-fc80-account-create-update-jh5xv" Jan 26 17:06:26 crc kubenswrapper[4754]: I0126 17:06:26.212783 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e2c6020a-c674-43b0-9698-b6b31e0ab5b8-operator-scripts\") pod \"glance-fc80-account-create-update-jh5xv\" (UID: \"e2c6020a-c674-43b0-9698-b6b31e0ab5b8\") " pod="openstack/glance-fc80-account-create-update-jh5xv" Jan 26 17:06:26 crc kubenswrapper[4754]: I0126 17:06:26.212898 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/07ac2529-0a5d-4a79-95cf-2b5040e67f46-operator-scripts\") pod \"glance-db-create-plgbz\" (UID: \"07ac2529-0a5d-4a79-95cf-2b5040e67f46\") " pod="openstack/glance-db-create-plgbz" Jan 26 17:06:26 crc kubenswrapper[4754]: I0126 17:06:26.314876 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e2c6020a-c674-43b0-9698-b6b31e0ab5b8-operator-scripts\") pod \"glance-fc80-account-create-update-jh5xv\" (UID: \"e2c6020a-c674-43b0-9698-b6b31e0ab5b8\") " pod="openstack/glance-fc80-account-create-update-jh5xv" Jan 26 17:06:26 crc kubenswrapper[4754]: I0126 17:06:26.314918 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/07ac2529-0a5d-4a79-95cf-2b5040e67f46-operator-scripts\") pod \"glance-db-create-plgbz\" (UID: \"07ac2529-0a5d-4a79-95cf-2b5040e67f46\") " pod="openstack/glance-db-create-plgbz" Jan 26 17:06:26 crc kubenswrapper[4754]: I0126 17:06:26.315039 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7cbcg\" (UniqueName: \"kubernetes.io/projected/07ac2529-0a5d-4a79-95cf-2b5040e67f46-kube-api-access-7cbcg\") pod \"glance-db-create-plgbz\" (UID: \"07ac2529-0a5d-4a79-95cf-2b5040e67f46\") " pod="openstack/glance-db-create-plgbz" Jan 26 17:06:26 crc kubenswrapper[4754]: I0126 17:06:26.315542 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j8zwl\" (UniqueName: \"kubernetes.io/projected/e2c6020a-c674-43b0-9698-b6b31e0ab5b8-kube-api-access-j8zwl\") pod \"glance-fc80-account-create-update-jh5xv\" (UID: \"e2c6020a-c674-43b0-9698-b6b31e0ab5b8\") " pod="openstack/glance-fc80-account-create-update-jh5xv" Jan 26 17:06:26 crc kubenswrapper[4754]: I0126 17:06:26.315903 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e2c6020a-c674-43b0-9698-b6b31e0ab5b8-operator-scripts\") pod \"glance-fc80-account-create-update-jh5xv\" (UID: \"e2c6020a-c674-43b0-9698-b6b31e0ab5b8\") " pod="openstack/glance-fc80-account-create-update-jh5xv" Jan 26 17:06:26 crc kubenswrapper[4754]: I0126 17:06:26.316327 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/07ac2529-0a5d-4a79-95cf-2b5040e67f46-operator-scripts\") pod \"glance-db-create-plgbz\" (UID: \"07ac2529-0a5d-4a79-95cf-2b5040e67f46\") " pod="openstack/glance-db-create-plgbz" Jan 26 17:06:26 crc kubenswrapper[4754]: I0126 17:06:26.336966 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7cbcg\" (UniqueName: \"kubernetes.io/projected/07ac2529-0a5d-4a79-95cf-2b5040e67f46-kube-api-access-7cbcg\") pod \"glance-db-create-plgbz\" (UID: \"07ac2529-0a5d-4a79-95cf-2b5040e67f46\") " pod="openstack/glance-db-create-plgbz" Jan 26 17:06:26 crc kubenswrapper[4754]: I0126 17:06:26.337387 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j8zwl\" (UniqueName: \"kubernetes.io/projected/e2c6020a-c674-43b0-9698-b6b31e0ab5b8-kube-api-access-j8zwl\") pod \"glance-fc80-account-create-update-jh5xv\" (UID: \"e2c6020a-c674-43b0-9698-b6b31e0ab5b8\") " pod="openstack/glance-fc80-account-create-update-jh5xv" Jan 26 17:06:26 crc kubenswrapper[4754]: I0126 17:06:26.370315 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-plgbz" Jan 26 17:06:26 crc kubenswrapper[4754]: I0126 17:06:26.462279 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-fc80-account-create-update-jh5xv" Jan 26 17:06:26 crc kubenswrapper[4754]: I0126 17:06:26.841169 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-plgbz"] Jan 26 17:06:26 crc kubenswrapper[4754]: W0126 17:06:26.849520 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod07ac2529_0a5d_4a79_95cf_2b5040e67f46.slice/crio-37a1abe949c9a09b07d9183243f5bad5bf49e5ca628bc0e1a39ab9bb25710a30 WatchSource:0}: Error finding container 37a1abe949c9a09b07d9183243f5bad5bf49e5ca628bc0e1a39ab9bb25710a30: Status 404 returned error can't find the container with id 37a1abe949c9a09b07d9183243f5bad5bf49e5ca628bc0e1a39ab9bb25710a30 Jan 26 17:06:26 crc kubenswrapper[4754]: I0126 17:06:26.958529 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-fc80-account-create-update-jh5xv"] Jan 26 17:06:26 crc kubenswrapper[4754]: W0126 17:06:26.961492 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode2c6020a_c674_43b0_9698_b6b31e0ab5b8.slice/crio-84db69390c57afbb8f02d1cf32b8d53df3043329a6e3380ae0e701825296a9b7 WatchSource:0}: Error finding container 84db69390c57afbb8f02d1cf32b8d53df3043329a6e3380ae0e701825296a9b7: Status 404 returned error can't find the container with id 84db69390c57afbb8f02d1cf32b8d53df3043329a6e3380ae0e701825296a9b7 Jan 26 17:06:27 crc kubenswrapper[4754]: I0126 17:06:27.194536 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-fc80-account-create-update-jh5xv" event={"ID":"e2c6020a-c674-43b0-9698-b6b31e0ab5b8","Type":"ContainerStarted","Data":"e6362fc04381ff9a7db06b0751f7ace044043ec05a105674da46640b67248fc0"} Jan 26 17:06:27 crc kubenswrapper[4754]: I0126 17:06:27.194584 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-fc80-account-create-update-jh5xv" event={"ID":"e2c6020a-c674-43b0-9698-b6b31e0ab5b8","Type":"ContainerStarted","Data":"84db69390c57afbb8f02d1cf32b8d53df3043329a6e3380ae0e701825296a9b7"} Jan 26 17:06:27 crc kubenswrapper[4754]: I0126 17:06:27.195949 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-plgbz" event={"ID":"07ac2529-0a5d-4a79-95cf-2b5040e67f46","Type":"ContainerStarted","Data":"e84f43eb249c92fd72b6e9bc45e2e579ad9889dd0ac77ec6d06f9ec1235954ec"} Jan 26 17:06:27 crc kubenswrapper[4754]: I0126 17:06:27.195990 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-plgbz" event={"ID":"07ac2529-0a5d-4a79-95cf-2b5040e67f46","Type":"ContainerStarted","Data":"37a1abe949c9a09b07d9183243f5bad5bf49e5ca628bc0e1a39ab9bb25710a30"} Jan 26 17:06:27 crc kubenswrapper[4754]: I0126 17:06:27.234944 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-create-plgbz" podStartSLOduration=1.2349239810000001 podStartE2EDuration="1.234923981s" podCreationTimestamp="2026-01-26 17:06:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 17:06:27.233600564 +0000 UTC m=+1153.757780998" watchObservedRunningTime="2026-01-26 17:06:27.234923981 +0000 UTC m=+1153.759104415" Jan 26 17:06:27 crc kubenswrapper[4754]: I0126 17:06:27.235749 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-fc80-account-create-update-jh5xv" podStartSLOduration=1.235741075 podStartE2EDuration="1.235741075s" podCreationTimestamp="2026-01-26 17:06:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 17:06:27.217362861 +0000 UTC m=+1153.741543495" watchObservedRunningTime="2026-01-26 17:06:27.235741075 +0000 UTC m=+1153.759921509" Jan 26 17:06:27 crc kubenswrapper[4754]: I0126 17:06:27.991487 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/root-account-create-update-p6j95"] Jan 26 17:06:27 crc kubenswrapper[4754]: I0126 17:06:27.993207 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-p6j95" Jan 26 17:06:27 crc kubenswrapper[4754]: I0126 17:06:27.996329 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-mariadb-root-db-secret" Jan 26 17:06:28 crc kubenswrapper[4754]: I0126 17:06:28.006335 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-p6j95"] Jan 26 17:06:28 crc kubenswrapper[4754]: I0126 17:06:28.142739 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7566c04c-bf1d-432b-b2e5-6a989f1f6321-operator-scripts\") pod \"root-account-create-update-p6j95\" (UID: \"7566c04c-bf1d-432b-b2e5-6a989f1f6321\") " pod="openstack/root-account-create-update-p6j95" Jan 26 17:06:28 crc kubenswrapper[4754]: I0126 17:06:28.143109 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c4t6k\" (UniqueName: \"kubernetes.io/projected/7566c04c-bf1d-432b-b2e5-6a989f1f6321-kube-api-access-c4t6k\") pod \"root-account-create-update-p6j95\" (UID: \"7566c04c-bf1d-432b-b2e5-6a989f1f6321\") " pod="openstack/root-account-create-update-p6j95" Jan 26 17:06:28 crc kubenswrapper[4754]: I0126 17:06:28.143186 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/19a0730c-46d0-4029-a86b-812ba1664dcc-etc-swift\") pod \"swift-storage-0\" (UID: \"19a0730c-46d0-4029-a86b-812ba1664dcc\") " pod="openstack/swift-storage-0" Jan 26 17:06:28 crc kubenswrapper[4754]: E0126 17:06:28.143427 4754 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Jan 26 17:06:28 crc kubenswrapper[4754]: E0126 17:06:28.143450 4754 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Jan 26 17:06:28 crc kubenswrapper[4754]: E0126 17:06:28.143502 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/19a0730c-46d0-4029-a86b-812ba1664dcc-etc-swift podName:19a0730c-46d0-4029-a86b-812ba1664dcc nodeName:}" failed. No retries permitted until 2026-01-26 17:06:32.143484666 +0000 UTC m=+1158.667665110 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/19a0730c-46d0-4029-a86b-812ba1664dcc-etc-swift") pod "swift-storage-0" (UID: "19a0730c-46d0-4029-a86b-812ba1664dcc") : configmap "swift-ring-files" not found Jan 26 17:06:28 crc kubenswrapper[4754]: I0126 17:06:28.207188 4754 generic.go:334] "Generic (PLEG): container finished" podID="e2c6020a-c674-43b0-9698-b6b31e0ab5b8" containerID="e6362fc04381ff9a7db06b0751f7ace044043ec05a105674da46640b67248fc0" exitCode=0 Jan 26 17:06:28 crc kubenswrapper[4754]: I0126 17:06:28.207281 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-fc80-account-create-update-jh5xv" event={"ID":"e2c6020a-c674-43b0-9698-b6b31e0ab5b8","Type":"ContainerDied","Data":"e6362fc04381ff9a7db06b0751f7ace044043ec05a105674da46640b67248fc0"} Jan 26 17:06:28 crc kubenswrapper[4754]: I0126 17:06:28.209632 4754 generic.go:334] "Generic (PLEG): container finished" podID="07ac2529-0a5d-4a79-95cf-2b5040e67f46" containerID="e84f43eb249c92fd72b6e9bc45e2e579ad9889dd0ac77ec6d06f9ec1235954ec" exitCode=0 Jan 26 17:06:28 crc kubenswrapper[4754]: I0126 17:06:28.209733 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-plgbz" event={"ID":"07ac2529-0a5d-4a79-95cf-2b5040e67f46","Type":"ContainerDied","Data":"e84f43eb249c92fd72b6e9bc45e2e579ad9889dd0ac77ec6d06f9ec1235954ec"} Jan 26 17:06:28 crc kubenswrapper[4754]: I0126 17:06:28.245083 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7566c04c-bf1d-432b-b2e5-6a989f1f6321-operator-scripts\") pod \"root-account-create-update-p6j95\" (UID: \"7566c04c-bf1d-432b-b2e5-6a989f1f6321\") " pod="openstack/root-account-create-update-p6j95" Jan 26 17:06:28 crc kubenswrapper[4754]: I0126 17:06:28.245148 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c4t6k\" (UniqueName: \"kubernetes.io/projected/7566c04c-bf1d-432b-b2e5-6a989f1f6321-kube-api-access-c4t6k\") pod \"root-account-create-update-p6j95\" (UID: \"7566c04c-bf1d-432b-b2e5-6a989f1f6321\") " pod="openstack/root-account-create-update-p6j95" Jan 26 17:06:28 crc kubenswrapper[4754]: I0126 17:06:28.245940 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7566c04c-bf1d-432b-b2e5-6a989f1f6321-operator-scripts\") pod \"root-account-create-update-p6j95\" (UID: \"7566c04c-bf1d-432b-b2e5-6a989f1f6321\") " pod="openstack/root-account-create-update-p6j95" Jan 26 17:06:28 crc kubenswrapper[4754]: I0126 17:06:28.267484 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c4t6k\" (UniqueName: \"kubernetes.io/projected/7566c04c-bf1d-432b-b2e5-6a989f1f6321-kube-api-access-c4t6k\") pod \"root-account-create-update-p6j95\" (UID: \"7566c04c-bf1d-432b-b2e5-6a989f1f6321\") " pod="openstack/root-account-create-update-p6j95" Jan 26 17:06:28 crc kubenswrapper[4754]: I0126 17:06:28.320405 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-p6j95" Jan 26 17:06:29 crc kubenswrapper[4754]: I0126 17:06:29.854507 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-plgbz" Jan 26 17:06:29 crc kubenswrapper[4754]: I0126 17:06:29.858889 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-fc80-account-create-update-jh5xv" Jan 26 17:06:29 crc kubenswrapper[4754]: I0126 17:06:29.981353 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7cbcg\" (UniqueName: \"kubernetes.io/projected/07ac2529-0a5d-4a79-95cf-2b5040e67f46-kube-api-access-7cbcg\") pod \"07ac2529-0a5d-4a79-95cf-2b5040e67f46\" (UID: \"07ac2529-0a5d-4a79-95cf-2b5040e67f46\") " Jan 26 17:06:29 crc kubenswrapper[4754]: I0126 17:06:29.981839 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e2c6020a-c674-43b0-9698-b6b31e0ab5b8-operator-scripts\") pod \"e2c6020a-c674-43b0-9698-b6b31e0ab5b8\" (UID: \"e2c6020a-c674-43b0-9698-b6b31e0ab5b8\") " Jan 26 17:06:29 crc kubenswrapper[4754]: I0126 17:06:29.981924 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j8zwl\" (UniqueName: \"kubernetes.io/projected/e2c6020a-c674-43b0-9698-b6b31e0ab5b8-kube-api-access-j8zwl\") pod \"e2c6020a-c674-43b0-9698-b6b31e0ab5b8\" (UID: \"e2c6020a-c674-43b0-9698-b6b31e0ab5b8\") " Jan 26 17:06:29 crc kubenswrapper[4754]: I0126 17:06:29.982157 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/07ac2529-0a5d-4a79-95cf-2b5040e67f46-operator-scripts\") pod \"07ac2529-0a5d-4a79-95cf-2b5040e67f46\" (UID: \"07ac2529-0a5d-4a79-95cf-2b5040e67f46\") " Jan 26 17:06:29 crc kubenswrapper[4754]: I0126 17:06:29.982833 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e2c6020a-c674-43b0-9698-b6b31e0ab5b8-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "e2c6020a-c674-43b0-9698-b6b31e0ab5b8" (UID: "e2c6020a-c674-43b0-9698-b6b31e0ab5b8"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:06:29 crc kubenswrapper[4754]: I0126 17:06:29.983094 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/07ac2529-0a5d-4a79-95cf-2b5040e67f46-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "07ac2529-0a5d-4a79-95cf-2b5040e67f46" (UID: "07ac2529-0a5d-4a79-95cf-2b5040e67f46"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:06:29 crc kubenswrapper[4754]: I0126 17:06:29.993719 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/07ac2529-0a5d-4a79-95cf-2b5040e67f46-kube-api-access-7cbcg" (OuterVolumeSpecName: "kube-api-access-7cbcg") pod "07ac2529-0a5d-4a79-95cf-2b5040e67f46" (UID: "07ac2529-0a5d-4a79-95cf-2b5040e67f46"). InnerVolumeSpecName "kube-api-access-7cbcg". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:06:29 crc kubenswrapper[4754]: I0126 17:06:29.993828 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e2c6020a-c674-43b0-9698-b6b31e0ab5b8-kube-api-access-j8zwl" (OuterVolumeSpecName: "kube-api-access-j8zwl") pod "e2c6020a-c674-43b0-9698-b6b31e0ab5b8" (UID: "e2c6020a-c674-43b0-9698-b6b31e0ab5b8"). InnerVolumeSpecName "kube-api-access-j8zwl". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:06:30 crc kubenswrapper[4754]: I0126 17:06:30.084794 4754 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/07ac2529-0a5d-4a79-95cf-2b5040e67f46-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 26 17:06:30 crc kubenswrapper[4754]: I0126 17:06:30.084826 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7cbcg\" (UniqueName: \"kubernetes.io/projected/07ac2529-0a5d-4a79-95cf-2b5040e67f46-kube-api-access-7cbcg\") on node \"crc\" DevicePath \"\"" Jan 26 17:06:30 crc kubenswrapper[4754]: I0126 17:06:30.084837 4754 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e2c6020a-c674-43b0-9698-b6b31e0ab5b8-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 26 17:06:30 crc kubenswrapper[4754]: I0126 17:06:30.084845 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j8zwl\" (UniqueName: \"kubernetes.io/projected/e2c6020a-c674-43b0-9698-b6b31e0ab5b8-kube-api-access-j8zwl\") on node \"crc\" DevicePath \"\"" Jan 26 17:06:30 crc kubenswrapper[4754]: I0126 17:06:30.210809 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-p6j95"] Jan 26 17:06:30 crc kubenswrapper[4754]: I0126 17:06:30.229002 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-p6j95" event={"ID":"7566c04c-bf1d-432b-b2e5-6a989f1f6321","Type":"ContainerStarted","Data":"c89cfbb65005c52646b78371635e862d97fd1aa6b0f41e09f9590612678068fc"} Jan 26 17:06:30 crc kubenswrapper[4754]: I0126 17:06:30.230657 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-sl9kb" event={"ID":"cb88274a-904f-4827-9518-81b79a0e6a42","Type":"ContainerStarted","Data":"5871f81906f8fa3b33f30e4738c9af6bcc767d3e04bf1d4b2a2b30271ade6144"} Jan 26 17:06:30 crc kubenswrapper[4754]: I0126 17:06:30.234194 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-fc80-account-create-update-jh5xv" event={"ID":"e2c6020a-c674-43b0-9698-b6b31e0ab5b8","Type":"ContainerDied","Data":"84db69390c57afbb8f02d1cf32b8d53df3043329a6e3380ae0e701825296a9b7"} Jan 26 17:06:30 crc kubenswrapper[4754]: I0126 17:06:30.234229 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="84db69390c57afbb8f02d1cf32b8d53df3043329a6e3380ae0e701825296a9b7" Jan 26 17:06:30 crc kubenswrapper[4754]: I0126 17:06:30.234304 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-fc80-account-create-update-jh5xv" Jan 26 17:06:30 crc kubenswrapper[4754]: I0126 17:06:30.240296 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-plgbz" event={"ID":"07ac2529-0a5d-4a79-95cf-2b5040e67f46","Type":"ContainerDied","Data":"37a1abe949c9a09b07d9183243f5bad5bf49e5ca628bc0e1a39ab9bb25710a30"} Jan 26 17:06:30 crc kubenswrapper[4754]: I0126 17:06:30.240351 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="37a1abe949c9a09b07d9183243f5bad5bf49e5ca628bc0e1a39ab9bb25710a30" Jan 26 17:06:30 crc kubenswrapper[4754]: I0126 17:06:30.240372 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-plgbz" Jan 26 17:06:30 crc kubenswrapper[4754]: I0126 17:06:30.264678 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-sl9kb" podStartSLOduration=2.316995886 podStartE2EDuration="6.264646169s" podCreationTimestamp="2026-01-26 17:06:24 +0000 UTC" firstStartedPulling="2026-01-26 17:06:25.80460733 +0000 UTC m=+1152.328787764" lastFinishedPulling="2026-01-26 17:06:29.752257613 +0000 UTC m=+1156.276438047" observedRunningTime="2026-01-26 17:06:30.261037458 +0000 UTC m=+1156.785217892" watchObservedRunningTime="2026-01-26 17:06:30.264646169 +0000 UTC m=+1156.788826603" Jan 26 17:06:31 crc kubenswrapper[4754]: I0126 17:06:31.248661 4754 generic.go:334] "Generic (PLEG): container finished" podID="7566c04c-bf1d-432b-b2e5-6a989f1f6321" containerID="b8309f296df385f6aabf8972db308f5cb5f3d17557ebc9629b2a87cdec9cc273" exitCode=0 Jan 26 17:06:31 crc kubenswrapper[4754]: I0126 17:06:31.248780 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-p6j95" event={"ID":"7566c04c-bf1d-432b-b2e5-6a989f1f6321","Type":"ContainerDied","Data":"b8309f296df385f6aabf8972db308f5cb5f3d17557ebc9629b2a87cdec9cc273"} Jan 26 17:06:31 crc kubenswrapper[4754]: I0126 17:06:31.419782 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-qbqp5"] Jan 26 17:06:31 crc kubenswrapper[4754]: E0126 17:06:31.420197 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="07ac2529-0a5d-4a79-95cf-2b5040e67f46" containerName="mariadb-database-create" Jan 26 17:06:31 crc kubenswrapper[4754]: I0126 17:06:31.420218 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="07ac2529-0a5d-4a79-95cf-2b5040e67f46" containerName="mariadb-database-create" Jan 26 17:06:31 crc kubenswrapper[4754]: E0126 17:06:31.420237 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e2c6020a-c674-43b0-9698-b6b31e0ab5b8" containerName="mariadb-account-create-update" Jan 26 17:06:31 crc kubenswrapper[4754]: I0126 17:06:31.420246 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="e2c6020a-c674-43b0-9698-b6b31e0ab5b8" containerName="mariadb-account-create-update" Jan 26 17:06:31 crc kubenswrapper[4754]: I0126 17:06:31.420439 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="07ac2529-0a5d-4a79-95cf-2b5040e67f46" containerName="mariadb-database-create" Jan 26 17:06:31 crc kubenswrapper[4754]: I0126 17:06:31.420465 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="e2c6020a-c674-43b0-9698-b6b31e0ab5b8" containerName="mariadb-account-create-update" Jan 26 17:06:31 crc kubenswrapper[4754]: I0126 17:06:31.421061 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-qbqp5" Jan 26 17:06:31 crc kubenswrapper[4754]: I0126 17:06:31.423076 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-hph75" Jan 26 17:06:31 crc kubenswrapper[4754]: I0126 17:06:31.423525 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Jan 26 17:06:31 crc kubenswrapper[4754]: I0126 17:06:31.435379 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-qbqp5"] Jan 26 17:06:31 crc kubenswrapper[4754]: I0126 17:06:31.510584 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bae8ecc7-8fac-4a63-9e99-4e7543eae5af-combined-ca-bundle\") pod \"glance-db-sync-qbqp5\" (UID: \"bae8ecc7-8fac-4a63-9e99-4e7543eae5af\") " pod="openstack/glance-db-sync-qbqp5" Jan 26 17:06:31 crc kubenswrapper[4754]: I0126 17:06:31.510651 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z4sm2\" (UniqueName: \"kubernetes.io/projected/bae8ecc7-8fac-4a63-9e99-4e7543eae5af-kube-api-access-z4sm2\") pod \"glance-db-sync-qbqp5\" (UID: \"bae8ecc7-8fac-4a63-9e99-4e7543eae5af\") " pod="openstack/glance-db-sync-qbqp5" Jan 26 17:06:31 crc kubenswrapper[4754]: I0126 17:06:31.510706 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/bae8ecc7-8fac-4a63-9e99-4e7543eae5af-db-sync-config-data\") pod \"glance-db-sync-qbqp5\" (UID: \"bae8ecc7-8fac-4a63-9e99-4e7543eae5af\") " pod="openstack/glance-db-sync-qbqp5" Jan 26 17:06:31 crc kubenswrapper[4754]: I0126 17:06:31.510736 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bae8ecc7-8fac-4a63-9e99-4e7543eae5af-config-data\") pod \"glance-db-sync-qbqp5\" (UID: \"bae8ecc7-8fac-4a63-9e99-4e7543eae5af\") " pod="openstack/glance-db-sync-qbqp5" Jan 26 17:06:31 crc kubenswrapper[4754]: I0126 17:06:31.612459 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z4sm2\" (UniqueName: \"kubernetes.io/projected/bae8ecc7-8fac-4a63-9e99-4e7543eae5af-kube-api-access-z4sm2\") pod \"glance-db-sync-qbqp5\" (UID: \"bae8ecc7-8fac-4a63-9e99-4e7543eae5af\") " pod="openstack/glance-db-sync-qbqp5" Jan 26 17:06:31 crc kubenswrapper[4754]: I0126 17:06:31.612547 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/bae8ecc7-8fac-4a63-9e99-4e7543eae5af-db-sync-config-data\") pod \"glance-db-sync-qbqp5\" (UID: \"bae8ecc7-8fac-4a63-9e99-4e7543eae5af\") " pod="openstack/glance-db-sync-qbqp5" Jan 26 17:06:31 crc kubenswrapper[4754]: I0126 17:06:31.612586 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bae8ecc7-8fac-4a63-9e99-4e7543eae5af-config-data\") pod \"glance-db-sync-qbqp5\" (UID: \"bae8ecc7-8fac-4a63-9e99-4e7543eae5af\") " pod="openstack/glance-db-sync-qbqp5" Jan 26 17:06:31 crc kubenswrapper[4754]: I0126 17:06:31.612740 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bae8ecc7-8fac-4a63-9e99-4e7543eae5af-combined-ca-bundle\") pod \"glance-db-sync-qbqp5\" (UID: \"bae8ecc7-8fac-4a63-9e99-4e7543eae5af\") " pod="openstack/glance-db-sync-qbqp5" Jan 26 17:06:31 crc kubenswrapper[4754]: I0126 17:06:31.619501 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bae8ecc7-8fac-4a63-9e99-4e7543eae5af-config-data\") pod \"glance-db-sync-qbqp5\" (UID: \"bae8ecc7-8fac-4a63-9e99-4e7543eae5af\") " pod="openstack/glance-db-sync-qbqp5" Jan 26 17:06:31 crc kubenswrapper[4754]: I0126 17:06:31.620440 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/bae8ecc7-8fac-4a63-9e99-4e7543eae5af-db-sync-config-data\") pod \"glance-db-sync-qbqp5\" (UID: \"bae8ecc7-8fac-4a63-9e99-4e7543eae5af\") " pod="openstack/glance-db-sync-qbqp5" Jan 26 17:06:31 crc kubenswrapper[4754]: I0126 17:06:31.621367 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bae8ecc7-8fac-4a63-9e99-4e7543eae5af-combined-ca-bundle\") pod \"glance-db-sync-qbqp5\" (UID: \"bae8ecc7-8fac-4a63-9e99-4e7543eae5af\") " pod="openstack/glance-db-sync-qbqp5" Jan 26 17:06:31 crc kubenswrapper[4754]: I0126 17:06:31.628813 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z4sm2\" (UniqueName: \"kubernetes.io/projected/bae8ecc7-8fac-4a63-9e99-4e7543eae5af-kube-api-access-z4sm2\") pod \"glance-db-sync-qbqp5\" (UID: \"bae8ecc7-8fac-4a63-9e99-4e7543eae5af\") " pod="openstack/glance-db-sync-qbqp5" Jan 26 17:06:31 crc kubenswrapper[4754]: I0126 17:06:31.740115 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-qbqp5" Jan 26 17:06:32 crc kubenswrapper[4754]: I0126 17:06:32.222649 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/19a0730c-46d0-4029-a86b-812ba1664dcc-etc-swift\") pod \"swift-storage-0\" (UID: \"19a0730c-46d0-4029-a86b-812ba1664dcc\") " pod="openstack/swift-storage-0" Jan 26 17:06:32 crc kubenswrapper[4754]: E0126 17:06:32.222820 4754 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Jan 26 17:06:32 crc kubenswrapper[4754]: E0126 17:06:32.223137 4754 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Jan 26 17:06:32 crc kubenswrapper[4754]: E0126 17:06:32.223197 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/19a0730c-46d0-4029-a86b-812ba1664dcc-etc-swift podName:19a0730c-46d0-4029-a86b-812ba1664dcc nodeName:}" failed. No retries permitted until 2026-01-26 17:06:40.223178738 +0000 UTC m=+1166.747359162 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/19a0730c-46d0-4029-a86b-812ba1664dcc-etc-swift") pod "swift-storage-0" (UID: "19a0730c-46d0-4029-a86b-812ba1664dcc") : configmap "swift-ring-files" not found Jan 26 17:06:32 crc kubenswrapper[4754]: I0126 17:06:32.411292 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-qbqp5"] Jan 26 17:06:32 crc kubenswrapper[4754]: I0126 17:06:32.678456 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-p6j95" Jan 26 17:06:32 crc kubenswrapper[4754]: I0126 17:06:32.834860 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c4t6k\" (UniqueName: \"kubernetes.io/projected/7566c04c-bf1d-432b-b2e5-6a989f1f6321-kube-api-access-c4t6k\") pod \"7566c04c-bf1d-432b-b2e5-6a989f1f6321\" (UID: \"7566c04c-bf1d-432b-b2e5-6a989f1f6321\") " Jan 26 17:06:32 crc kubenswrapper[4754]: I0126 17:06:32.834896 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7566c04c-bf1d-432b-b2e5-6a989f1f6321-operator-scripts\") pod \"7566c04c-bf1d-432b-b2e5-6a989f1f6321\" (UID: \"7566c04c-bf1d-432b-b2e5-6a989f1f6321\") " Jan 26 17:06:32 crc kubenswrapper[4754]: I0126 17:06:32.836051 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7566c04c-bf1d-432b-b2e5-6a989f1f6321-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "7566c04c-bf1d-432b-b2e5-6a989f1f6321" (UID: "7566c04c-bf1d-432b-b2e5-6a989f1f6321"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:06:32 crc kubenswrapper[4754]: I0126 17:06:32.847523 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7566c04c-bf1d-432b-b2e5-6a989f1f6321-kube-api-access-c4t6k" (OuterVolumeSpecName: "kube-api-access-c4t6k") pod "7566c04c-bf1d-432b-b2e5-6a989f1f6321" (UID: "7566c04c-bf1d-432b-b2e5-6a989f1f6321"). InnerVolumeSpecName "kube-api-access-c4t6k". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:06:32 crc kubenswrapper[4754]: I0126 17:06:32.936432 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c4t6k\" (UniqueName: \"kubernetes.io/projected/7566c04c-bf1d-432b-b2e5-6a989f1f6321-kube-api-access-c4t6k\") on node \"crc\" DevicePath \"\"" Jan 26 17:06:32 crc kubenswrapper[4754]: I0126 17:06:32.936473 4754 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7566c04c-bf1d-432b-b2e5-6a989f1f6321-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 26 17:06:33 crc kubenswrapper[4754]: I0126 17:06:33.173789 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Jan 26 17:06:33 crc kubenswrapper[4754]: I0126 17:06:33.274201 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-p6j95" event={"ID":"7566c04c-bf1d-432b-b2e5-6a989f1f6321","Type":"ContainerDied","Data":"c89cfbb65005c52646b78371635e862d97fd1aa6b0f41e09f9590612678068fc"} Jan 26 17:06:33 crc kubenswrapper[4754]: I0126 17:06:33.274242 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c89cfbb65005c52646b78371635e862d97fd1aa6b0f41e09f9590612678068fc" Jan 26 17:06:33 crc kubenswrapper[4754]: I0126 17:06:33.274246 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-p6j95" Jan 26 17:06:33 crc kubenswrapper[4754]: I0126 17:06:33.275354 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-qbqp5" event={"ID":"bae8ecc7-8fac-4a63-9e99-4e7543eae5af","Type":"ContainerStarted","Data":"d1d22ae82d685b4b2ac68b19cf28f22b7537f243a37e0d0cee27ebbc64da0f64"} Jan 26 17:06:33 crc kubenswrapper[4754]: I0126 17:06:33.434849 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-698758b865-z5pcq" Jan 26 17:06:33 crc kubenswrapper[4754]: I0126 17:06:33.484321 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-r9jrq"] Jan 26 17:06:33 crc kubenswrapper[4754]: I0126 17:06:33.484596 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-86db49b7ff-r9jrq" podUID="0df51b10-10ea-4822-ad9b-e952ba4673f8" containerName="dnsmasq-dns" containerID="cri-o://cd0b28dc0eb9a531434655b61d84d44919718ba6e136e2e784afe94300f332f5" gracePeriod=10 Jan 26 17:06:34 crc kubenswrapper[4754]: I0126 17:06:34.283132 4754 generic.go:334] "Generic (PLEG): container finished" podID="0df51b10-10ea-4822-ad9b-e952ba4673f8" containerID="cd0b28dc0eb9a531434655b61d84d44919718ba6e136e2e784afe94300f332f5" exitCode=0 Jan 26 17:06:34 crc kubenswrapper[4754]: I0126 17:06:34.283205 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-r9jrq" event={"ID":"0df51b10-10ea-4822-ad9b-e952ba4673f8","Type":"ContainerDied","Data":"cd0b28dc0eb9a531434655b61d84d44919718ba6e136e2e784afe94300f332f5"} Jan 26 17:06:34 crc kubenswrapper[4754]: I0126 17:06:34.361840 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/root-account-create-update-p6j95"] Jan 26 17:06:34 crc kubenswrapper[4754]: I0126 17:06:34.370801 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/root-account-create-update-p6j95"] Jan 26 17:06:35 crc kubenswrapper[4754]: I0126 17:06:35.157720 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-r9jrq" Jan 26 17:06:35 crc kubenswrapper[4754]: I0126 17:06:35.278951 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0df51b10-10ea-4822-ad9b-e952ba4673f8-ovsdbserver-nb\") pod \"0df51b10-10ea-4822-ad9b-e952ba4673f8\" (UID: \"0df51b10-10ea-4822-ad9b-e952ba4673f8\") " Jan 26 17:06:35 crc kubenswrapper[4754]: I0126 17:06:35.279012 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0df51b10-10ea-4822-ad9b-e952ba4673f8-ovsdbserver-sb\") pod \"0df51b10-10ea-4822-ad9b-e952ba4673f8\" (UID: \"0df51b10-10ea-4822-ad9b-e952ba4673f8\") " Jan 26 17:06:35 crc kubenswrapper[4754]: I0126 17:06:35.279386 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0df51b10-10ea-4822-ad9b-e952ba4673f8-dns-svc\") pod \"0df51b10-10ea-4822-ad9b-e952ba4673f8\" (UID: \"0df51b10-10ea-4822-ad9b-e952ba4673f8\") " Jan 26 17:06:35 crc kubenswrapper[4754]: I0126 17:06:35.280480 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0df51b10-10ea-4822-ad9b-e952ba4673f8-config\") pod \"0df51b10-10ea-4822-ad9b-e952ba4673f8\" (UID: \"0df51b10-10ea-4822-ad9b-e952ba4673f8\") " Jan 26 17:06:35 crc kubenswrapper[4754]: I0126 17:06:35.280567 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2s5m4\" (UniqueName: \"kubernetes.io/projected/0df51b10-10ea-4822-ad9b-e952ba4673f8-kube-api-access-2s5m4\") pod \"0df51b10-10ea-4822-ad9b-e952ba4673f8\" (UID: \"0df51b10-10ea-4822-ad9b-e952ba4673f8\") " Jan 26 17:06:35 crc kubenswrapper[4754]: I0126 17:06:35.285540 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0df51b10-10ea-4822-ad9b-e952ba4673f8-kube-api-access-2s5m4" (OuterVolumeSpecName: "kube-api-access-2s5m4") pod "0df51b10-10ea-4822-ad9b-e952ba4673f8" (UID: "0df51b10-10ea-4822-ad9b-e952ba4673f8"). InnerVolumeSpecName "kube-api-access-2s5m4". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:06:35 crc kubenswrapper[4754]: I0126 17:06:35.292783 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-r9jrq" event={"ID":"0df51b10-10ea-4822-ad9b-e952ba4673f8","Type":"ContainerDied","Data":"e8f45055324a26f98dc4083fe586bd50ba46807545bfeaf4613b880ea1fa974e"} Jan 26 17:06:35 crc kubenswrapper[4754]: I0126 17:06:35.292828 4754 scope.go:117] "RemoveContainer" containerID="cd0b28dc0eb9a531434655b61d84d44919718ba6e136e2e784afe94300f332f5" Jan 26 17:06:35 crc kubenswrapper[4754]: I0126 17:06:35.292935 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-r9jrq" Jan 26 17:06:35 crc kubenswrapper[4754]: I0126 17:06:35.323179 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0df51b10-10ea-4822-ad9b-e952ba4673f8-config" (OuterVolumeSpecName: "config") pod "0df51b10-10ea-4822-ad9b-e952ba4673f8" (UID: "0df51b10-10ea-4822-ad9b-e952ba4673f8"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:06:35 crc kubenswrapper[4754]: I0126 17:06:35.324438 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0df51b10-10ea-4822-ad9b-e952ba4673f8-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "0df51b10-10ea-4822-ad9b-e952ba4673f8" (UID: "0df51b10-10ea-4822-ad9b-e952ba4673f8"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:06:35 crc kubenswrapper[4754]: I0126 17:06:35.329771 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0df51b10-10ea-4822-ad9b-e952ba4673f8-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "0df51b10-10ea-4822-ad9b-e952ba4673f8" (UID: "0df51b10-10ea-4822-ad9b-e952ba4673f8"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:06:35 crc kubenswrapper[4754]: I0126 17:06:35.361641 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0df51b10-10ea-4822-ad9b-e952ba4673f8-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "0df51b10-10ea-4822-ad9b-e952ba4673f8" (UID: "0df51b10-10ea-4822-ad9b-e952ba4673f8"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:06:35 crc kubenswrapper[4754]: I0126 17:06:35.382501 4754 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0df51b10-10ea-4822-ad9b-e952ba4673f8-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Jan 26 17:06:35 crc kubenswrapper[4754]: I0126 17:06:35.382533 4754 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0df51b10-10ea-4822-ad9b-e952ba4673f8-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Jan 26 17:06:35 crc kubenswrapper[4754]: I0126 17:06:35.382541 4754 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0df51b10-10ea-4822-ad9b-e952ba4673f8-dns-svc\") on node \"crc\" DevicePath \"\"" Jan 26 17:06:35 crc kubenswrapper[4754]: I0126 17:06:35.382554 4754 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0df51b10-10ea-4822-ad9b-e952ba4673f8-config\") on node \"crc\" DevicePath \"\"" Jan 26 17:06:35 crc kubenswrapper[4754]: I0126 17:06:35.382563 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2s5m4\" (UniqueName: \"kubernetes.io/projected/0df51b10-10ea-4822-ad9b-e952ba4673f8-kube-api-access-2s5m4\") on node \"crc\" DevicePath \"\"" Jan 26 17:06:35 crc kubenswrapper[4754]: I0126 17:06:35.403734 4754 scope.go:117] "RemoveContainer" containerID="62dc8ee2c3e18b05a5ad1bdbc9c5d7af30f0f506b69d408c389b431fec3b3fb8" Jan 26 17:06:35 crc kubenswrapper[4754]: I0126 17:06:35.626864 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-r9jrq"] Jan 26 17:06:35 crc kubenswrapper[4754]: I0126 17:06:35.637480 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-r9jrq"] Jan 26 17:06:35 crc kubenswrapper[4754]: I0126 17:06:35.776198 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0df51b10-10ea-4822-ad9b-e952ba4673f8" path="/var/lib/kubelet/pods/0df51b10-10ea-4822-ad9b-e952ba4673f8/volumes" Jan 26 17:06:35 crc kubenswrapper[4754]: I0126 17:06:35.776777 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7566c04c-bf1d-432b-b2e5-6a989f1f6321" path="/var/lib/kubelet/pods/7566c04c-bf1d-432b-b2e5-6a989f1f6321/volumes" Jan 26 17:06:35 crc kubenswrapper[4754]: I0126 17:06:35.962676 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Jan 26 17:06:37 crc kubenswrapper[4754]: I0126 17:06:37.129541 4754 patch_prober.go:28] interesting pod/machine-config-daemon-x65wv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 26 17:06:37 crc kubenswrapper[4754]: I0126 17:06:37.129612 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 26 17:06:37 crc kubenswrapper[4754]: I0126 17:06:37.129682 4754 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" Jan 26 17:06:37 crc kubenswrapper[4754]: I0126 17:06:37.130469 4754 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"a1662bcf9213c8129b38140c4208b765dd0d2dd18def4d53852919fa5dab4d3e"} pod="openshift-machine-config-operator/machine-config-daemon-x65wv" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 26 17:06:37 crc kubenswrapper[4754]: I0126 17:06:37.130527 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" containerName="machine-config-daemon" containerID="cri-o://a1662bcf9213c8129b38140c4208b765dd0d2dd18def4d53852919fa5dab4d3e" gracePeriod=600 Jan 26 17:06:38 crc kubenswrapper[4754]: I0126 17:06:38.014370 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/root-account-create-update-gwx8r"] Jan 26 17:06:38 crc kubenswrapper[4754]: E0126 17:06:38.015059 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7566c04c-bf1d-432b-b2e5-6a989f1f6321" containerName="mariadb-account-create-update" Jan 26 17:06:38 crc kubenswrapper[4754]: I0126 17:06:38.015083 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="7566c04c-bf1d-432b-b2e5-6a989f1f6321" containerName="mariadb-account-create-update" Jan 26 17:06:38 crc kubenswrapper[4754]: E0126 17:06:38.015105 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0df51b10-10ea-4822-ad9b-e952ba4673f8" containerName="init" Jan 26 17:06:38 crc kubenswrapper[4754]: I0126 17:06:38.015113 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="0df51b10-10ea-4822-ad9b-e952ba4673f8" containerName="init" Jan 26 17:06:38 crc kubenswrapper[4754]: E0126 17:06:38.015140 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0df51b10-10ea-4822-ad9b-e952ba4673f8" containerName="dnsmasq-dns" Jan 26 17:06:38 crc kubenswrapper[4754]: I0126 17:06:38.015148 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="0df51b10-10ea-4822-ad9b-e952ba4673f8" containerName="dnsmasq-dns" Jan 26 17:06:38 crc kubenswrapper[4754]: I0126 17:06:38.015402 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="7566c04c-bf1d-432b-b2e5-6a989f1f6321" containerName="mariadb-account-create-update" Jan 26 17:06:38 crc kubenswrapper[4754]: I0126 17:06:38.015425 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="0df51b10-10ea-4822-ad9b-e952ba4673f8" containerName="dnsmasq-dns" Jan 26 17:06:38 crc kubenswrapper[4754]: I0126 17:06:38.016045 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-gwx8r" Jan 26 17:06:38 crc kubenswrapper[4754]: I0126 17:06:38.020822 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-mariadb-root-db-secret" Jan 26 17:06:38 crc kubenswrapper[4754]: I0126 17:06:38.023423 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-gwx8r"] Jan 26 17:06:38 crc kubenswrapper[4754]: I0126 17:06:38.128434 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q954r\" (UniqueName: \"kubernetes.io/projected/afc71d32-c836-4b6a-aae7-339d2582da56-kube-api-access-q954r\") pod \"root-account-create-update-gwx8r\" (UID: \"afc71d32-c836-4b6a-aae7-339d2582da56\") " pod="openstack/root-account-create-update-gwx8r" Jan 26 17:06:38 crc kubenswrapper[4754]: I0126 17:06:38.128733 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/afc71d32-c836-4b6a-aae7-339d2582da56-operator-scripts\") pod \"root-account-create-update-gwx8r\" (UID: \"afc71d32-c836-4b6a-aae7-339d2582da56\") " pod="openstack/root-account-create-update-gwx8r" Jan 26 17:06:38 crc kubenswrapper[4754]: I0126 17:06:38.230877 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q954r\" (UniqueName: \"kubernetes.io/projected/afc71d32-c836-4b6a-aae7-339d2582da56-kube-api-access-q954r\") pod \"root-account-create-update-gwx8r\" (UID: \"afc71d32-c836-4b6a-aae7-339d2582da56\") " pod="openstack/root-account-create-update-gwx8r" Jan 26 17:06:38 crc kubenswrapper[4754]: I0126 17:06:38.231014 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/afc71d32-c836-4b6a-aae7-339d2582da56-operator-scripts\") pod \"root-account-create-update-gwx8r\" (UID: \"afc71d32-c836-4b6a-aae7-339d2582da56\") " pod="openstack/root-account-create-update-gwx8r" Jan 26 17:06:38 crc kubenswrapper[4754]: I0126 17:06:38.232037 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/afc71d32-c836-4b6a-aae7-339d2582da56-operator-scripts\") pod \"root-account-create-update-gwx8r\" (UID: \"afc71d32-c836-4b6a-aae7-339d2582da56\") " pod="openstack/root-account-create-update-gwx8r" Jan 26 17:06:38 crc kubenswrapper[4754]: I0126 17:06:38.277318 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q954r\" (UniqueName: \"kubernetes.io/projected/afc71d32-c836-4b6a-aae7-339d2582da56-kube-api-access-q954r\") pod \"root-account-create-update-gwx8r\" (UID: \"afc71d32-c836-4b6a-aae7-339d2582da56\") " pod="openstack/root-account-create-update-gwx8r" Jan 26 17:06:38 crc kubenswrapper[4754]: I0126 17:06:38.321362 4754 generic.go:334] "Generic (PLEG): container finished" podID="8c3718a4-f354-4284-92e0-fdfb45a692bd" containerID="a1662bcf9213c8129b38140c4208b765dd0d2dd18def4d53852919fa5dab4d3e" exitCode=0 Jan 26 17:06:38 crc kubenswrapper[4754]: I0126 17:06:38.321429 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" event={"ID":"8c3718a4-f354-4284-92e0-fdfb45a692bd","Type":"ContainerDied","Data":"a1662bcf9213c8129b38140c4208b765dd0d2dd18def4d53852919fa5dab4d3e"} Jan 26 17:06:38 crc kubenswrapper[4754]: I0126 17:06:38.321467 4754 scope.go:117] "RemoveContainer" containerID="2d157bed70bdc2504afc3419bb56bdffbceb22c7b7e8e1c5315d44974048523d" Jan 26 17:06:38 crc kubenswrapper[4754]: I0126 17:06:38.345712 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-gwx8r" Jan 26 17:06:38 crc kubenswrapper[4754]: I0126 17:06:38.772969 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-gwx8r"] Jan 26 17:06:38 crc kubenswrapper[4754]: W0126 17:06:38.794620 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podafc71d32_c836_4b6a_aae7_339d2582da56.slice/crio-0b5b7af38a2a9623e055f41f8498b5feb4e6f4837a7f1471efc22524321c04b9 WatchSource:0}: Error finding container 0b5b7af38a2a9623e055f41f8498b5feb4e6f4837a7f1471efc22524321c04b9: Status 404 returned error can't find the container with id 0b5b7af38a2a9623e055f41f8498b5feb4e6f4837a7f1471efc22524321c04b9 Jan 26 17:06:39 crc kubenswrapper[4754]: I0126 17:06:39.328659 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-gwx8r" event={"ID":"afc71d32-c836-4b6a-aae7-339d2582da56","Type":"ContainerStarted","Data":"0b5b7af38a2a9623e055f41f8498b5feb4e6f4837a7f1471efc22524321c04b9"} Jan 26 17:06:40 crc kubenswrapper[4754]: I0126 17:06:40.269351 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/19a0730c-46d0-4029-a86b-812ba1664dcc-etc-swift\") pod \"swift-storage-0\" (UID: \"19a0730c-46d0-4029-a86b-812ba1664dcc\") " pod="openstack/swift-storage-0" Jan 26 17:06:40 crc kubenswrapper[4754]: E0126 17:06:40.269622 4754 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Jan 26 17:06:40 crc kubenswrapper[4754]: E0126 17:06:40.270014 4754 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Jan 26 17:06:40 crc kubenswrapper[4754]: E0126 17:06:40.270129 4754 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/19a0730c-46d0-4029-a86b-812ba1664dcc-etc-swift podName:19a0730c-46d0-4029-a86b-812ba1664dcc nodeName:}" failed. No retries permitted until 2026-01-26 17:06:56.270093002 +0000 UTC m=+1182.794273476 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/19a0730c-46d0-4029-a86b-812ba1664dcc-etc-swift") pod "swift-storage-0" (UID: "19a0730c-46d0-4029-a86b-812ba1664dcc") : configmap "swift-ring-files" not found Jan 26 17:06:41 crc kubenswrapper[4754]: I0126 17:06:41.351612 4754 generic.go:334] "Generic (PLEG): container finished" podID="afc71d32-c836-4b6a-aae7-339d2582da56" containerID="890780a4f6df86cd0474c7787d35ee7d2ca45b9ed51a3fabc57b69c719947625" exitCode=0 Jan 26 17:06:41 crc kubenswrapper[4754]: I0126 17:06:41.351723 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-gwx8r" event={"ID":"afc71d32-c836-4b6a-aae7-339d2582da56","Type":"ContainerDied","Data":"890780a4f6df86cd0474c7787d35ee7d2ca45b9ed51a3fabc57b69c719947625"} Jan 26 17:06:41 crc kubenswrapper[4754]: I0126 17:06:41.366113 4754 generic.go:334] "Generic (PLEG): container finished" podID="cb88274a-904f-4827-9518-81b79a0e6a42" containerID="5871f81906f8fa3b33f30e4738c9af6bcc767d3e04bf1d4b2a2b30271ade6144" exitCode=0 Jan 26 17:06:41 crc kubenswrapper[4754]: I0126 17:06:41.366230 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-sl9kb" event={"ID":"cb88274a-904f-4827-9518-81b79a0e6a42","Type":"ContainerDied","Data":"5871f81906f8fa3b33f30e4738c9af6bcc767d3e04bf1d4b2a2b30271ade6144"} Jan 26 17:06:41 crc kubenswrapper[4754]: I0126 17:06:41.612643 4754 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-8bgjb" podUID="6db16a49-6566-42bf-91ad-c34be46e7800" containerName="ovn-controller" probeResult="failure" output=< Jan 26 17:06:41 crc kubenswrapper[4754]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Jan 26 17:06:41 crc kubenswrapper[4754]: > Jan 26 17:06:41 crc kubenswrapper[4754]: I0126 17:06:41.654502 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-7psgg" Jan 26 17:06:41 crc kubenswrapper[4754]: I0126 17:06:41.661100 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-7psgg" Jan 26 17:06:41 crc kubenswrapper[4754]: I0126 17:06:41.876481 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-8bgjb-config-bbncl"] Jan 26 17:06:41 crc kubenswrapper[4754]: I0126 17:06:41.878740 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-8bgjb-config-bbncl" Jan 26 17:06:41 crc kubenswrapper[4754]: I0126 17:06:41.880950 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Jan 26 17:06:41 crc kubenswrapper[4754]: I0126 17:06:41.895142 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-8bgjb-config-bbncl"] Jan 26 17:06:42 crc kubenswrapper[4754]: I0126 17:06:42.006432 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/4146f13c-4a4c-4f9c-95d2-5dcedd19d362-var-run\") pod \"ovn-controller-8bgjb-config-bbncl\" (UID: \"4146f13c-4a4c-4f9c-95d2-5dcedd19d362\") " pod="openstack/ovn-controller-8bgjb-config-bbncl" Jan 26 17:06:42 crc kubenswrapper[4754]: I0126 17:06:42.006649 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/4146f13c-4a4c-4f9c-95d2-5dcedd19d362-var-run-ovn\") pod \"ovn-controller-8bgjb-config-bbncl\" (UID: \"4146f13c-4a4c-4f9c-95d2-5dcedd19d362\") " pod="openstack/ovn-controller-8bgjb-config-bbncl" Jan 26 17:06:42 crc kubenswrapper[4754]: I0126 17:06:42.006774 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/4146f13c-4a4c-4f9c-95d2-5dcedd19d362-additional-scripts\") pod \"ovn-controller-8bgjb-config-bbncl\" (UID: \"4146f13c-4a4c-4f9c-95d2-5dcedd19d362\") " pod="openstack/ovn-controller-8bgjb-config-bbncl" Jan 26 17:06:42 crc kubenswrapper[4754]: I0126 17:06:42.006817 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qw5wq\" (UniqueName: \"kubernetes.io/projected/4146f13c-4a4c-4f9c-95d2-5dcedd19d362-kube-api-access-qw5wq\") pod \"ovn-controller-8bgjb-config-bbncl\" (UID: \"4146f13c-4a4c-4f9c-95d2-5dcedd19d362\") " pod="openstack/ovn-controller-8bgjb-config-bbncl" Jan 26 17:06:42 crc kubenswrapper[4754]: I0126 17:06:42.006862 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4146f13c-4a4c-4f9c-95d2-5dcedd19d362-scripts\") pod \"ovn-controller-8bgjb-config-bbncl\" (UID: \"4146f13c-4a4c-4f9c-95d2-5dcedd19d362\") " pod="openstack/ovn-controller-8bgjb-config-bbncl" Jan 26 17:06:42 crc kubenswrapper[4754]: I0126 17:06:42.006931 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/4146f13c-4a4c-4f9c-95d2-5dcedd19d362-var-log-ovn\") pod \"ovn-controller-8bgjb-config-bbncl\" (UID: \"4146f13c-4a4c-4f9c-95d2-5dcedd19d362\") " pod="openstack/ovn-controller-8bgjb-config-bbncl" Jan 26 17:06:42 crc kubenswrapper[4754]: I0126 17:06:42.108958 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/4146f13c-4a4c-4f9c-95d2-5dcedd19d362-additional-scripts\") pod \"ovn-controller-8bgjb-config-bbncl\" (UID: \"4146f13c-4a4c-4f9c-95d2-5dcedd19d362\") " pod="openstack/ovn-controller-8bgjb-config-bbncl" Jan 26 17:06:42 crc kubenswrapper[4754]: I0126 17:06:42.109035 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qw5wq\" (UniqueName: \"kubernetes.io/projected/4146f13c-4a4c-4f9c-95d2-5dcedd19d362-kube-api-access-qw5wq\") pod \"ovn-controller-8bgjb-config-bbncl\" (UID: \"4146f13c-4a4c-4f9c-95d2-5dcedd19d362\") " pod="openstack/ovn-controller-8bgjb-config-bbncl" Jan 26 17:06:42 crc kubenswrapper[4754]: I0126 17:06:42.109125 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4146f13c-4a4c-4f9c-95d2-5dcedd19d362-scripts\") pod \"ovn-controller-8bgjb-config-bbncl\" (UID: \"4146f13c-4a4c-4f9c-95d2-5dcedd19d362\") " pod="openstack/ovn-controller-8bgjb-config-bbncl" Jan 26 17:06:42 crc kubenswrapper[4754]: I0126 17:06:42.109206 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/4146f13c-4a4c-4f9c-95d2-5dcedd19d362-var-log-ovn\") pod \"ovn-controller-8bgjb-config-bbncl\" (UID: \"4146f13c-4a4c-4f9c-95d2-5dcedd19d362\") " pod="openstack/ovn-controller-8bgjb-config-bbncl" Jan 26 17:06:42 crc kubenswrapper[4754]: I0126 17:06:42.109235 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/4146f13c-4a4c-4f9c-95d2-5dcedd19d362-var-run\") pod \"ovn-controller-8bgjb-config-bbncl\" (UID: \"4146f13c-4a4c-4f9c-95d2-5dcedd19d362\") " pod="openstack/ovn-controller-8bgjb-config-bbncl" Jan 26 17:06:42 crc kubenswrapper[4754]: I0126 17:06:42.109348 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/4146f13c-4a4c-4f9c-95d2-5dcedd19d362-var-run-ovn\") pod \"ovn-controller-8bgjb-config-bbncl\" (UID: \"4146f13c-4a4c-4f9c-95d2-5dcedd19d362\") " pod="openstack/ovn-controller-8bgjb-config-bbncl" Jan 26 17:06:42 crc kubenswrapper[4754]: I0126 17:06:42.109684 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/4146f13c-4a4c-4f9c-95d2-5dcedd19d362-var-log-ovn\") pod \"ovn-controller-8bgjb-config-bbncl\" (UID: \"4146f13c-4a4c-4f9c-95d2-5dcedd19d362\") " pod="openstack/ovn-controller-8bgjb-config-bbncl" Jan 26 17:06:42 crc kubenswrapper[4754]: I0126 17:06:42.109693 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/4146f13c-4a4c-4f9c-95d2-5dcedd19d362-var-run\") pod \"ovn-controller-8bgjb-config-bbncl\" (UID: \"4146f13c-4a4c-4f9c-95d2-5dcedd19d362\") " pod="openstack/ovn-controller-8bgjb-config-bbncl" Jan 26 17:06:42 crc kubenswrapper[4754]: I0126 17:06:42.109693 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/4146f13c-4a4c-4f9c-95d2-5dcedd19d362-var-run-ovn\") pod \"ovn-controller-8bgjb-config-bbncl\" (UID: \"4146f13c-4a4c-4f9c-95d2-5dcedd19d362\") " pod="openstack/ovn-controller-8bgjb-config-bbncl" Jan 26 17:06:42 crc kubenswrapper[4754]: I0126 17:06:42.111055 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/4146f13c-4a4c-4f9c-95d2-5dcedd19d362-additional-scripts\") pod \"ovn-controller-8bgjb-config-bbncl\" (UID: \"4146f13c-4a4c-4f9c-95d2-5dcedd19d362\") " pod="openstack/ovn-controller-8bgjb-config-bbncl" Jan 26 17:06:42 crc kubenswrapper[4754]: I0126 17:06:42.123261 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4146f13c-4a4c-4f9c-95d2-5dcedd19d362-scripts\") pod \"ovn-controller-8bgjb-config-bbncl\" (UID: \"4146f13c-4a4c-4f9c-95d2-5dcedd19d362\") " pod="openstack/ovn-controller-8bgjb-config-bbncl" Jan 26 17:06:42 crc kubenswrapper[4754]: I0126 17:06:42.129336 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qw5wq\" (UniqueName: \"kubernetes.io/projected/4146f13c-4a4c-4f9c-95d2-5dcedd19d362-kube-api-access-qw5wq\") pod \"ovn-controller-8bgjb-config-bbncl\" (UID: \"4146f13c-4a4c-4f9c-95d2-5dcedd19d362\") " pod="openstack/ovn-controller-8bgjb-config-bbncl" Jan 26 17:06:42 crc kubenswrapper[4754]: I0126 17:06:42.207577 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-8bgjb-config-bbncl" Jan 26 17:06:46 crc kubenswrapper[4754]: I0126 17:06:46.622631 4754 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-8bgjb" podUID="6db16a49-6566-42bf-91ad-c34be46e7800" containerName="ovn-controller" probeResult="failure" output=< Jan 26 17:06:46 crc kubenswrapper[4754]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Jan 26 17:06:46 crc kubenswrapper[4754]: > Jan 26 17:06:47 crc kubenswrapper[4754]: I0126 17:06:47.431438 4754 generic.go:334] "Generic (PLEG): container finished" podID="48810c5e-3c40-4cdc-8bab-47efa97e76fa" containerID="397f275bbe0a6c7fe3d67ba536948ee64336fe00addd514ba132a01fca3d4cb6" exitCode=0 Jan 26 17:06:47 crc kubenswrapper[4754]: I0126 17:06:47.431538 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"48810c5e-3c40-4cdc-8bab-47efa97e76fa","Type":"ContainerDied","Data":"397f275bbe0a6c7fe3d67ba536948ee64336fe00addd514ba132a01fca3d4cb6"} Jan 26 17:06:47 crc kubenswrapper[4754]: I0126 17:06:47.434014 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-gwx8r" event={"ID":"afc71d32-c836-4b6a-aae7-339d2582da56","Type":"ContainerDied","Data":"0b5b7af38a2a9623e055f41f8498b5feb4e6f4837a7f1471efc22524321c04b9"} Jan 26 17:06:47 crc kubenswrapper[4754]: I0126 17:06:47.434049 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0b5b7af38a2a9623e055f41f8498b5feb4e6f4837a7f1471efc22524321c04b9" Jan 26 17:06:47 crc kubenswrapper[4754]: I0126 17:06:47.437636 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-sl9kb" event={"ID":"cb88274a-904f-4827-9518-81b79a0e6a42","Type":"ContainerDied","Data":"7c406fb5a434d2616d1651f2bf56181f757b5fc67b31825b896ace5a92d83a88"} Jan 26 17:06:47 crc kubenswrapper[4754]: I0126 17:06:47.437680 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7c406fb5a434d2616d1651f2bf56181f757b5fc67b31825b896ace5a92d83a88" Jan 26 17:06:47 crc kubenswrapper[4754]: I0126 17:06:47.438790 4754 generic.go:334] "Generic (PLEG): container finished" podID="29be5081-5097-4cd1-b35b-192df0ce8faf" containerID="3f67cb54c8ac7414b3b3fe5d6bc7d975fa28e160e0a72a1773a4721747eb4df3" exitCode=0 Jan 26 17:06:47 crc kubenswrapper[4754]: I0126 17:06:47.438833 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"29be5081-5097-4cd1-b35b-192df0ce8faf","Type":"ContainerDied","Data":"3f67cb54c8ac7414b3b3fe5d6bc7d975fa28e160e0a72a1773a4721747eb4df3"} Jan 26 17:06:47 crc kubenswrapper[4754]: I0126 17:06:47.579176 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-sl9kb" Jan 26 17:06:47 crc kubenswrapper[4754]: I0126 17:06:47.622123 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-gwx8r" Jan 26 17:06:47 crc kubenswrapper[4754]: I0126 17:06:47.703032 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/cb88274a-904f-4827-9518-81b79a0e6a42-dispersionconf\") pod \"cb88274a-904f-4827-9518-81b79a0e6a42\" (UID: \"cb88274a-904f-4827-9518-81b79a0e6a42\") " Jan 26 17:06:47 crc kubenswrapper[4754]: I0126 17:06:47.703075 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/cb88274a-904f-4827-9518-81b79a0e6a42-swiftconf\") pod \"cb88274a-904f-4827-9518-81b79a0e6a42\" (UID: \"cb88274a-904f-4827-9518-81b79a0e6a42\") " Jan 26 17:06:47 crc kubenswrapper[4754]: I0126 17:06:47.703099 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q954r\" (UniqueName: \"kubernetes.io/projected/afc71d32-c836-4b6a-aae7-339d2582da56-kube-api-access-q954r\") pod \"afc71d32-c836-4b6a-aae7-339d2582da56\" (UID: \"afc71d32-c836-4b6a-aae7-339d2582da56\") " Jan 26 17:06:47 crc kubenswrapper[4754]: I0126 17:06:47.703134 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/cb88274a-904f-4827-9518-81b79a0e6a42-etc-swift\") pod \"cb88274a-904f-4827-9518-81b79a0e6a42\" (UID: \"cb88274a-904f-4827-9518-81b79a0e6a42\") " Jan 26 17:06:47 crc kubenswrapper[4754]: I0126 17:06:47.703170 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/afc71d32-c836-4b6a-aae7-339d2582da56-operator-scripts\") pod \"afc71d32-c836-4b6a-aae7-339d2582da56\" (UID: \"afc71d32-c836-4b6a-aae7-339d2582da56\") " Jan 26 17:06:47 crc kubenswrapper[4754]: I0126 17:06:47.703199 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mqx8z\" (UniqueName: \"kubernetes.io/projected/cb88274a-904f-4827-9518-81b79a0e6a42-kube-api-access-mqx8z\") pod \"cb88274a-904f-4827-9518-81b79a0e6a42\" (UID: \"cb88274a-904f-4827-9518-81b79a0e6a42\") " Jan 26 17:06:47 crc kubenswrapper[4754]: I0126 17:06:47.703298 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/cb88274a-904f-4827-9518-81b79a0e6a42-ring-data-devices\") pod \"cb88274a-904f-4827-9518-81b79a0e6a42\" (UID: \"cb88274a-904f-4827-9518-81b79a0e6a42\") " Jan 26 17:06:47 crc kubenswrapper[4754]: I0126 17:06:47.703312 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cb88274a-904f-4827-9518-81b79a0e6a42-scripts\") pod \"cb88274a-904f-4827-9518-81b79a0e6a42\" (UID: \"cb88274a-904f-4827-9518-81b79a0e6a42\") " Jan 26 17:06:47 crc kubenswrapper[4754]: I0126 17:06:47.703340 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb88274a-904f-4827-9518-81b79a0e6a42-combined-ca-bundle\") pod \"cb88274a-904f-4827-9518-81b79a0e6a42\" (UID: \"cb88274a-904f-4827-9518-81b79a0e6a42\") " Jan 26 17:06:47 crc kubenswrapper[4754]: I0126 17:06:47.704072 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/afc71d32-c836-4b6a-aae7-339d2582da56-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "afc71d32-c836-4b6a-aae7-339d2582da56" (UID: "afc71d32-c836-4b6a-aae7-339d2582da56"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:06:47 crc kubenswrapper[4754]: I0126 17:06:47.704137 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cb88274a-904f-4827-9518-81b79a0e6a42-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "cb88274a-904f-4827-9518-81b79a0e6a42" (UID: "cb88274a-904f-4827-9518-81b79a0e6a42"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:06:47 crc kubenswrapper[4754]: I0126 17:06:47.705898 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cb88274a-904f-4827-9518-81b79a0e6a42-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "cb88274a-904f-4827-9518-81b79a0e6a42" (UID: "cb88274a-904f-4827-9518-81b79a0e6a42"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:06:47 crc kubenswrapper[4754]: I0126 17:06:47.711509 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/afc71d32-c836-4b6a-aae7-339d2582da56-kube-api-access-q954r" (OuterVolumeSpecName: "kube-api-access-q954r") pod "afc71d32-c836-4b6a-aae7-339d2582da56" (UID: "afc71d32-c836-4b6a-aae7-339d2582da56"). InnerVolumeSpecName "kube-api-access-q954r". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:06:47 crc kubenswrapper[4754]: I0126 17:06:47.711808 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cb88274a-904f-4827-9518-81b79a0e6a42-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "cb88274a-904f-4827-9518-81b79a0e6a42" (UID: "cb88274a-904f-4827-9518-81b79a0e6a42"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:06:47 crc kubenswrapper[4754]: I0126 17:06:47.715378 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cb88274a-904f-4827-9518-81b79a0e6a42-kube-api-access-mqx8z" (OuterVolumeSpecName: "kube-api-access-mqx8z") pod "cb88274a-904f-4827-9518-81b79a0e6a42" (UID: "cb88274a-904f-4827-9518-81b79a0e6a42"). InnerVolumeSpecName "kube-api-access-mqx8z". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:06:47 crc kubenswrapper[4754]: I0126 17:06:47.722534 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cb88274a-904f-4827-9518-81b79a0e6a42-scripts" (OuterVolumeSpecName: "scripts") pod "cb88274a-904f-4827-9518-81b79a0e6a42" (UID: "cb88274a-904f-4827-9518-81b79a0e6a42"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:06:47 crc kubenswrapper[4754]: I0126 17:06:47.730497 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cb88274a-904f-4827-9518-81b79a0e6a42-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "cb88274a-904f-4827-9518-81b79a0e6a42" (UID: "cb88274a-904f-4827-9518-81b79a0e6a42"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:06:47 crc kubenswrapper[4754]: I0126 17:06:47.734731 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cb88274a-904f-4827-9518-81b79a0e6a42-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cb88274a-904f-4827-9518-81b79a0e6a42" (UID: "cb88274a-904f-4827-9518-81b79a0e6a42"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:06:47 crc kubenswrapper[4754]: I0126 17:06:47.805554 4754 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb88274a-904f-4827-9518-81b79a0e6a42-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 26 17:06:47 crc kubenswrapper[4754]: I0126 17:06:47.805592 4754 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/cb88274a-904f-4827-9518-81b79a0e6a42-dispersionconf\") on node \"crc\" DevicePath \"\"" Jan 26 17:06:47 crc kubenswrapper[4754]: I0126 17:06:47.805600 4754 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/cb88274a-904f-4827-9518-81b79a0e6a42-swiftconf\") on node \"crc\" DevicePath \"\"" Jan 26 17:06:47 crc kubenswrapper[4754]: I0126 17:06:47.805610 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q954r\" (UniqueName: \"kubernetes.io/projected/afc71d32-c836-4b6a-aae7-339d2582da56-kube-api-access-q954r\") on node \"crc\" DevicePath \"\"" Jan 26 17:06:47 crc kubenswrapper[4754]: I0126 17:06:47.805620 4754 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/cb88274a-904f-4827-9518-81b79a0e6a42-etc-swift\") on node \"crc\" DevicePath \"\"" Jan 26 17:06:47 crc kubenswrapper[4754]: I0126 17:06:47.805629 4754 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/afc71d32-c836-4b6a-aae7-339d2582da56-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 26 17:06:47 crc kubenswrapper[4754]: I0126 17:06:47.805637 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mqx8z\" (UniqueName: \"kubernetes.io/projected/cb88274a-904f-4827-9518-81b79a0e6a42-kube-api-access-mqx8z\") on node \"crc\" DevicePath \"\"" Jan 26 17:06:47 crc kubenswrapper[4754]: I0126 17:06:47.805649 4754 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/cb88274a-904f-4827-9518-81b79a0e6a42-ring-data-devices\") on node \"crc\" DevicePath \"\"" Jan 26 17:06:47 crc kubenswrapper[4754]: I0126 17:06:47.805657 4754 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cb88274a-904f-4827-9518-81b79a0e6a42-scripts\") on node \"crc\" DevicePath \"\"" Jan 26 17:06:47 crc kubenswrapper[4754]: I0126 17:06:47.891115 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-8bgjb-config-bbncl"] Jan 26 17:06:47 crc kubenswrapper[4754]: W0126 17:06:47.894822 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4146f13c_4a4c_4f9c_95d2_5dcedd19d362.slice/crio-a5ae207512f923c6f06d2dcdbb90e3e0c1649cad3808b2749490b609326ae2d8 WatchSource:0}: Error finding container a5ae207512f923c6f06d2dcdbb90e3e0c1649cad3808b2749490b609326ae2d8: Status 404 returned error can't find the container with id a5ae207512f923c6f06d2dcdbb90e3e0c1649cad3808b2749490b609326ae2d8 Jan 26 17:06:48 crc kubenswrapper[4754]: I0126 17:06:48.446798 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-8bgjb-config-bbncl" event={"ID":"4146f13c-4a4c-4f9c-95d2-5dcedd19d362","Type":"ContainerStarted","Data":"a5ae207512f923c6f06d2dcdbb90e3e0c1649cad3808b2749490b609326ae2d8"} Jan 26 17:06:48 crc kubenswrapper[4754]: I0126 17:06:48.446811 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-sl9kb" Jan 26 17:06:48 crc kubenswrapper[4754]: I0126 17:06:48.446823 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-gwx8r" Jan 26 17:06:49 crc kubenswrapper[4754]: I0126 17:06:49.404973 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/root-account-create-update-gwx8r"] Jan 26 17:06:49 crc kubenswrapper[4754]: I0126 17:06:49.424913 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/root-account-create-update-gwx8r"] Jan 26 17:06:49 crc kubenswrapper[4754]: I0126 17:06:49.467120 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-qbqp5" event={"ID":"bae8ecc7-8fac-4a63-9e99-4e7543eae5af","Type":"ContainerStarted","Data":"e330185b9b6205d6f4d57fde1a2bb14997226cc270954bfc37385ba8ca09676b"} Jan 26 17:06:49 crc kubenswrapper[4754]: I0126 17:06:49.469780 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-8bgjb-config-bbncl" event={"ID":"4146f13c-4a4c-4f9c-95d2-5dcedd19d362","Type":"ContainerStarted","Data":"316130fcccc0bf944a3b73afaa9f2d0f8d0d8fa2c0ccc61283a5ae24c9d6366c"} Jan 26 17:06:49 crc kubenswrapper[4754]: I0126 17:06:49.471826 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"29be5081-5097-4cd1-b35b-192df0ce8faf","Type":"ContainerStarted","Data":"0b6e796b8cc0ee4cab5cb13a889864aa381f1854cef6c88cc1c854144900c02a"} Jan 26 17:06:49 crc kubenswrapper[4754]: I0126 17:06:49.472246 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Jan 26 17:06:49 crc kubenswrapper[4754]: I0126 17:06:49.484797 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"48810c5e-3c40-4cdc-8bab-47efa97e76fa","Type":"ContainerStarted","Data":"c74f5c36b9d479df65eb065e767730beafd576f0ba47ddcc2bc9a3d7da87b416"} Jan 26 17:06:49 crc kubenswrapper[4754]: I0126 17:06:49.485133 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Jan 26 17:06:49 crc kubenswrapper[4754]: I0126 17:06:49.497532 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" event={"ID":"8c3718a4-f354-4284-92e0-fdfb45a692bd","Type":"ContainerStarted","Data":"7bdda0a504a055c787cfe163a27b6598986e7e593ba6562f6669a36ea09cda24"} Jan 26 17:06:49 crc kubenswrapper[4754]: I0126 17:06:49.500424 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-qbqp5" podStartSLOduration=3.411010774 podStartE2EDuration="18.500400407s" podCreationTimestamp="2026-01-26 17:06:31 +0000 UTC" firstStartedPulling="2026-01-26 17:06:32.436378405 +0000 UTC m=+1158.960558839" lastFinishedPulling="2026-01-26 17:06:47.525768038 +0000 UTC m=+1174.049948472" observedRunningTime="2026-01-26 17:06:49.487526847 +0000 UTC m=+1176.011707281" watchObservedRunningTime="2026-01-26 17:06:49.500400407 +0000 UTC m=+1176.024580841" Jan 26 17:06:49 crc kubenswrapper[4754]: I0126 17:06:49.548899 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-8bgjb-config-bbncl" podStartSLOduration=8.548878462 podStartE2EDuration="8.548878462s" podCreationTimestamp="2026-01-26 17:06:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 17:06:49.548870591 +0000 UTC m=+1176.073051025" watchObservedRunningTime="2026-01-26 17:06:49.548878462 +0000 UTC m=+1176.073058896" Jan 26 17:06:49 crc kubenswrapper[4754]: I0126 17:06:49.556609 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=-9223371963.298185 podStartE2EDuration="1m13.556590837s" podCreationTimestamp="2026-01-26 17:05:36 +0000 UTC" firstStartedPulling="2026-01-26 17:05:38.632975856 +0000 UTC m=+1105.157156290" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 17:06:49.523933405 +0000 UTC m=+1176.048113839" watchObservedRunningTime="2026-01-26 17:06:49.556590837 +0000 UTC m=+1176.080771261" Jan 26 17:06:49 crc kubenswrapper[4754]: I0126 17:06:49.593400 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=39.816435186 podStartE2EDuration="1m13.593383175s" podCreationTimestamp="2026-01-26 17:05:36 +0000 UTC" firstStartedPulling="2026-01-26 17:05:38.546394567 +0000 UTC m=+1105.070575001" lastFinishedPulling="2026-01-26 17:06:12.323342556 +0000 UTC m=+1138.847522990" observedRunningTime="2026-01-26 17:06:49.578037926 +0000 UTC m=+1176.102218370" watchObservedRunningTime="2026-01-26 17:06:49.593383175 +0000 UTC m=+1176.117563609" Jan 26 17:06:49 crc kubenswrapper[4754]: I0126 17:06:49.777041 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="afc71d32-c836-4b6a-aae7-339d2582da56" path="/var/lib/kubelet/pods/afc71d32-c836-4b6a-aae7-339d2582da56/volumes" Jan 26 17:06:50 crc kubenswrapper[4754]: I0126 17:06:50.506159 4754 generic.go:334] "Generic (PLEG): container finished" podID="4146f13c-4a4c-4f9c-95d2-5dcedd19d362" containerID="316130fcccc0bf944a3b73afaa9f2d0f8d0d8fa2c0ccc61283a5ae24c9d6366c" exitCode=0 Jan 26 17:06:50 crc kubenswrapper[4754]: I0126 17:06:50.506257 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-8bgjb-config-bbncl" event={"ID":"4146f13c-4a4c-4f9c-95d2-5dcedd19d362","Type":"ContainerDied","Data":"316130fcccc0bf944a3b73afaa9f2d0f8d0d8fa2c0ccc61283a5ae24c9d6366c"} Jan 26 17:06:51 crc kubenswrapper[4754]: I0126 17:06:51.611892 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-8bgjb" Jan 26 17:06:51 crc kubenswrapper[4754]: I0126 17:06:51.861198 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-8bgjb-config-bbncl" Jan 26 17:06:51 crc kubenswrapper[4754]: I0126 17:06:51.983421 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/4146f13c-4a4c-4f9c-95d2-5dcedd19d362-var-log-ovn\") pod \"4146f13c-4a4c-4f9c-95d2-5dcedd19d362\" (UID: \"4146f13c-4a4c-4f9c-95d2-5dcedd19d362\") " Jan 26 17:06:51 crc kubenswrapper[4754]: I0126 17:06:51.983559 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4146f13c-4a4c-4f9c-95d2-5dcedd19d362-scripts\") pod \"4146f13c-4a4c-4f9c-95d2-5dcedd19d362\" (UID: \"4146f13c-4a4c-4f9c-95d2-5dcedd19d362\") " Jan 26 17:06:51 crc kubenswrapper[4754]: I0126 17:06:51.983604 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/4146f13c-4a4c-4f9c-95d2-5dcedd19d362-var-run\") pod \"4146f13c-4a4c-4f9c-95d2-5dcedd19d362\" (UID: \"4146f13c-4a4c-4f9c-95d2-5dcedd19d362\") " Jan 26 17:06:51 crc kubenswrapper[4754]: I0126 17:06:51.983649 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qw5wq\" (UniqueName: \"kubernetes.io/projected/4146f13c-4a4c-4f9c-95d2-5dcedd19d362-kube-api-access-qw5wq\") pod \"4146f13c-4a4c-4f9c-95d2-5dcedd19d362\" (UID: \"4146f13c-4a4c-4f9c-95d2-5dcedd19d362\") " Jan 26 17:06:51 crc kubenswrapper[4754]: I0126 17:06:51.983633 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4146f13c-4a4c-4f9c-95d2-5dcedd19d362-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "4146f13c-4a4c-4f9c-95d2-5dcedd19d362" (UID: "4146f13c-4a4c-4f9c-95d2-5dcedd19d362"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 17:06:51 crc kubenswrapper[4754]: I0126 17:06:51.983792 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/4146f13c-4a4c-4f9c-95d2-5dcedd19d362-var-run-ovn\") pod \"4146f13c-4a4c-4f9c-95d2-5dcedd19d362\" (UID: \"4146f13c-4a4c-4f9c-95d2-5dcedd19d362\") " Jan 26 17:06:51 crc kubenswrapper[4754]: I0126 17:06:51.983754 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4146f13c-4a4c-4f9c-95d2-5dcedd19d362-var-run" (OuterVolumeSpecName: "var-run") pod "4146f13c-4a4c-4f9c-95d2-5dcedd19d362" (UID: "4146f13c-4a4c-4f9c-95d2-5dcedd19d362"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 17:06:51 crc kubenswrapper[4754]: I0126 17:06:51.983909 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/4146f13c-4a4c-4f9c-95d2-5dcedd19d362-additional-scripts\") pod \"4146f13c-4a4c-4f9c-95d2-5dcedd19d362\" (UID: \"4146f13c-4a4c-4f9c-95d2-5dcedd19d362\") " Jan 26 17:06:51 crc kubenswrapper[4754]: I0126 17:06:51.983814 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4146f13c-4a4c-4f9c-95d2-5dcedd19d362-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "4146f13c-4a4c-4f9c-95d2-5dcedd19d362" (UID: "4146f13c-4a4c-4f9c-95d2-5dcedd19d362"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 17:06:51 crc kubenswrapper[4754]: I0126 17:06:51.984322 4754 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/4146f13c-4a4c-4f9c-95d2-5dcedd19d362-var-log-ovn\") on node \"crc\" DevicePath \"\"" Jan 26 17:06:51 crc kubenswrapper[4754]: I0126 17:06:51.984342 4754 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/4146f13c-4a4c-4f9c-95d2-5dcedd19d362-var-run\") on node \"crc\" DevicePath \"\"" Jan 26 17:06:51 crc kubenswrapper[4754]: I0126 17:06:51.984354 4754 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/4146f13c-4a4c-4f9c-95d2-5dcedd19d362-var-run-ovn\") on node \"crc\" DevicePath \"\"" Jan 26 17:06:51 crc kubenswrapper[4754]: I0126 17:06:51.984549 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4146f13c-4a4c-4f9c-95d2-5dcedd19d362-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "4146f13c-4a4c-4f9c-95d2-5dcedd19d362" (UID: "4146f13c-4a4c-4f9c-95d2-5dcedd19d362"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:06:51 crc kubenswrapper[4754]: I0126 17:06:51.984741 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4146f13c-4a4c-4f9c-95d2-5dcedd19d362-scripts" (OuterVolumeSpecName: "scripts") pod "4146f13c-4a4c-4f9c-95d2-5dcedd19d362" (UID: "4146f13c-4a4c-4f9c-95d2-5dcedd19d362"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:06:51 crc kubenswrapper[4754]: I0126 17:06:51.989896 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4146f13c-4a4c-4f9c-95d2-5dcedd19d362-kube-api-access-qw5wq" (OuterVolumeSpecName: "kube-api-access-qw5wq") pod "4146f13c-4a4c-4f9c-95d2-5dcedd19d362" (UID: "4146f13c-4a4c-4f9c-95d2-5dcedd19d362"). InnerVolumeSpecName "kube-api-access-qw5wq". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:06:52 crc kubenswrapper[4754]: I0126 17:06:52.086138 4754 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4146f13c-4a4c-4f9c-95d2-5dcedd19d362-scripts\") on node \"crc\" DevicePath \"\"" Jan 26 17:06:52 crc kubenswrapper[4754]: I0126 17:06:52.086164 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qw5wq\" (UniqueName: \"kubernetes.io/projected/4146f13c-4a4c-4f9c-95d2-5dcedd19d362-kube-api-access-qw5wq\") on node \"crc\" DevicePath \"\"" Jan 26 17:06:52 crc kubenswrapper[4754]: I0126 17:06:52.086177 4754 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/4146f13c-4a4c-4f9c-95d2-5dcedd19d362-additional-scripts\") on node \"crc\" DevicePath \"\"" Jan 26 17:06:52 crc kubenswrapper[4754]: I0126 17:06:52.529833 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-8bgjb-config-bbncl" event={"ID":"4146f13c-4a4c-4f9c-95d2-5dcedd19d362","Type":"ContainerDied","Data":"a5ae207512f923c6f06d2dcdbb90e3e0c1649cad3808b2749490b609326ae2d8"} Jan 26 17:06:52 crc kubenswrapper[4754]: I0126 17:06:52.530479 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a5ae207512f923c6f06d2dcdbb90e3e0c1649cad3808b2749490b609326ae2d8" Jan 26 17:06:52 crc kubenswrapper[4754]: I0126 17:06:52.529912 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-8bgjb-config-bbncl" Jan 26 17:06:52 crc kubenswrapper[4754]: I0126 17:06:52.944832 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-8bgjb-config-bbncl"] Jan 26 17:06:52 crc kubenswrapper[4754]: I0126 17:06:52.950584 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-8bgjb-config-bbncl"] Jan 26 17:06:53 crc kubenswrapper[4754]: I0126 17:06:53.015615 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-8bgjb-config-hk6pv"] Jan 26 17:06:53 crc kubenswrapper[4754]: E0126 17:06:53.016055 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="afc71d32-c836-4b6a-aae7-339d2582da56" containerName="mariadb-account-create-update" Jan 26 17:06:53 crc kubenswrapper[4754]: I0126 17:06:53.016077 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="afc71d32-c836-4b6a-aae7-339d2582da56" containerName="mariadb-account-create-update" Jan 26 17:06:53 crc kubenswrapper[4754]: E0126 17:06:53.016100 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb88274a-904f-4827-9518-81b79a0e6a42" containerName="swift-ring-rebalance" Jan 26 17:06:53 crc kubenswrapper[4754]: I0126 17:06:53.016108 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb88274a-904f-4827-9518-81b79a0e6a42" containerName="swift-ring-rebalance" Jan 26 17:06:53 crc kubenswrapper[4754]: E0126 17:06:53.016121 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4146f13c-4a4c-4f9c-95d2-5dcedd19d362" containerName="ovn-config" Jan 26 17:06:53 crc kubenswrapper[4754]: I0126 17:06:53.016128 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="4146f13c-4a4c-4f9c-95d2-5dcedd19d362" containerName="ovn-config" Jan 26 17:06:53 crc kubenswrapper[4754]: I0126 17:06:53.016296 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="afc71d32-c836-4b6a-aae7-339d2582da56" containerName="mariadb-account-create-update" Jan 26 17:06:53 crc kubenswrapper[4754]: I0126 17:06:53.016311 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="cb88274a-904f-4827-9518-81b79a0e6a42" containerName="swift-ring-rebalance" Jan 26 17:06:53 crc kubenswrapper[4754]: I0126 17:06:53.016441 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="4146f13c-4a4c-4f9c-95d2-5dcedd19d362" containerName="ovn-config" Jan 26 17:06:53 crc kubenswrapper[4754]: I0126 17:06:53.017019 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-8bgjb-config-hk6pv" Jan 26 17:06:53 crc kubenswrapper[4754]: I0126 17:06:53.020657 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Jan 26 17:06:53 crc kubenswrapper[4754]: I0126 17:06:53.031362 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-8bgjb-config-hk6pv"] Jan 26 17:06:53 crc kubenswrapper[4754]: I0126 17:06:53.056133 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/root-account-create-update-wjjd9"] Jan 26 17:06:53 crc kubenswrapper[4754]: I0126 17:06:53.058389 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-wjjd9" Jan 26 17:06:53 crc kubenswrapper[4754]: I0126 17:06:53.061754 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-mariadb-root-db-secret" Jan 26 17:06:53 crc kubenswrapper[4754]: I0126 17:06:53.078573 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-wjjd9"] Jan 26 17:06:53 crc kubenswrapper[4754]: I0126 17:06:53.104121 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2vznw\" (UniqueName: \"kubernetes.io/projected/e71ee5fd-3be8-49ff-96cd-3f5e67a55bad-kube-api-access-2vznw\") pod \"ovn-controller-8bgjb-config-hk6pv\" (UID: \"e71ee5fd-3be8-49ff-96cd-3f5e67a55bad\") " pod="openstack/ovn-controller-8bgjb-config-hk6pv" Jan 26 17:06:53 crc kubenswrapper[4754]: I0126 17:06:53.104201 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/e71ee5fd-3be8-49ff-96cd-3f5e67a55bad-var-run-ovn\") pod \"ovn-controller-8bgjb-config-hk6pv\" (UID: \"e71ee5fd-3be8-49ff-96cd-3f5e67a55bad\") " pod="openstack/ovn-controller-8bgjb-config-hk6pv" Jan 26 17:06:53 crc kubenswrapper[4754]: I0126 17:06:53.104249 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/e71ee5fd-3be8-49ff-96cd-3f5e67a55bad-var-run\") pod \"ovn-controller-8bgjb-config-hk6pv\" (UID: \"e71ee5fd-3be8-49ff-96cd-3f5e67a55bad\") " pod="openstack/ovn-controller-8bgjb-config-hk6pv" Jan 26 17:06:53 crc kubenswrapper[4754]: I0126 17:06:53.104280 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4qk8l\" (UniqueName: \"kubernetes.io/projected/05f7ca48-d18a-42cf-921f-43c3fc4111ff-kube-api-access-4qk8l\") pod \"root-account-create-update-wjjd9\" (UID: \"05f7ca48-d18a-42cf-921f-43c3fc4111ff\") " pod="openstack/root-account-create-update-wjjd9" Jan 26 17:06:53 crc kubenswrapper[4754]: I0126 17:06:53.104333 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e71ee5fd-3be8-49ff-96cd-3f5e67a55bad-scripts\") pod \"ovn-controller-8bgjb-config-hk6pv\" (UID: \"e71ee5fd-3be8-49ff-96cd-3f5e67a55bad\") " pod="openstack/ovn-controller-8bgjb-config-hk6pv" Jan 26 17:06:53 crc kubenswrapper[4754]: I0126 17:06:53.104358 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/e71ee5fd-3be8-49ff-96cd-3f5e67a55bad-var-log-ovn\") pod \"ovn-controller-8bgjb-config-hk6pv\" (UID: \"e71ee5fd-3be8-49ff-96cd-3f5e67a55bad\") " pod="openstack/ovn-controller-8bgjb-config-hk6pv" Jan 26 17:06:53 crc kubenswrapper[4754]: I0126 17:06:53.104391 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/e71ee5fd-3be8-49ff-96cd-3f5e67a55bad-additional-scripts\") pod \"ovn-controller-8bgjb-config-hk6pv\" (UID: \"e71ee5fd-3be8-49ff-96cd-3f5e67a55bad\") " pod="openstack/ovn-controller-8bgjb-config-hk6pv" Jan 26 17:06:53 crc kubenswrapper[4754]: I0126 17:06:53.104416 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/05f7ca48-d18a-42cf-921f-43c3fc4111ff-operator-scripts\") pod \"root-account-create-update-wjjd9\" (UID: \"05f7ca48-d18a-42cf-921f-43c3fc4111ff\") " pod="openstack/root-account-create-update-wjjd9" Jan 26 17:06:53 crc kubenswrapper[4754]: I0126 17:06:53.205916 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2vznw\" (UniqueName: \"kubernetes.io/projected/e71ee5fd-3be8-49ff-96cd-3f5e67a55bad-kube-api-access-2vznw\") pod \"ovn-controller-8bgjb-config-hk6pv\" (UID: \"e71ee5fd-3be8-49ff-96cd-3f5e67a55bad\") " pod="openstack/ovn-controller-8bgjb-config-hk6pv" Jan 26 17:06:53 crc kubenswrapper[4754]: I0126 17:06:53.206000 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/e71ee5fd-3be8-49ff-96cd-3f5e67a55bad-var-run-ovn\") pod \"ovn-controller-8bgjb-config-hk6pv\" (UID: \"e71ee5fd-3be8-49ff-96cd-3f5e67a55bad\") " pod="openstack/ovn-controller-8bgjb-config-hk6pv" Jan 26 17:06:53 crc kubenswrapper[4754]: I0126 17:06:53.206050 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/e71ee5fd-3be8-49ff-96cd-3f5e67a55bad-var-run\") pod \"ovn-controller-8bgjb-config-hk6pv\" (UID: \"e71ee5fd-3be8-49ff-96cd-3f5e67a55bad\") " pod="openstack/ovn-controller-8bgjb-config-hk6pv" Jan 26 17:06:53 crc kubenswrapper[4754]: I0126 17:06:53.206096 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4qk8l\" (UniqueName: \"kubernetes.io/projected/05f7ca48-d18a-42cf-921f-43c3fc4111ff-kube-api-access-4qk8l\") pod \"root-account-create-update-wjjd9\" (UID: \"05f7ca48-d18a-42cf-921f-43c3fc4111ff\") " pod="openstack/root-account-create-update-wjjd9" Jan 26 17:06:53 crc kubenswrapper[4754]: I0126 17:06:53.206153 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e71ee5fd-3be8-49ff-96cd-3f5e67a55bad-scripts\") pod \"ovn-controller-8bgjb-config-hk6pv\" (UID: \"e71ee5fd-3be8-49ff-96cd-3f5e67a55bad\") " pod="openstack/ovn-controller-8bgjb-config-hk6pv" Jan 26 17:06:53 crc kubenswrapper[4754]: I0126 17:06:53.206189 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/e71ee5fd-3be8-49ff-96cd-3f5e67a55bad-var-log-ovn\") pod \"ovn-controller-8bgjb-config-hk6pv\" (UID: \"e71ee5fd-3be8-49ff-96cd-3f5e67a55bad\") " pod="openstack/ovn-controller-8bgjb-config-hk6pv" Jan 26 17:06:53 crc kubenswrapper[4754]: I0126 17:06:53.206215 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/e71ee5fd-3be8-49ff-96cd-3f5e67a55bad-additional-scripts\") pod \"ovn-controller-8bgjb-config-hk6pv\" (UID: \"e71ee5fd-3be8-49ff-96cd-3f5e67a55bad\") " pod="openstack/ovn-controller-8bgjb-config-hk6pv" Jan 26 17:06:53 crc kubenswrapper[4754]: I0126 17:06:53.206250 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/05f7ca48-d18a-42cf-921f-43c3fc4111ff-operator-scripts\") pod \"root-account-create-update-wjjd9\" (UID: \"05f7ca48-d18a-42cf-921f-43c3fc4111ff\") " pod="openstack/root-account-create-update-wjjd9" Jan 26 17:06:53 crc kubenswrapper[4754]: I0126 17:06:53.206440 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/e71ee5fd-3be8-49ff-96cd-3f5e67a55bad-var-run\") pod \"ovn-controller-8bgjb-config-hk6pv\" (UID: \"e71ee5fd-3be8-49ff-96cd-3f5e67a55bad\") " pod="openstack/ovn-controller-8bgjb-config-hk6pv" Jan 26 17:06:53 crc kubenswrapper[4754]: I0126 17:06:53.206552 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/e71ee5fd-3be8-49ff-96cd-3f5e67a55bad-var-log-ovn\") pod \"ovn-controller-8bgjb-config-hk6pv\" (UID: \"e71ee5fd-3be8-49ff-96cd-3f5e67a55bad\") " pod="openstack/ovn-controller-8bgjb-config-hk6pv" Jan 26 17:06:53 crc kubenswrapper[4754]: I0126 17:06:53.206971 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/e71ee5fd-3be8-49ff-96cd-3f5e67a55bad-var-run-ovn\") pod \"ovn-controller-8bgjb-config-hk6pv\" (UID: \"e71ee5fd-3be8-49ff-96cd-3f5e67a55bad\") " pod="openstack/ovn-controller-8bgjb-config-hk6pv" Jan 26 17:06:53 crc kubenswrapper[4754]: I0126 17:06:53.207263 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/05f7ca48-d18a-42cf-921f-43c3fc4111ff-operator-scripts\") pod \"root-account-create-update-wjjd9\" (UID: \"05f7ca48-d18a-42cf-921f-43c3fc4111ff\") " pod="openstack/root-account-create-update-wjjd9" Jan 26 17:06:53 crc kubenswrapper[4754]: I0126 17:06:53.207389 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/e71ee5fd-3be8-49ff-96cd-3f5e67a55bad-additional-scripts\") pod \"ovn-controller-8bgjb-config-hk6pv\" (UID: \"e71ee5fd-3be8-49ff-96cd-3f5e67a55bad\") " pod="openstack/ovn-controller-8bgjb-config-hk6pv" Jan 26 17:06:53 crc kubenswrapper[4754]: I0126 17:06:53.208489 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e71ee5fd-3be8-49ff-96cd-3f5e67a55bad-scripts\") pod \"ovn-controller-8bgjb-config-hk6pv\" (UID: \"e71ee5fd-3be8-49ff-96cd-3f5e67a55bad\") " pod="openstack/ovn-controller-8bgjb-config-hk6pv" Jan 26 17:06:53 crc kubenswrapper[4754]: I0126 17:06:53.227130 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2vznw\" (UniqueName: \"kubernetes.io/projected/e71ee5fd-3be8-49ff-96cd-3f5e67a55bad-kube-api-access-2vznw\") pod \"ovn-controller-8bgjb-config-hk6pv\" (UID: \"e71ee5fd-3be8-49ff-96cd-3f5e67a55bad\") " pod="openstack/ovn-controller-8bgjb-config-hk6pv" Jan 26 17:06:53 crc kubenswrapper[4754]: I0126 17:06:53.227643 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4qk8l\" (UniqueName: \"kubernetes.io/projected/05f7ca48-d18a-42cf-921f-43c3fc4111ff-kube-api-access-4qk8l\") pod \"root-account-create-update-wjjd9\" (UID: \"05f7ca48-d18a-42cf-921f-43c3fc4111ff\") " pod="openstack/root-account-create-update-wjjd9" Jan 26 17:06:53 crc kubenswrapper[4754]: I0126 17:06:53.333624 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-8bgjb-config-hk6pv" Jan 26 17:06:53 crc kubenswrapper[4754]: I0126 17:06:53.378751 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-wjjd9" Jan 26 17:06:53 crc kubenswrapper[4754]: I0126 17:06:53.776346 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4146f13c-4a4c-4f9c-95d2-5dcedd19d362" path="/var/lib/kubelet/pods/4146f13c-4a4c-4f9c-95d2-5dcedd19d362/volumes" Jan 26 17:06:53 crc kubenswrapper[4754]: I0126 17:06:53.797900 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-8bgjb-config-hk6pv"] Jan 26 17:06:53 crc kubenswrapper[4754]: I0126 17:06:53.923634 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-wjjd9"] Jan 26 17:06:53 crc kubenswrapper[4754]: W0126 17:06:53.938790 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod05f7ca48_d18a_42cf_921f_43c3fc4111ff.slice/crio-b709c92b9fca849837b78494197fece43e924fd2d5ca9eb04c815516871577f4 WatchSource:0}: Error finding container b709c92b9fca849837b78494197fece43e924fd2d5ca9eb04c815516871577f4: Status 404 returned error can't find the container with id b709c92b9fca849837b78494197fece43e924fd2d5ca9eb04c815516871577f4 Jan 26 17:06:54 crc kubenswrapper[4754]: I0126 17:06:54.545472 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-8bgjb-config-hk6pv" event={"ID":"e71ee5fd-3be8-49ff-96cd-3f5e67a55bad","Type":"ContainerStarted","Data":"731e4cde8e4b02895e7472b4abf99a79bcdecc2d1f09279f1b0a24e3811126f1"} Jan 26 17:06:54 crc kubenswrapper[4754]: I0126 17:06:54.547506 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-wjjd9" event={"ID":"05f7ca48-d18a-42cf-921f-43c3fc4111ff","Type":"ContainerStarted","Data":"b709c92b9fca849837b78494197fece43e924fd2d5ca9eb04c815516871577f4"} Jan 26 17:06:55 crc kubenswrapper[4754]: I0126 17:06:55.578279 4754 generic.go:334] "Generic (PLEG): container finished" podID="e71ee5fd-3be8-49ff-96cd-3f5e67a55bad" containerID="fe88ff978302d35023844df531d072c3bf22651302bb9dd2dddce4e3e4df7ecc" exitCode=0 Jan 26 17:06:55 crc kubenswrapper[4754]: I0126 17:06:55.578402 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-8bgjb-config-hk6pv" event={"ID":"e71ee5fd-3be8-49ff-96cd-3f5e67a55bad","Type":"ContainerDied","Data":"fe88ff978302d35023844df531d072c3bf22651302bb9dd2dddce4e3e4df7ecc"} Jan 26 17:06:55 crc kubenswrapper[4754]: I0126 17:06:55.580259 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-wjjd9" event={"ID":"05f7ca48-d18a-42cf-921f-43c3fc4111ff","Type":"ContainerStarted","Data":"420afe55ba4cf912e47db77c3348ccb28bb43686b51ba463c7e761755d35e137"} Jan 26 17:06:55 crc kubenswrapper[4754]: I0126 17:06:55.613767 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/root-account-create-update-wjjd9" podStartSLOduration=2.613749709 podStartE2EDuration="2.613749709s" podCreationTimestamp="2026-01-26 17:06:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 17:06:55.610751955 +0000 UTC m=+1182.134932469" watchObservedRunningTime="2026-01-26 17:06:55.613749709 +0000 UTC m=+1182.137930143" Jan 26 17:06:56 crc kubenswrapper[4754]: I0126 17:06:56.363390 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/19a0730c-46d0-4029-a86b-812ba1664dcc-etc-swift\") pod \"swift-storage-0\" (UID: \"19a0730c-46d0-4029-a86b-812ba1664dcc\") " pod="openstack/swift-storage-0" Jan 26 17:06:56 crc kubenswrapper[4754]: I0126 17:06:56.381448 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/19a0730c-46d0-4029-a86b-812ba1664dcc-etc-swift\") pod \"swift-storage-0\" (UID: \"19a0730c-46d0-4029-a86b-812ba1664dcc\") " pod="openstack/swift-storage-0" Jan 26 17:06:56 crc kubenswrapper[4754]: I0126 17:06:56.589053 4754 generic.go:334] "Generic (PLEG): container finished" podID="05f7ca48-d18a-42cf-921f-43c3fc4111ff" containerID="420afe55ba4cf912e47db77c3348ccb28bb43686b51ba463c7e761755d35e137" exitCode=0 Jan 26 17:06:56 crc kubenswrapper[4754]: I0126 17:06:56.589162 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-wjjd9" event={"ID":"05f7ca48-d18a-42cf-921f-43c3fc4111ff","Type":"ContainerDied","Data":"420afe55ba4cf912e47db77c3348ccb28bb43686b51ba463c7e761755d35e137"} Jan 26 17:06:56 crc kubenswrapper[4754]: I0126 17:06:56.635849 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Jan 26 17:06:56 crc kubenswrapper[4754]: I0126 17:06:56.910339 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-8bgjb-config-hk6pv" Jan 26 17:06:56 crc kubenswrapper[4754]: I0126 17:06:56.973653 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2vznw\" (UniqueName: \"kubernetes.io/projected/e71ee5fd-3be8-49ff-96cd-3f5e67a55bad-kube-api-access-2vznw\") pod \"e71ee5fd-3be8-49ff-96cd-3f5e67a55bad\" (UID: \"e71ee5fd-3be8-49ff-96cd-3f5e67a55bad\") " Jan 26 17:06:56 crc kubenswrapper[4754]: I0126 17:06:56.973729 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e71ee5fd-3be8-49ff-96cd-3f5e67a55bad-scripts\") pod \"e71ee5fd-3be8-49ff-96cd-3f5e67a55bad\" (UID: \"e71ee5fd-3be8-49ff-96cd-3f5e67a55bad\") " Jan 26 17:06:56 crc kubenswrapper[4754]: I0126 17:06:56.973845 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/e71ee5fd-3be8-49ff-96cd-3f5e67a55bad-var-run\") pod \"e71ee5fd-3be8-49ff-96cd-3f5e67a55bad\" (UID: \"e71ee5fd-3be8-49ff-96cd-3f5e67a55bad\") " Jan 26 17:06:56 crc kubenswrapper[4754]: I0126 17:06:56.973976 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/e71ee5fd-3be8-49ff-96cd-3f5e67a55bad-var-log-ovn\") pod \"e71ee5fd-3be8-49ff-96cd-3f5e67a55bad\" (UID: \"e71ee5fd-3be8-49ff-96cd-3f5e67a55bad\") " Jan 26 17:06:56 crc kubenswrapper[4754]: I0126 17:06:56.974030 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/e71ee5fd-3be8-49ff-96cd-3f5e67a55bad-var-run-ovn\") pod \"e71ee5fd-3be8-49ff-96cd-3f5e67a55bad\" (UID: \"e71ee5fd-3be8-49ff-96cd-3f5e67a55bad\") " Jan 26 17:06:56 crc kubenswrapper[4754]: I0126 17:06:56.974021 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e71ee5fd-3be8-49ff-96cd-3f5e67a55bad-var-run" (OuterVolumeSpecName: "var-run") pod "e71ee5fd-3be8-49ff-96cd-3f5e67a55bad" (UID: "e71ee5fd-3be8-49ff-96cd-3f5e67a55bad"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 17:06:56 crc kubenswrapper[4754]: I0126 17:06:56.974078 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/e71ee5fd-3be8-49ff-96cd-3f5e67a55bad-additional-scripts\") pod \"e71ee5fd-3be8-49ff-96cd-3f5e67a55bad\" (UID: \"e71ee5fd-3be8-49ff-96cd-3f5e67a55bad\") " Jan 26 17:06:56 crc kubenswrapper[4754]: I0126 17:06:56.974088 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e71ee5fd-3be8-49ff-96cd-3f5e67a55bad-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "e71ee5fd-3be8-49ff-96cd-3f5e67a55bad" (UID: "e71ee5fd-3be8-49ff-96cd-3f5e67a55bad"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 17:06:56 crc kubenswrapper[4754]: I0126 17:06:56.974121 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e71ee5fd-3be8-49ff-96cd-3f5e67a55bad-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "e71ee5fd-3be8-49ff-96cd-3f5e67a55bad" (UID: "e71ee5fd-3be8-49ff-96cd-3f5e67a55bad"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 17:06:56 crc kubenswrapper[4754]: I0126 17:06:56.974409 4754 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/e71ee5fd-3be8-49ff-96cd-3f5e67a55bad-var-run\") on node \"crc\" DevicePath \"\"" Jan 26 17:06:56 crc kubenswrapper[4754]: I0126 17:06:56.974429 4754 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/e71ee5fd-3be8-49ff-96cd-3f5e67a55bad-var-log-ovn\") on node \"crc\" DevicePath \"\"" Jan 26 17:06:56 crc kubenswrapper[4754]: I0126 17:06:56.974438 4754 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/e71ee5fd-3be8-49ff-96cd-3f5e67a55bad-var-run-ovn\") on node \"crc\" DevicePath \"\"" Jan 26 17:06:56 crc kubenswrapper[4754]: I0126 17:06:56.975022 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e71ee5fd-3be8-49ff-96cd-3f5e67a55bad-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "e71ee5fd-3be8-49ff-96cd-3f5e67a55bad" (UID: "e71ee5fd-3be8-49ff-96cd-3f5e67a55bad"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:06:56 crc kubenswrapper[4754]: I0126 17:06:56.975266 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e71ee5fd-3be8-49ff-96cd-3f5e67a55bad-scripts" (OuterVolumeSpecName: "scripts") pod "e71ee5fd-3be8-49ff-96cd-3f5e67a55bad" (UID: "e71ee5fd-3be8-49ff-96cd-3f5e67a55bad"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:06:56 crc kubenswrapper[4754]: I0126 17:06:56.979823 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e71ee5fd-3be8-49ff-96cd-3f5e67a55bad-kube-api-access-2vznw" (OuterVolumeSpecName: "kube-api-access-2vznw") pod "e71ee5fd-3be8-49ff-96cd-3f5e67a55bad" (UID: "e71ee5fd-3be8-49ff-96cd-3f5e67a55bad"). InnerVolumeSpecName "kube-api-access-2vznw". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:06:57 crc kubenswrapper[4754]: I0126 17:06:57.076161 4754 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/e71ee5fd-3be8-49ff-96cd-3f5e67a55bad-additional-scripts\") on node \"crc\" DevicePath \"\"" Jan 26 17:06:57 crc kubenswrapper[4754]: I0126 17:06:57.076201 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2vznw\" (UniqueName: \"kubernetes.io/projected/e71ee5fd-3be8-49ff-96cd-3f5e67a55bad-kube-api-access-2vznw\") on node \"crc\" DevicePath \"\"" Jan 26 17:06:57 crc kubenswrapper[4754]: I0126 17:06:57.076212 4754 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e71ee5fd-3be8-49ff-96cd-3f5e67a55bad-scripts\") on node \"crc\" DevicePath \"\"" Jan 26 17:06:57 crc kubenswrapper[4754]: I0126 17:06:57.238751 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Jan 26 17:06:57 crc kubenswrapper[4754]: W0126 17:06:57.251908 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod19a0730c_46d0_4029_a86b_812ba1664dcc.slice/crio-70f9f598f784041417accd28f331f3df030ebf5e17c708f5382600f8509cd6d4 WatchSource:0}: Error finding container 70f9f598f784041417accd28f331f3df030ebf5e17c708f5382600f8509cd6d4: Status 404 returned error can't find the container with id 70f9f598f784041417accd28f331f3df030ebf5e17c708f5382600f8509cd6d4 Jan 26 17:06:57 crc kubenswrapper[4754]: I0126 17:06:57.602449 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-8bgjb-config-hk6pv" Jan 26 17:06:57 crc kubenswrapper[4754]: I0126 17:06:57.602488 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-8bgjb-config-hk6pv" event={"ID":"e71ee5fd-3be8-49ff-96cd-3f5e67a55bad","Type":"ContainerDied","Data":"731e4cde8e4b02895e7472b4abf99a79bcdecc2d1f09279f1b0a24e3811126f1"} Jan 26 17:06:57 crc kubenswrapper[4754]: I0126 17:06:57.602936 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="731e4cde8e4b02895e7472b4abf99a79bcdecc2d1f09279f1b0a24e3811126f1" Jan 26 17:06:57 crc kubenswrapper[4754]: I0126 17:06:57.603573 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"19a0730c-46d0-4029-a86b-812ba1664dcc","Type":"ContainerStarted","Data":"70f9f598f784041417accd28f331f3df030ebf5e17c708f5382600f8509cd6d4"} Jan 26 17:06:57 crc kubenswrapper[4754]: I0126 17:06:57.917768 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-wjjd9" Jan 26 17:06:57 crc kubenswrapper[4754]: I0126 17:06:57.977361 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-8bgjb-config-hk6pv"] Jan 26 17:06:57 crc kubenswrapper[4754]: I0126 17:06:57.990559 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4qk8l\" (UniqueName: \"kubernetes.io/projected/05f7ca48-d18a-42cf-921f-43c3fc4111ff-kube-api-access-4qk8l\") pod \"05f7ca48-d18a-42cf-921f-43c3fc4111ff\" (UID: \"05f7ca48-d18a-42cf-921f-43c3fc4111ff\") " Jan 26 17:06:57 crc kubenswrapper[4754]: I0126 17:06:57.990832 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/05f7ca48-d18a-42cf-921f-43c3fc4111ff-operator-scripts\") pod \"05f7ca48-d18a-42cf-921f-43c3fc4111ff\" (UID: \"05f7ca48-d18a-42cf-921f-43c3fc4111ff\") " Jan 26 17:06:57 crc kubenswrapper[4754]: I0126 17:06:57.991138 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-8bgjb-config-hk6pv"] Jan 26 17:06:57 crc kubenswrapper[4754]: I0126 17:06:57.991456 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/05f7ca48-d18a-42cf-921f-43c3fc4111ff-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "05f7ca48-d18a-42cf-921f-43c3fc4111ff" (UID: "05f7ca48-d18a-42cf-921f-43c3fc4111ff"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:06:58 crc kubenswrapper[4754]: I0126 17:06:57.995317 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/05f7ca48-d18a-42cf-921f-43c3fc4111ff-kube-api-access-4qk8l" (OuterVolumeSpecName: "kube-api-access-4qk8l") pod "05f7ca48-d18a-42cf-921f-43c3fc4111ff" (UID: "05f7ca48-d18a-42cf-921f-43c3fc4111ff"). InnerVolumeSpecName "kube-api-access-4qk8l". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:06:58 crc kubenswrapper[4754]: I0126 17:06:58.016189 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-8bgjb-config-hfs7f"] Jan 26 17:06:58 crc kubenswrapper[4754]: E0126 17:06:58.016536 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e71ee5fd-3be8-49ff-96cd-3f5e67a55bad" containerName="ovn-config" Jan 26 17:06:58 crc kubenswrapper[4754]: I0126 17:06:58.016558 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="e71ee5fd-3be8-49ff-96cd-3f5e67a55bad" containerName="ovn-config" Jan 26 17:06:58 crc kubenswrapper[4754]: E0126 17:06:58.016594 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="05f7ca48-d18a-42cf-921f-43c3fc4111ff" containerName="mariadb-account-create-update" Jan 26 17:06:58 crc kubenswrapper[4754]: I0126 17:06:58.016604 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="05f7ca48-d18a-42cf-921f-43c3fc4111ff" containerName="mariadb-account-create-update" Jan 26 17:06:58 crc kubenswrapper[4754]: I0126 17:06:58.016802 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="05f7ca48-d18a-42cf-921f-43c3fc4111ff" containerName="mariadb-account-create-update" Jan 26 17:06:58 crc kubenswrapper[4754]: I0126 17:06:58.016823 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="e71ee5fd-3be8-49ff-96cd-3f5e67a55bad" containerName="ovn-config" Jan 26 17:06:58 crc kubenswrapper[4754]: I0126 17:06:58.017872 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-8bgjb-config-hfs7f" Jan 26 17:06:58 crc kubenswrapper[4754]: I0126 17:06:58.020718 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Jan 26 17:06:58 crc kubenswrapper[4754]: I0126 17:06:58.026220 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-8bgjb-config-hfs7f"] Jan 26 17:06:58 crc kubenswrapper[4754]: I0126 17:06:58.092183 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kl9sg\" (UniqueName: \"kubernetes.io/projected/fdbb773a-337d-44f0-8f43-380cbc74bd43-kube-api-access-kl9sg\") pod \"ovn-controller-8bgjb-config-hfs7f\" (UID: \"fdbb773a-337d-44f0-8f43-380cbc74bd43\") " pod="openstack/ovn-controller-8bgjb-config-hfs7f" Jan 26 17:06:58 crc kubenswrapper[4754]: I0126 17:06:58.092286 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fdbb773a-337d-44f0-8f43-380cbc74bd43-scripts\") pod \"ovn-controller-8bgjb-config-hfs7f\" (UID: \"fdbb773a-337d-44f0-8f43-380cbc74bd43\") " pod="openstack/ovn-controller-8bgjb-config-hfs7f" Jan 26 17:06:58 crc kubenswrapper[4754]: I0126 17:06:58.092319 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/fdbb773a-337d-44f0-8f43-380cbc74bd43-additional-scripts\") pod \"ovn-controller-8bgjb-config-hfs7f\" (UID: \"fdbb773a-337d-44f0-8f43-380cbc74bd43\") " pod="openstack/ovn-controller-8bgjb-config-hfs7f" Jan 26 17:06:58 crc kubenswrapper[4754]: I0126 17:06:58.092428 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/fdbb773a-337d-44f0-8f43-380cbc74bd43-var-run\") pod \"ovn-controller-8bgjb-config-hfs7f\" (UID: \"fdbb773a-337d-44f0-8f43-380cbc74bd43\") " pod="openstack/ovn-controller-8bgjb-config-hfs7f" Jan 26 17:06:58 crc kubenswrapper[4754]: I0126 17:06:58.092625 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/fdbb773a-337d-44f0-8f43-380cbc74bd43-var-log-ovn\") pod \"ovn-controller-8bgjb-config-hfs7f\" (UID: \"fdbb773a-337d-44f0-8f43-380cbc74bd43\") " pod="openstack/ovn-controller-8bgjb-config-hfs7f" Jan 26 17:06:58 crc kubenswrapper[4754]: I0126 17:06:58.092769 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/fdbb773a-337d-44f0-8f43-380cbc74bd43-var-run-ovn\") pod \"ovn-controller-8bgjb-config-hfs7f\" (UID: \"fdbb773a-337d-44f0-8f43-380cbc74bd43\") " pod="openstack/ovn-controller-8bgjb-config-hfs7f" Jan 26 17:06:58 crc kubenswrapper[4754]: I0126 17:06:58.092921 4754 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/05f7ca48-d18a-42cf-921f-43c3fc4111ff-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 26 17:06:58 crc kubenswrapper[4754]: I0126 17:06:58.092938 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4qk8l\" (UniqueName: \"kubernetes.io/projected/05f7ca48-d18a-42cf-921f-43c3fc4111ff-kube-api-access-4qk8l\") on node \"crc\" DevicePath \"\"" Jan 26 17:06:58 crc kubenswrapper[4754]: I0126 17:06:58.194469 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fdbb773a-337d-44f0-8f43-380cbc74bd43-scripts\") pod \"ovn-controller-8bgjb-config-hfs7f\" (UID: \"fdbb773a-337d-44f0-8f43-380cbc74bd43\") " pod="openstack/ovn-controller-8bgjb-config-hfs7f" Jan 26 17:06:58 crc kubenswrapper[4754]: I0126 17:06:58.194524 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/fdbb773a-337d-44f0-8f43-380cbc74bd43-additional-scripts\") pod \"ovn-controller-8bgjb-config-hfs7f\" (UID: \"fdbb773a-337d-44f0-8f43-380cbc74bd43\") " pod="openstack/ovn-controller-8bgjb-config-hfs7f" Jan 26 17:06:58 crc kubenswrapper[4754]: I0126 17:06:58.194647 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/fdbb773a-337d-44f0-8f43-380cbc74bd43-var-run\") pod \"ovn-controller-8bgjb-config-hfs7f\" (UID: \"fdbb773a-337d-44f0-8f43-380cbc74bd43\") " pod="openstack/ovn-controller-8bgjb-config-hfs7f" Jan 26 17:06:58 crc kubenswrapper[4754]: I0126 17:06:58.194738 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/fdbb773a-337d-44f0-8f43-380cbc74bd43-var-log-ovn\") pod \"ovn-controller-8bgjb-config-hfs7f\" (UID: \"fdbb773a-337d-44f0-8f43-380cbc74bd43\") " pod="openstack/ovn-controller-8bgjb-config-hfs7f" Jan 26 17:06:58 crc kubenswrapper[4754]: I0126 17:06:58.194776 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/fdbb773a-337d-44f0-8f43-380cbc74bd43-var-run-ovn\") pod \"ovn-controller-8bgjb-config-hfs7f\" (UID: \"fdbb773a-337d-44f0-8f43-380cbc74bd43\") " pod="openstack/ovn-controller-8bgjb-config-hfs7f" Jan 26 17:06:58 crc kubenswrapper[4754]: I0126 17:06:58.194873 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kl9sg\" (UniqueName: \"kubernetes.io/projected/fdbb773a-337d-44f0-8f43-380cbc74bd43-kube-api-access-kl9sg\") pod \"ovn-controller-8bgjb-config-hfs7f\" (UID: \"fdbb773a-337d-44f0-8f43-380cbc74bd43\") " pod="openstack/ovn-controller-8bgjb-config-hfs7f" Jan 26 17:06:58 crc kubenswrapper[4754]: I0126 17:06:58.195046 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/fdbb773a-337d-44f0-8f43-380cbc74bd43-var-run\") pod \"ovn-controller-8bgjb-config-hfs7f\" (UID: \"fdbb773a-337d-44f0-8f43-380cbc74bd43\") " pod="openstack/ovn-controller-8bgjb-config-hfs7f" Jan 26 17:06:58 crc kubenswrapper[4754]: I0126 17:06:58.195071 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/fdbb773a-337d-44f0-8f43-380cbc74bd43-var-log-ovn\") pod \"ovn-controller-8bgjb-config-hfs7f\" (UID: \"fdbb773a-337d-44f0-8f43-380cbc74bd43\") " pod="openstack/ovn-controller-8bgjb-config-hfs7f" Jan 26 17:06:58 crc kubenswrapper[4754]: I0126 17:06:58.195046 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/fdbb773a-337d-44f0-8f43-380cbc74bd43-var-run-ovn\") pod \"ovn-controller-8bgjb-config-hfs7f\" (UID: \"fdbb773a-337d-44f0-8f43-380cbc74bd43\") " pod="openstack/ovn-controller-8bgjb-config-hfs7f" Jan 26 17:06:58 crc kubenswrapper[4754]: I0126 17:06:58.195548 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/fdbb773a-337d-44f0-8f43-380cbc74bd43-additional-scripts\") pod \"ovn-controller-8bgjb-config-hfs7f\" (UID: \"fdbb773a-337d-44f0-8f43-380cbc74bd43\") " pod="openstack/ovn-controller-8bgjb-config-hfs7f" Jan 26 17:06:58 crc kubenswrapper[4754]: I0126 17:06:58.196987 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fdbb773a-337d-44f0-8f43-380cbc74bd43-scripts\") pod \"ovn-controller-8bgjb-config-hfs7f\" (UID: \"fdbb773a-337d-44f0-8f43-380cbc74bd43\") " pod="openstack/ovn-controller-8bgjb-config-hfs7f" Jan 26 17:06:58 crc kubenswrapper[4754]: I0126 17:06:58.210723 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kl9sg\" (UniqueName: \"kubernetes.io/projected/fdbb773a-337d-44f0-8f43-380cbc74bd43-kube-api-access-kl9sg\") pod \"ovn-controller-8bgjb-config-hfs7f\" (UID: \"fdbb773a-337d-44f0-8f43-380cbc74bd43\") " pod="openstack/ovn-controller-8bgjb-config-hfs7f" Jan 26 17:06:58 crc kubenswrapper[4754]: I0126 17:06:58.347868 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-8bgjb-config-hfs7f" Jan 26 17:06:58 crc kubenswrapper[4754]: I0126 17:06:58.613327 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-wjjd9" event={"ID":"05f7ca48-d18a-42cf-921f-43c3fc4111ff","Type":"ContainerDied","Data":"b709c92b9fca849837b78494197fece43e924fd2d5ca9eb04c815516871577f4"} Jan 26 17:06:58 crc kubenswrapper[4754]: I0126 17:06:58.613374 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-wjjd9" Jan 26 17:06:58 crc kubenswrapper[4754]: I0126 17:06:58.613376 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b709c92b9fca849837b78494197fece43e924fd2d5ca9eb04c815516871577f4" Jan 26 17:06:59 crc kubenswrapper[4754]: I0126 17:06:59.410377 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/root-account-create-update-wjjd9"] Jan 26 17:06:59 crc kubenswrapper[4754]: I0126 17:06:59.417483 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/root-account-create-update-wjjd9"] Jan 26 17:06:59 crc kubenswrapper[4754]: I0126 17:06:59.509121 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-8bgjb-config-hfs7f"] Jan 26 17:06:59 crc kubenswrapper[4754]: I0126 17:06:59.624134 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-8bgjb-config-hfs7f" event={"ID":"fdbb773a-337d-44f0-8f43-380cbc74bd43","Type":"ContainerStarted","Data":"985cdc799ae91fc2a36682e3fdc0601c3463cc5f54facf69eacdccb20193a712"} Jan 26 17:06:59 crc kubenswrapper[4754]: I0126 17:06:59.777631 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="05f7ca48-d18a-42cf-921f-43c3fc4111ff" path="/var/lib/kubelet/pods/05f7ca48-d18a-42cf-921f-43c3fc4111ff/volumes" Jan 26 17:06:59 crc kubenswrapper[4754]: I0126 17:06:59.778654 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e71ee5fd-3be8-49ff-96cd-3f5e67a55bad" path="/var/lib/kubelet/pods/e71ee5fd-3be8-49ff-96cd-3f5e67a55bad/volumes" Jan 26 17:07:00 crc kubenswrapper[4754]: I0126 17:07:00.638988 4754 generic.go:334] "Generic (PLEG): container finished" podID="fdbb773a-337d-44f0-8f43-380cbc74bd43" containerID="877b01f8012ec09c2df866b01599ea7ed3a790707891024efe46f3aaca9cc418" exitCode=0 Jan 26 17:07:00 crc kubenswrapper[4754]: I0126 17:07:00.639072 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-8bgjb-config-hfs7f" event={"ID":"fdbb773a-337d-44f0-8f43-380cbc74bd43","Type":"ContainerDied","Data":"877b01f8012ec09c2df866b01599ea7ed3a790707891024efe46f3aaca9cc418"} Jan 26 17:07:00 crc kubenswrapper[4754]: I0126 17:07:00.642168 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"19a0730c-46d0-4029-a86b-812ba1664dcc","Type":"ContainerStarted","Data":"f8607df2dc80f3490367dc22934ae3c73dc6502ab9eeadec3ad38afdf0580eeb"} Jan 26 17:07:00 crc kubenswrapper[4754]: I0126 17:07:00.642201 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"19a0730c-46d0-4029-a86b-812ba1664dcc","Type":"ContainerStarted","Data":"17b479b537c3cc3743c3e5b6118422033545dc31d29761d90042208bc79e113a"} Jan 26 17:07:01 crc kubenswrapper[4754]: I0126 17:07:01.653488 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"19a0730c-46d0-4029-a86b-812ba1664dcc","Type":"ContainerStarted","Data":"6af822d1954cf14140fd28e384fe47ea7e8cd324ba078689b6ac6fa647c37036"} Jan 26 17:07:01 crc kubenswrapper[4754]: I0126 17:07:01.653797 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"19a0730c-46d0-4029-a86b-812ba1664dcc","Type":"ContainerStarted","Data":"6ed9d2a5b9eab8b34d72e0cd6ec834a2adbdae6d6ac9b78724c4626497324c11"} Jan 26 17:07:01 crc kubenswrapper[4754]: I0126 17:07:01.934085 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-8bgjb-config-hfs7f" Jan 26 17:07:02 crc kubenswrapper[4754]: I0126 17:07:02.053584 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fdbb773a-337d-44f0-8f43-380cbc74bd43-scripts\") pod \"fdbb773a-337d-44f0-8f43-380cbc74bd43\" (UID: \"fdbb773a-337d-44f0-8f43-380cbc74bd43\") " Jan 26 17:07:02 crc kubenswrapper[4754]: I0126 17:07:02.053721 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kl9sg\" (UniqueName: \"kubernetes.io/projected/fdbb773a-337d-44f0-8f43-380cbc74bd43-kube-api-access-kl9sg\") pod \"fdbb773a-337d-44f0-8f43-380cbc74bd43\" (UID: \"fdbb773a-337d-44f0-8f43-380cbc74bd43\") " Jan 26 17:07:02 crc kubenswrapper[4754]: I0126 17:07:02.053883 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/fdbb773a-337d-44f0-8f43-380cbc74bd43-var-run-ovn\") pod \"fdbb773a-337d-44f0-8f43-380cbc74bd43\" (UID: \"fdbb773a-337d-44f0-8f43-380cbc74bd43\") " Jan 26 17:07:02 crc kubenswrapper[4754]: I0126 17:07:02.053998 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/fdbb773a-337d-44f0-8f43-380cbc74bd43-var-run\") pod \"fdbb773a-337d-44f0-8f43-380cbc74bd43\" (UID: \"fdbb773a-337d-44f0-8f43-380cbc74bd43\") " Jan 26 17:07:02 crc kubenswrapper[4754]: I0126 17:07:02.054057 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/fdbb773a-337d-44f0-8f43-380cbc74bd43-additional-scripts\") pod \"fdbb773a-337d-44f0-8f43-380cbc74bd43\" (UID: \"fdbb773a-337d-44f0-8f43-380cbc74bd43\") " Jan 26 17:07:02 crc kubenswrapper[4754]: I0126 17:07:02.054139 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/fdbb773a-337d-44f0-8f43-380cbc74bd43-var-log-ovn\") pod \"fdbb773a-337d-44f0-8f43-380cbc74bd43\" (UID: \"fdbb773a-337d-44f0-8f43-380cbc74bd43\") " Jan 26 17:07:02 crc kubenswrapper[4754]: I0126 17:07:02.054063 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fdbb773a-337d-44f0-8f43-380cbc74bd43-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "fdbb773a-337d-44f0-8f43-380cbc74bd43" (UID: "fdbb773a-337d-44f0-8f43-380cbc74bd43"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 17:07:02 crc kubenswrapper[4754]: I0126 17:07:02.054084 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fdbb773a-337d-44f0-8f43-380cbc74bd43-var-run" (OuterVolumeSpecName: "var-run") pod "fdbb773a-337d-44f0-8f43-380cbc74bd43" (UID: "fdbb773a-337d-44f0-8f43-380cbc74bd43"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 17:07:02 crc kubenswrapper[4754]: I0126 17:07:02.054290 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fdbb773a-337d-44f0-8f43-380cbc74bd43-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "fdbb773a-337d-44f0-8f43-380cbc74bd43" (UID: "fdbb773a-337d-44f0-8f43-380cbc74bd43"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 17:07:02 crc kubenswrapper[4754]: I0126 17:07:02.054794 4754 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/fdbb773a-337d-44f0-8f43-380cbc74bd43-var-run-ovn\") on node \"crc\" DevicePath \"\"" Jan 26 17:07:02 crc kubenswrapper[4754]: I0126 17:07:02.054816 4754 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/fdbb773a-337d-44f0-8f43-380cbc74bd43-var-run\") on node \"crc\" DevicePath \"\"" Jan 26 17:07:02 crc kubenswrapper[4754]: I0126 17:07:02.054824 4754 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/fdbb773a-337d-44f0-8f43-380cbc74bd43-var-log-ovn\") on node \"crc\" DevicePath \"\"" Jan 26 17:07:02 crc kubenswrapper[4754]: I0126 17:07:02.054889 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fdbb773a-337d-44f0-8f43-380cbc74bd43-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "fdbb773a-337d-44f0-8f43-380cbc74bd43" (UID: "fdbb773a-337d-44f0-8f43-380cbc74bd43"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:07:02 crc kubenswrapper[4754]: I0126 17:07:02.055177 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fdbb773a-337d-44f0-8f43-380cbc74bd43-scripts" (OuterVolumeSpecName: "scripts") pod "fdbb773a-337d-44f0-8f43-380cbc74bd43" (UID: "fdbb773a-337d-44f0-8f43-380cbc74bd43"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:07:02 crc kubenswrapper[4754]: I0126 17:07:02.059427 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fdbb773a-337d-44f0-8f43-380cbc74bd43-kube-api-access-kl9sg" (OuterVolumeSpecName: "kube-api-access-kl9sg") pod "fdbb773a-337d-44f0-8f43-380cbc74bd43" (UID: "fdbb773a-337d-44f0-8f43-380cbc74bd43"). InnerVolumeSpecName "kube-api-access-kl9sg". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:07:02 crc kubenswrapper[4754]: I0126 17:07:02.155830 4754 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fdbb773a-337d-44f0-8f43-380cbc74bd43-scripts\") on node \"crc\" DevicePath \"\"" Jan 26 17:07:02 crc kubenswrapper[4754]: I0126 17:07:02.155862 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kl9sg\" (UniqueName: \"kubernetes.io/projected/fdbb773a-337d-44f0-8f43-380cbc74bd43-kube-api-access-kl9sg\") on node \"crc\" DevicePath \"\"" Jan 26 17:07:02 crc kubenswrapper[4754]: I0126 17:07:02.155871 4754 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/fdbb773a-337d-44f0-8f43-380cbc74bd43-additional-scripts\") on node \"crc\" DevicePath \"\"" Jan 26 17:07:02 crc kubenswrapper[4754]: I0126 17:07:02.664589 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-8bgjb-config-hfs7f" event={"ID":"fdbb773a-337d-44f0-8f43-380cbc74bd43","Type":"ContainerDied","Data":"985cdc799ae91fc2a36682e3fdc0601c3463cc5f54facf69eacdccb20193a712"} Jan 26 17:07:02 crc kubenswrapper[4754]: I0126 17:07:02.664639 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="985cdc799ae91fc2a36682e3fdc0601c3463cc5f54facf69eacdccb20193a712" Jan 26 17:07:02 crc kubenswrapper[4754]: I0126 17:07:02.665508 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-8bgjb-config-hfs7f" Jan 26 17:07:03 crc kubenswrapper[4754]: I0126 17:07:03.030806 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-8bgjb-config-hfs7f"] Jan 26 17:07:03 crc kubenswrapper[4754]: I0126 17:07:03.037160 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-8bgjb-config-hfs7f"] Jan 26 17:07:03 crc kubenswrapper[4754]: I0126 17:07:03.064489 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/root-account-create-update-bp5dl"] Jan 26 17:07:03 crc kubenswrapper[4754]: E0126 17:07:03.064970 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fdbb773a-337d-44f0-8f43-380cbc74bd43" containerName="ovn-config" Jan 26 17:07:03 crc kubenswrapper[4754]: I0126 17:07:03.064994 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="fdbb773a-337d-44f0-8f43-380cbc74bd43" containerName="ovn-config" Jan 26 17:07:03 crc kubenswrapper[4754]: I0126 17:07:03.065226 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="fdbb773a-337d-44f0-8f43-380cbc74bd43" containerName="ovn-config" Jan 26 17:07:03 crc kubenswrapper[4754]: I0126 17:07:03.066048 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-bp5dl" Jan 26 17:07:03 crc kubenswrapper[4754]: I0126 17:07:03.067930 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-mariadb-root-db-secret" Jan 26 17:07:03 crc kubenswrapper[4754]: I0126 17:07:03.082755 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-bp5dl"] Jan 26 17:07:03 crc kubenswrapper[4754]: I0126 17:07:03.157533 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-8bgjb-config-cnrml"] Jan 26 17:07:03 crc kubenswrapper[4754]: I0126 17:07:03.158728 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-8bgjb-config-cnrml" Jan 26 17:07:03 crc kubenswrapper[4754]: I0126 17:07:03.161791 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Jan 26 17:07:03 crc kubenswrapper[4754]: I0126 17:07:03.166329 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-8bgjb-config-cnrml"] Jan 26 17:07:03 crc kubenswrapper[4754]: I0126 17:07:03.184150 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4ad7fbd5-cd5f-4637-b2a1-8c71df037ed6-operator-scripts\") pod \"root-account-create-update-bp5dl\" (UID: \"4ad7fbd5-cd5f-4637-b2a1-8c71df037ed6\") " pod="openstack/root-account-create-update-bp5dl" Jan 26 17:07:03 crc kubenswrapper[4754]: I0126 17:07:03.184349 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2k6kq\" (UniqueName: \"kubernetes.io/projected/4ad7fbd5-cd5f-4637-b2a1-8c71df037ed6-kube-api-access-2k6kq\") pod \"root-account-create-update-bp5dl\" (UID: \"4ad7fbd5-cd5f-4637-b2a1-8c71df037ed6\") " pod="openstack/root-account-create-update-bp5dl" Jan 26 17:07:03 crc kubenswrapper[4754]: I0126 17:07:03.286168 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/944939e4-26fe-4583-9c20-4092f6f5dab3-var-run-ovn\") pod \"ovn-controller-8bgjb-config-cnrml\" (UID: \"944939e4-26fe-4583-9c20-4092f6f5dab3\") " pod="openstack/ovn-controller-8bgjb-config-cnrml" Jan 26 17:07:03 crc kubenswrapper[4754]: I0126 17:07:03.286221 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2k6kq\" (UniqueName: \"kubernetes.io/projected/4ad7fbd5-cd5f-4637-b2a1-8c71df037ed6-kube-api-access-2k6kq\") pod \"root-account-create-update-bp5dl\" (UID: \"4ad7fbd5-cd5f-4637-b2a1-8c71df037ed6\") " pod="openstack/root-account-create-update-bp5dl" Jan 26 17:07:03 crc kubenswrapper[4754]: I0126 17:07:03.286258 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zpvhq\" (UniqueName: \"kubernetes.io/projected/944939e4-26fe-4583-9c20-4092f6f5dab3-kube-api-access-zpvhq\") pod \"ovn-controller-8bgjb-config-cnrml\" (UID: \"944939e4-26fe-4583-9c20-4092f6f5dab3\") " pod="openstack/ovn-controller-8bgjb-config-cnrml" Jan 26 17:07:03 crc kubenswrapper[4754]: I0126 17:07:03.286295 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/944939e4-26fe-4583-9c20-4092f6f5dab3-var-run\") pod \"ovn-controller-8bgjb-config-cnrml\" (UID: \"944939e4-26fe-4583-9c20-4092f6f5dab3\") " pod="openstack/ovn-controller-8bgjb-config-cnrml" Jan 26 17:07:03 crc kubenswrapper[4754]: I0126 17:07:03.286331 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4ad7fbd5-cd5f-4637-b2a1-8c71df037ed6-operator-scripts\") pod \"root-account-create-update-bp5dl\" (UID: \"4ad7fbd5-cd5f-4637-b2a1-8c71df037ed6\") " pod="openstack/root-account-create-update-bp5dl" Jan 26 17:07:03 crc kubenswrapper[4754]: I0126 17:07:03.286365 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/944939e4-26fe-4583-9c20-4092f6f5dab3-var-log-ovn\") pod \"ovn-controller-8bgjb-config-cnrml\" (UID: \"944939e4-26fe-4583-9c20-4092f6f5dab3\") " pod="openstack/ovn-controller-8bgjb-config-cnrml" Jan 26 17:07:03 crc kubenswrapper[4754]: I0126 17:07:03.286389 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/944939e4-26fe-4583-9c20-4092f6f5dab3-additional-scripts\") pod \"ovn-controller-8bgjb-config-cnrml\" (UID: \"944939e4-26fe-4583-9c20-4092f6f5dab3\") " pod="openstack/ovn-controller-8bgjb-config-cnrml" Jan 26 17:07:03 crc kubenswrapper[4754]: I0126 17:07:03.286417 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/944939e4-26fe-4583-9c20-4092f6f5dab3-scripts\") pod \"ovn-controller-8bgjb-config-cnrml\" (UID: \"944939e4-26fe-4583-9c20-4092f6f5dab3\") " pod="openstack/ovn-controller-8bgjb-config-cnrml" Jan 26 17:07:03 crc kubenswrapper[4754]: I0126 17:07:03.287238 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4ad7fbd5-cd5f-4637-b2a1-8c71df037ed6-operator-scripts\") pod \"root-account-create-update-bp5dl\" (UID: \"4ad7fbd5-cd5f-4637-b2a1-8c71df037ed6\") " pod="openstack/root-account-create-update-bp5dl" Jan 26 17:07:03 crc kubenswrapper[4754]: I0126 17:07:03.307169 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2k6kq\" (UniqueName: \"kubernetes.io/projected/4ad7fbd5-cd5f-4637-b2a1-8c71df037ed6-kube-api-access-2k6kq\") pod \"root-account-create-update-bp5dl\" (UID: \"4ad7fbd5-cd5f-4637-b2a1-8c71df037ed6\") " pod="openstack/root-account-create-update-bp5dl" Jan 26 17:07:03 crc kubenswrapper[4754]: I0126 17:07:03.385012 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-bp5dl" Jan 26 17:07:03 crc kubenswrapper[4754]: I0126 17:07:03.388353 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zpvhq\" (UniqueName: \"kubernetes.io/projected/944939e4-26fe-4583-9c20-4092f6f5dab3-kube-api-access-zpvhq\") pod \"ovn-controller-8bgjb-config-cnrml\" (UID: \"944939e4-26fe-4583-9c20-4092f6f5dab3\") " pod="openstack/ovn-controller-8bgjb-config-cnrml" Jan 26 17:07:03 crc kubenswrapper[4754]: I0126 17:07:03.388417 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/944939e4-26fe-4583-9c20-4092f6f5dab3-var-run\") pod \"ovn-controller-8bgjb-config-cnrml\" (UID: \"944939e4-26fe-4583-9c20-4092f6f5dab3\") " pod="openstack/ovn-controller-8bgjb-config-cnrml" Jan 26 17:07:03 crc kubenswrapper[4754]: I0126 17:07:03.388467 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/944939e4-26fe-4583-9c20-4092f6f5dab3-var-log-ovn\") pod \"ovn-controller-8bgjb-config-cnrml\" (UID: \"944939e4-26fe-4583-9c20-4092f6f5dab3\") " pod="openstack/ovn-controller-8bgjb-config-cnrml" Jan 26 17:07:03 crc kubenswrapper[4754]: I0126 17:07:03.388491 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/944939e4-26fe-4583-9c20-4092f6f5dab3-additional-scripts\") pod \"ovn-controller-8bgjb-config-cnrml\" (UID: \"944939e4-26fe-4583-9c20-4092f6f5dab3\") " pod="openstack/ovn-controller-8bgjb-config-cnrml" Jan 26 17:07:03 crc kubenswrapper[4754]: I0126 17:07:03.388511 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/944939e4-26fe-4583-9c20-4092f6f5dab3-scripts\") pod \"ovn-controller-8bgjb-config-cnrml\" (UID: \"944939e4-26fe-4583-9c20-4092f6f5dab3\") " pod="openstack/ovn-controller-8bgjb-config-cnrml" Jan 26 17:07:03 crc kubenswrapper[4754]: I0126 17:07:03.388592 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/944939e4-26fe-4583-9c20-4092f6f5dab3-var-run-ovn\") pod \"ovn-controller-8bgjb-config-cnrml\" (UID: \"944939e4-26fe-4583-9c20-4092f6f5dab3\") " pod="openstack/ovn-controller-8bgjb-config-cnrml" Jan 26 17:07:03 crc kubenswrapper[4754]: I0126 17:07:03.388803 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/944939e4-26fe-4583-9c20-4092f6f5dab3-var-run-ovn\") pod \"ovn-controller-8bgjb-config-cnrml\" (UID: \"944939e4-26fe-4583-9c20-4092f6f5dab3\") " pod="openstack/ovn-controller-8bgjb-config-cnrml" Jan 26 17:07:03 crc kubenswrapper[4754]: I0126 17:07:03.388816 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/944939e4-26fe-4583-9c20-4092f6f5dab3-var-log-ovn\") pod \"ovn-controller-8bgjb-config-cnrml\" (UID: \"944939e4-26fe-4583-9c20-4092f6f5dab3\") " pod="openstack/ovn-controller-8bgjb-config-cnrml" Jan 26 17:07:03 crc kubenswrapper[4754]: I0126 17:07:03.388806 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/944939e4-26fe-4583-9c20-4092f6f5dab3-var-run\") pod \"ovn-controller-8bgjb-config-cnrml\" (UID: \"944939e4-26fe-4583-9c20-4092f6f5dab3\") " pod="openstack/ovn-controller-8bgjb-config-cnrml" Jan 26 17:07:03 crc kubenswrapper[4754]: I0126 17:07:03.389891 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/944939e4-26fe-4583-9c20-4092f6f5dab3-additional-scripts\") pod \"ovn-controller-8bgjb-config-cnrml\" (UID: \"944939e4-26fe-4583-9c20-4092f6f5dab3\") " pod="openstack/ovn-controller-8bgjb-config-cnrml" Jan 26 17:07:03 crc kubenswrapper[4754]: I0126 17:07:03.391118 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/944939e4-26fe-4583-9c20-4092f6f5dab3-scripts\") pod \"ovn-controller-8bgjb-config-cnrml\" (UID: \"944939e4-26fe-4583-9c20-4092f6f5dab3\") " pod="openstack/ovn-controller-8bgjb-config-cnrml" Jan 26 17:07:03 crc kubenswrapper[4754]: I0126 17:07:03.407961 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zpvhq\" (UniqueName: \"kubernetes.io/projected/944939e4-26fe-4583-9c20-4092f6f5dab3-kube-api-access-zpvhq\") pod \"ovn-controller-8bgjb-config-cnrml\" (UID: \"944939e4-26fe-4583-9c20-4092f6f5dab3\") " pod="openstack/ovn-controller-8bgjb-config-cnrml" Jan 26 17:07:03 crc kubenswrapper[4754]: I0126 17:07:03.479276 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-8bgjb-config-cnrml" Jan 26 17:07:03 crc kubenswrapper[4754]: I0126 17:07:03.779525 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fdbb773a-337d-44f0-8f43-380cbc74bd43" path="/var/lib/kubelet/pods/fdbb773a-337d-44f0-8f43-380cbc74bd43/volumes" Jan 26 17:07:03 crc kubenswrapper[4754]: I0126 17:07:03.868037 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-bp5dl"] Jan 26 17:07:04 crc kubenswrapper[4754]: I0126 17:07:04.020411 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-8bgjb-config-cnrml"] Jan 26 17:07:04 crc kubenswrapper[4754]: W0126 17:07:04.458870 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod944939e4_26fe_4583_9c20_4092f6f5dab3.slice/crio-c6d929c4e7c0419a6afc90b1152d7d188605c7801b3f190cd6b28b967353a90d WatchSource:0}: Error finding container c6d929c4e7c0419a6afc90b1152d7d188605c7801b3f190cd6b28b967353a90d: Status 404 returned error can't find the container with id c6d929c4e7c0419a6afc90b1152d7d188605c7801b3f190cd6b28b967353a90d Jan 26 17:07:04 crc kubenswrapper[4754]: W0126 17:07:04.460073 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4ad7fbd5_cd5f_4637_b2a1_8c71df037ed6.slice/crio-1ee169015c6fb6de3bdb02a7cfaef5bff5fbdd1a2ce21ae418861d6e663bf5aa WatchSource:0}: Error finding container 1ee169015c6fb6de3bdb02a7cfaef5bff5fbdd1a2ce21ae418861d6e663bf5aa: Status 404 returned error can't find the container with id 1ee169015c6fb6de3bdb02a7cfaef5bff5fbdd1a2ce21ae418861d6e663bf5aa Jan 26 17:07:04 crc kubenswrapper[4754]: I0126 17:07:04.682219 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-bp5dl" event={"ID":"4ad7fbd5-cd5f-4637-b2a1-8c71df037ed6","Type":"ContainerStarted","Data":"1ee169015c6fb6de3bdb02a7cfaef5bff5fbdd1a2ce21ae418861d6e663bf5aa"} Jan 26 17:07:04 crc kubenswrapper[4754]: I0126 17:07:04.683808 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-8bgjb-config-cnrml" event={"ID":"944939e4-26fe-4583-9c20-4092f6f5dab3","Type":"ContainerStarted","Data":"c6d929c4e7c0419a6afc90b1152d7d188605c7801b3f190cd6b28b967353a90d"} Jan 26 17:07:05 crc kubenswrapper[4754]: I0126 17:07:05.698686 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"19a0730c-46d0-4029-a86b-812ba1664dcc","Type":"ContainerStarted","Data":"5261c88f0581fe92975f9282ffd2956a0c6775aef4b55ee778e4704d6969a534"} Jan 26 17:07:05 crc kubenswrapper[4754]: I0126 17:07:05.699162 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"19a0730c-46d0-4029-a86b-812ba1664dcc","Type":"ContainerStarted","Data":"e0329d671e6facf2878184dde15169160dfc203e7b33a7f26e81e5a4ff8e8aa1"} Jan 26 17:07:05 crc kubenswrapper[4754]: I0126 17:07:05.699173 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"19a0730c-46d0-4029-a86b-812ba1664dcc","Type":"ContainerStarted","Data":"d2a96777ea6b37c4b988f732285c922b258d3600439898fc084f2b0f6dddc0bd"} Jan 26 17:07:05 crc kubenswrapper[4754]: I0126 17:07:05.699182 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"19a0730c-46d0-4029-a86b-812ba1664dcc","Type":"ContainerStarted","Data":"c4dcf143ccd226617eb683ebf8da6fe2e6e50ae5ce293e1d76b5b820b3843db9"} Jan 26 17:07:05 crc kubenswrapper[4754]: I0126 17:07:05.701302 4754 generic.go:334] "Generic (PLEG): container finished" podID="944939e4-26fe-4583-9c20-4092f6f5dab3" containerID="33a0ab55bbd24b76d21481a4d6e3bb9482f39506cf37531f314c5be370502159" exitCode=0 Jan 26 17:07:05 crc kubenswrapper[4754]: I0126 17:07:05.701639 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-8bgjb-config-cnrml" event={"ID":"944939e4-26fe-4583-9c20-4092f6f5dab3","Type":"ContainerDied","Data":"33a0ab55bbd24b76d21481a4d6e3bb9482f39506cf37531f314c5be370502159"} Jan 26 17:07:05 crc kubenswrapper[4754]: I0126 17:07:05.710103 4754 generic.go:334] "Generic (PLEG): container finished" podID="4ad7fbd5-cd5f-4637-b2a1-8c71df037ed6" containerID="18791aca8574779fa910ccd041b92873e7acc45391f6423c1c9e799ef87424a0" exitCode=0 Jan 26 17:07:05 crc kubenswrapper[4754]: I0126 17:07:05.710173 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-bp5dl" event={"ID":"4ad7fbd5-cd5f-4637-b2a1-8c71df037ed6","Type":"ContainerDied","Data":"18791aca8574779fa910ccd041b92873e7acc45391f6423c1c9e799ef87424a0"} Jan 26 17:07:06 crc kubenswrapper[4754]: I0126 17:07:06.724786 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"19a0730c-46d0-4029-a86b-812ba1664dcc","Type":"ContainerStarted","Data":"0a02640cce32f0eb253314abf104b9e67dcba124c5ac36638a16dc0bc7f24b0e"} Jan 26 17:07:06 crc kubenswrapper[4754]: I0126 17:07:06.725090 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"19a0730c-46d0-4029-a86b-812ba1664dcc","Type":"ContainerStarted","Data":"04c64be18d9a87867e47a282134fbb6dffd35da8fbfb3682df177d352fe043da"} Jan 26 17:07:07 crc kubenswrapper[4754]: I0126 17:07:07.010308 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-8bgjb-config-cnrml" Jan 26 17:07:07 crc kubenswrapper[4754]: I0126 17:07:07.064169 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-bp5dl" Jan 26 17:07:07 crc kubenswrapper[4754]: I0126 17:07:07.172586 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4ad7fbd5-cd5f-4637-b2a1-8c71df037ed6-operator-scripts\") pod \"4ad7fbd5-cd5f-4637-b2a1-8c71df037ed6\" (UID: \"4ad7fbd5-cd5f-4637-b2a1-8c71df037ed6\") " Jan 26 17:07:07 crc kubenswrapper[4754]: I0126 17:07:07.172660 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/944939e4-26fe-4583-9c20-4092f6f5dab3-scripts\") pod \"944939e4-26fe-4583-9c20-4092f6f5dab3\" (UID: \"944939e4-26fe-4583-9c20-4092f6f5dab3\") " Jan 26 17:07:07 crc kubenswrapper[4754]: I0126 17:07:07.172717 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/944939e4-26fe-4583-9c20-4092f6f5dab3-var-run-ovn\") pod \"944939e4-26fe-4583-9c20-4092f6f5dab3\" (UID: \"944939e4-26fe-4583-9c20-4092f6f5dab3\") " Jan 26 17:07:07 crc kubenswrapper[4754]: I0126 17:07:07.172765 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zpvhq\" (UniqueName: \"kubernetes.io/projected/944939e4-26fe-4583-9c20-4092f6f5dab3-kube-api-access-zpvhq\") pod \"944939e4-26fe-4583-9c20-4092f6f5dab3\" (UID: \"944939e4-26fe-4583-9c20-4092f6f5dab3\") " Jan 26 17:07:07 crc kubenswrapper[4754]: I0126 17:07:07.172811 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/944939e4-26fe-4583-9c20-4092f6f5dab3-var-run\") pod \"944939e4-26fe-4583-9c20-4092f6f5dab3\" (UID: \"944939e4-26fe-4583-9c20-4092f6f5dab3\") " Jan 26 17:07:07 crc kubenswrapper[4754]: I0126 17:07:07.172868 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/944939e4-26fe-4583-9c20-4092f6f5dab3-additional-scripts\") pod \"944939e4-26fe-4583-9c20-4092f6f5dab3\" (UID: \"944939e4-26fe-4583-9c20-4092f6f5dab3\") " Jan 26 17:07:07 crc kubenswrapper[4754]: I0126 17:07:07.172855 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/944939e4-26fe-4583-9c20-4092f6f5dab3-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "944939e4-26fe-4583-9c20-4092f6f5dab3" (UID: "944939e4-26fe-4583-9c20-4092f6f5dab3"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 17:07:07 crc kubenswrapper[4754]: I0126 17:07:07.172922 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/944939e4-26fe-4583-9c20-4092f6f5dab3-var-run" (OuterVolumeSpecName: "var-run") pod "944939e4-26fe-4583-9c20-4092f6f5dab3" (UID: "944939e4-26fe-4583-9c20-4092f6f5dab3"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 17:07:07 crc kubenswrapper[4754]: I0126 17:07:07.172938 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2k6kq\" (UniqueName: \"kubernetes.io/projected/4ad7fbd5-cd5f-4637-b2a1-8c71df037ed6-kube-api-access-2k6kq\") pod \"4ad7fbd5-cd5f-4637-b2a1-8c71df037ed6\" (UID: \"4ad7fbd5-cd5f-4637-b2a1-8c71df037ed6\") " Jan 26 17:07:07 crc kubenswrapper[4754]: I0126 17:07:07.172981 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/944939e4-26fe-4583-9c20-4092f6f5dab3-var-log-ovn\") pod \"944939e4-26fe-4583-9c20-4092f6f5dab3\" (UID: \"944939e4-26fe-4583-9c20-4092f6f5dab3\") " Jan 26 17:07:07 crc kubenswrapper[4754]: I0126 17:07:07.173173 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/944939e4-26fe-4583-9c20-4092f6f5dab3-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "944939e4-26fe-4583-9c20-4092f6f5dab3" (UID: "944939e4-26fe-4583-9c20-4092f6f5dab3"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 17:07:07 crc kubenswrapper[4754]: I0126 17:07:07.173752 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4ad7fbd5-cd5f-4637-b2a1-8c71df037ed6-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "4ad7fbd5-cd5f-4637-b2a1-8c71df037ed6" (UID: "4ad7fbd5-cd5f-4637-b2a1-8c71df037ed6"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:07:07 crc kubenswrapper[4754]: I0126 17:07:07.173803 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/944939e4-26fe-4583-9c20-4092f6f5dab3-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "944939e4-26fe-4583-9c20-4092f6f5dab3" (UID: "944939e4-26fe-4583-9c20-4092f6f5dab3"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:07:07 crc kubenswrapper[4754]: I0126 17:07:07.173982 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/944939e4-26fe-4583-9c20-4092f6f5dab3-scripts" (OuterVolumeSpecName: "scripts") pod "944939e4-26fe-4583-9c20-4092f6f5dab3" (UID: "944939e4-26fe-4583-9c20-4092f6f5dab3"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:07:07 crc kubenswrapper[4754]: I0126 17:07:07.174097 4754 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4ad7fbd5-cd5f-4637-b2a1-8c71df037ed6-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 26 17:07:07 crc kubenswrapper[4754]: I0126 17:07:07.174132 4754 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/944939e4-26fe-4583-9c20-4092f6f5dab3-var-run-ovn\") on node \"crc\" DevicePath \"\"" Jan 26 17:07:07 crc kubenswrapper[4754]: I0126 17:07:07.174153 4754 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/944939e4-26fe-4583-9c20-4092f6f5dab3-var-run\") on node \"crc\" DevicePath \"\"" Jan 26 17:07:07 crc kubenswrapper[4754]: I0126 17:07:07.174170 4754 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/944939e4-26fe-4583-9c20-4092f6f5dab3-additional-scripts\") on node \"crc\" DevicePath \"\"" Jan 26 17:07:07 crc kubenswrapper[4754]: I0126 17:07:07.174187 4754 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/944939e4-26fe-4583-9c20-4092f6f5dab3-var-log-ovn\") on node \"crc\" DevicePath \"\"" Jan 26 17:07:07 crc kubenswrapper[4754]: I0126 17:07:07.176982 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/944939e4-26fe-4583-9c20-4092f6f5dab3-kube-api-access-zpvhq" (OuterVolumeSpecName: "kube-api-access-zpvhq") pod "944939e4-26fe-4583-9c20-4092f6f5dab3" (UID: "944939e4-26fe-4583-9c20-4092f6f5dab3"). InnerVolumeSpecName "kube-api-access-zpvhq". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:07:07 crc kubenswrapper[4754]: I0126 17:07:07.180867 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4ad7fbd5-cd5f-4637-b2a1-8c71df037ed6-kube-api-access-2k6kq" (OuterVolumeSpecName: "kube-api-access-2k6kq") pod "4ad7fbd5-cd5f-4637-b2a1-8c71df037ed6" (UID: "4ad7fbd5-cd5f-4637-b2a1-8c71df037ed6"). InnerVolumeSpecName "kube-api-access-2k6kq". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:07:07 crc kubenswrapper[4754]: I0126 17:07:07.275617 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zpvhq\" (UniqueName: \"kubernetes.io/projected/944939e4-26fe-4583-9c20-4092f6f5dab3-kube-api-access-zpvhq\") on node \"crc\" DevicePath \"\"" Jan 26 17:07:07 crc kubenswrapper[4754]: I0126 17:07:07.275954 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2k6kq\" (UniqueName: \"kubernetes.io/projected/4ad7fbd5-cd5f-4637-b2a1-8c71df037ed6-kube-api-access-2k6kq\") on node \"crc\" DevicePath \"\"" Jan 26 17:07:07 crc kubenswrapper[4754]: I0126 17:07:07.275971 4754 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/944939e4-26fe-4583-9c20-4092f6f5dab3-scripts\") on node \"crc\" DevicePath \"\"" Jan 26 17:07:07 crc kubenswrapper[4754]: I0126 17:07:07.751441 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"19a0730c-46d0-4029-a86b-812ba1664dcc","Type":"ContainerStarted","Data":"6c1ee40bd646127092a0410e6b7f1bce4bd9a2335cd8e3af51d09839106119f1"} Jan 26 17:07:07 crc kubenswrapper[4754]: I0126 17:07:07.751553 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"19a0730c-46d0-4029-a86b-812ba1664dcc","Type":"ContainerStarted","Data":"5e26d49cf1c041a07dda196774717582ef58fe378faf3fc5d8f8f0ffd7fbd9ed"} Jan 26 17:07:07 crc kubenswrapper[4754]: I0126 17:07:07.751572 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"19a0730c-46d0-4029-a86b-812ba1664dcc","Type":"ContainerStarted","Data":"42c85c2f047ea53356dd1b8da11c4304c3b1b978eb62dd1d2df220082115875f"} Jan 26 17:07:07 crc kubenswrapper[4754]: I0126 17:07:07.751589 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"19a0730c-46d0-4029-a86b-812ba1664dcc","Type":"ContainerStarted","Data":"b523629d60a6f1c6ef20aa25e6083d90b0cadfa787944bfbae20bcf71600cc66"} Jan 26 17:07:07 crc kubenswrapper[4754]: I0126 17:07:07.751607 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"19a0730c-46d0-4029-a86b-812ba1664dcc","Type":"ContainerStarted","Data":"f10530a7a19c67daab2d42da004e6f48d6675c580fff92a3dcf28016d7fcb746"} Jan 26 17:07:07 crc kubenswrapper[4754]: I0126 17:07:07.756252 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-8bgjb-config-cnrml" event={"ID":"944939e4-26fe-4583-9c20-4092f6f5dab3","Type":"ContainerDied","Data":"c6d929c4e7c0419a6afc90b1152d7d188605c7801b3f190cd6b28b967353a90d"} Jan 26 17:07:07 crc kubenswrapper[4754]: I0126 17:07:07.756291 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c6d929c4e7c0419a6afc90b1152d7d188605c7801b3f190cd6b28b967353a90d" Jan 26 17:07:07 crc kubenswrapper[4754]: I0126 17:07:07.756291 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-8bgjb-config-cnrml" Jan 26 17:07:07 crc kubenswrapper[4754]: I0126 17:07:07.766993 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-bp5dl" Jan 26 17:07:07 crc kubenswrapper[4754]: I0126 17:07:07.780470 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-bp5dl" event={"ID":"4ad7fbd5-cd5f-4637-b2a1-8c71df037ed6","Type":"ContainerDied","Data":"1ee169015c6fb6de3bdb02a7cfaef5bff5fbdd1a2ce21ae418861d6e663bf5aa"} Jan 26 17:07:07 crc kubenswrapper[4754]: I0126 17:07:07.780508 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1ee169015c6fb6de3bdb02a7cfaef5bff5fbdd1a2ce21ae418861d6e663bf5aa" Jan 26 17:07:07 crc kubenswrapper[4754]: I0126 17:07:07.794922 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=35.787531511 podStartE2EDuration="44.794901738s" podCreationTimestamp="2026-01-26 17:06:23 +0000 UTC" firstStartedPulling="2026-01-26 17:06:57.254762297 +0000 UTC m=+1183.778942731" lastFinishedPulling="2026-01-26 17:07:06.262132484 +0000 UTC m=+1192.786312958" observedRunningTime="2026-01-26 17:07:07.78889819 +0000 UTC m=+1194.313078624" watchObservedRunningTime="2026-01-26 17:07:07.794901738 +0000 UTC m=+1194.319082172" Jan 26 17:07:07 crc kubenswrapper[4754]: I0126 17:07:07.894008 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Jan 26 17:07:07 crc kubenswrapper[4754]: I0126 17:07:07.960886 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Jan 26 17:07:08 crc kubenswrapper[4754]: I0126 17:07:08.166906 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-8bgjb-config-cnrml"] Jan 26 17:07:08 crc kubenswrapper[4754]: I0126 17:07:08.178774 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-8bgjb-config-cnrml"] Jan 26 17:07:08 crc kubenswrapper[4754]: I0126 17:07:08.223557 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-764c5664d7-gmrtk"] Jan 26 17:07:08 crc kubenswrapper[4754]: E0126 17:07:08.223992 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ad7fbd5-cd5f-4637-b2a1-8c71df037ed6" containerName="mariadb-account-create-update" Jan 26 17:07:08 crc kubenswrapper[4754]: I0126 17:07:08.224073 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ad7fbd5-cd5f-4637-b2a1-8c71df037ed6" containerName="mariadb-account-create-update" Jan 26 17:07:08 crc kubenswrapper[4754]: E0126 17:07:08.224161 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="944939e4-26fe-4583-9c20-4092f6f5dab3" containerName="ovn-config" Jan 26 17:07:08 crc kubenswrapper[4754]: I0126 17:07:08.224223 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="944939e4-26fe-4583-9c20-4092f6f5dab3" containerName="ovn-config" Jan 26 17:07:08 crc kubenswrapper[4754]: I0126 17:07:08.224441 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="944939e4-26fe-4583-9c20-4092f6f5dab3" containerName="ovn-config" Jan 26 17:07:08 crc kubenswrapper[4754]: I0126 17:07:08.224532 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="4ad7fbd5-cd5f-4637-b2a1-8c71df037ed6" containerName="mariadb-account-create-update" Jan 26 17:07:08 crc kubenswrapper[4754]: I0126 17:07:08.226915 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-764c5664d7-gmrtk" Jan 26 17:07:08 crc kubenswrapper[4754]: I0126 17:07:08.234896 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Jan 26 17:07:08 crc kubenswrapper[4754]: I0126 17:07:08.280328 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-764c5664d7-gmrtk"] Jan 26 17:07:08 crc kubenswrapper[4754]: I0126 17:07:08.300259 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5d5e322d-aa8b-4f36-8257-09d1ccab02c6-ovsdbserver-nb\") pod \"dnsmasq-dns-764c5664d7-gmrtk\" (UID: \"5d5e322d-aa8b-4f36-8257-09d1ccab02c6\") " pod="openstack/dnsmasq-dns-764c5664d7-gmrtk" Jan 26 17:07:08 crc kubenswrapper[4754]: I0126 17:07:08.300526 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5d5e322d-aa8b-4f36-8257-09d1ccab02c6-dns-svc\") pod \"dnsmasq-dns-764c5664d7-gmrtk\" (UID: \"5d5e322d-aa8b-4f36-8257-09d1ccab02c6\") " pod="openstack/dnsmasq-dns-764c5664d7-gmrtk" Jan 26 17:07:08 crc kubenswrapper[4754]: I0126 17:07:08.300659 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5d5e322d-aa8b-4f36-8257-09d1ccab02c6-ovsdbserver-sb\") pod \"dnsmasq-dns-764c5664d7-gmrtk\" (UID: \"5d5e322d-aa8b-4f36-8257-09d1ccab02c6\") " pod="openstack/dnsmasq-dns-764c5664d7-gmrtk" Jan 26 17:07:08 crc kubenswrapper[4754]: I0126 17:07:08.300816 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6pczz\" (UniqueName: \"kubernetes.io/projected/5d5e322d-aa8b-4f36-8257-09d1ccab02c6-kube-api-access-6pczz\") pod \"dnsmasq-dns-764c5664d7-gmrtk\" (UID: \"5d5e322d-aa8b-4f36-8257-09d1ccab02c6\") " pod="openstack/dnsmasq-dns-764c5664d7-gmrtk" Jan 26 17:07:08 crc kubenswrapper[4754]: I0126 17:07:08.300966 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5d5e322d-aa8b-4f36-8257-09d1ccab02c6-dns-swift-storage-0\") pod \"dnsmasq-dns-764c5664d7-gmrtk\" (UID: \"5d5e322d-aa8b-4f36-8257-09d1ccab02c6\") " pod="openstack/dnsmasq-dns-764c5664d7-gmrtk" Jan 26 17:07:08 crc kubenswrapper[4754]: I0126 17:07:08.301084 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5d5e322d-aa8b-4f36-8257-09d1ccab02c6-config\") pod \"dnsmasq-dns-764c5664d7-gmrtk\" (UID: \"5d5e322d-aa8b-4f36-8257-09d1ccab02c6\") " pod="openstack/dnsmasq-dns-764c5664d7-gmrtk" Jan 26 17:07:08 crc kubenswrapper[4754]: I0126 17:07:08.402173 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5d5e322d-aa8b-4f36-8257-09d1ccab02c6-config\") pod \"dnsmasq-dns-764c5664d7-gmrtk\" (UID: \"5d5e322d-aa8b-4f36-8257-09d1ccab02c6\") " pod="openstack/dnsmasq-dns-764c5664d7-gmrtk" Jan 26 17:07:08 crc kubenswrapper[4754]: I0126 17:07:08.402474 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5d5e322d-aa8b-4f36-8257-09d1ccab02c6-ovsdbserver-nb\") pod \"dnsmasq-dns-764c5664d7-gmrtk\" (UID: \"5d5e322d-aa8b-4f36-8257-09d1ccab02c6\") " pod="openstack/dnsmasq-dns-764c5664d7-gmrtk" Jan 26 17:07:08 crc kubenswrapper[4754]: I0126 17:07:08.402501 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5d5e322d-aa8b-4f36-8257-09d1ccab02c6-dns-svc\") pod \"dnsmasq-dns-764c5664d7-gmrtk\" (UID: \"5d5e322d-aa8b-4f36-8257-09d1ccab02c6\") " pod="openstack/dnsmasq-dns-764c5664d7-gmrtk" Jan 26 17:07:08 crc kubenswrapper[4754]: I0126 17:07:08.402537 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5d5e322d-aa8b-4f36-8257-09d1ccab02c6-ovsdbserver-sb\") pod \"dnsmasq-dns-764c5664d7-gmrtk\" (UID: \"5d5e322d-aa8b-4f36-8257-09d1ccab02c6\") " pod="openstack/dnsmasq-dns-764c5664d7-gmrtk" Jan 26 17:07:08 crc kubenswrapper[4754]: I0126 17:07:08.402558 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6pczz\" (UniqueName: \"kubernetes.io/projected/5d5e322d-aa8b-4f36-8257-09d1ccab02c6-kube-api-access-6pczz\") pod \"dnsmasq-dns-764c5664d7-gmrtk\" (UID: \"5d5e322d-aa8b-4f36-8257-09d1ccab02c6\") " pod="openstack/dnsmasq-dns-764c5664d7-gmrtk" Jan 26 17:07:08 crc kubenswrapper[4754]: I0126 17:07:08.402627 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5d5e322d-aa8b-4f36-8257-09d1ccab02c6-dns-swift-storage-0\") pod \"dnsmasq-dns-764c5664d7-gmrtk\" (UID: \"5d5e322d-aa8b-4f36-8257-09d1ccab02c6\") " pod="openstack/dnsmasq-dns-764c5664d7-gmrtk" Jan 26 17:07:08 crc kubenswrapper[4754]: I0126 17:07:08.404526 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5d5e322d-aa8b-4f36-8257-09d1ccab02c6-config\") pod \"dnsmasq-dns-764c5664d7-gmrtk\" (UID: \"5d5e322d-aa8b-4f36-8257-09d1ccab02c6\") " pod="openstack/dnsmasq-dns-764c5664d7-gmrtk" Jan 26 17:07:08 crc kubenswrapper[4754]: I0126 17:07:08.404925 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5d5e322d-aa8b-4f36-8257-09d1ccab02c6-dns-swift-storage-0\") pod \"dnsmasq-dns-764c5664d7-gmrtk\" (UID: \"5d5e322d-aa8b-4f36-8257-09d1ccab02c6\") " pod="openstack/dnsmasq-dns-764c5664d7-gmrtk" Jan 26 17:07:08 crc kubenswrapper[4754]: I0126 17:07:08.405170 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5d5e322d-aa8b-4f36-8257-09d1ccab02c6-ovsdbserver-nb\") pod \"dnsmasq-dns-764c5664d7-gmrtk\" (UID: \"5d5e322d-aa8b-4f36-8257-09d1ccab02c6\") " pod="openstack/dnsmasq-dns-764c5664d7-gmrtk" Jan 26 17:07:08 crc kubenswrapper[4754]: I0126 17:07:08.405411 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5d5e322d-aa8b-4f36-8257-09d1ccab02c6-ovsdbserver-sb\") pod \"dnsmasq-dns-764c5664d7-gmrtk\" (UID: \"5d5e322d-aa8b-4f36-8257-09d1ccab02c6\") " pod="openstack/dnsmasq-dns-764c5664d7-gmrtk" Jan 26 17:07:08 crc kubenswrapper[4754]: I0126 17:07:08.418118 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5d5e322d-aa8b-4f36-8257-09d1ccab02c6-dns-svc\") pod \"dnsmasq-dns-764c5664d7-gmrtk\" (UID: \"5d5e322d-aa8b-4f36-8257-09d1ccab02c6\") " pod="openstack/dnsmasq-dns-764c5664d7-gmrtk" Jan 26 17:07:08 crc kubenswrapper[4754]: I0126 17:07:08.433561 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-899gb"] Jan 26 17:07:08 crc kubenswrapper[4754]: I0126 17:07:08.434814 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-899gb" Jan 26 17:07:08 crc kubenswrapper[4754]: I0126 17:07:08.443731 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6pczz\" (UniqueName: \"kubernetes.io/projected/5d5e322d-aa8b-4f36-8257-09d1ccab02c6-kube-api-access-6pczz\") pod \"dnsmasq-dns-764c5664d7-gmrtk\" (UID: \"5d5e322d-aa8b-4f36-8257-09d1ccab02c6\") " pod="openstack/dnsmasq-dns-764c5664d7-gmrtk" Jan 26 17:07:08 crc kubenswrapper[4754]: I0126 17:07:08.463468 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-899gb"] Jan 26 17:07:08 crc kubenswrapper[4754]: I0126 17:07:08.501642 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-764c5664d7-gmrtk" Jan 26 17:07:08 crc kubenswrapper[4754]: I0126 17:07:08.504499 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d4102ba6-b5f4-40a7-8551-300924ddba03-operator-scripts\") pod \"cinder-db-create-899gb\" (UID: \"d4102ba6-b5f4-40a7-8551-300924ddba03\") " pod="openstack/cinder-db-create-899gb" Jan 26 17:07:08 crc kubenswrapper[4754]: I0126 17:07:08.504556 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mf9jt\" (UniqueName: \"kubernetes.io/projected/d4102ba6-b5f4-40a7-8551-300924ddba03-kube-api-access-mf9jt\") pod \"cinder-db-create-899gb\" (UID: \"d4102ba6-b5f4-40a7-8551-300924ddba03\") " pod="openstack/cinder-db-create-899gb" Jan 26 17:07:08 crc kubenswrapper[4754]: I0126 17:07:08.527019 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-dcjt9"] Jan 26 17:07:08 crc kubenswrapper[4754]: I0126 17:07:08.528007 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-dcjt9" Jan 26 17:07:08 crc kubenswrapper[4754]: I0126 17:07:08.549950 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-dcjt9"] Jan 26 17:07:08 crc kubenswrapper[4754]: I0126 17:07:08.595372 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-b037-account-create-update-sb2vx"] Jan 26 17:07:08 crc kubenswrapper[4754]: I0126 17:07:08.596588 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-b037-account-create-update-sb2vx" Jan 26 17:07:08 crc kubenswrapper[4754]: I0126 17:07:08.600031 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Jan 26 17:07:08 crc kubenswrapper[4754]: I0126 17:07:08.605456 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mf9jt\" (UniqueName: \"kubernetes.io/projected/d4102ba6-b5f4-40a7-8551-300924ddba03-kube-api-access-mf9jt\") pod \"cinder-db-create-899gb\" (UID: \"d4102ba6-b5f4-40a7-8551-300924ddba03\") " pod="openstack/cinder-db-create-899gb" Jan 26 17:07:08 crc kubenswrapper[4754]: I0126 17:07:08.606068 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xvddn\" (UniqueName: \"kubernetes.io/projected/83bdc38e-6bbc-4ea8-b57c-9942707b4071-kube-api-access-xvddn\") pod \"barbican-db-create-dcjt9\" (UID: \"83bdc38e-6bbc-4ea8-b57c-9942707b4071\") " pod="openstack/barbican-db-create-dcjt9" Jan 26 17:07:08 crc kubenswrapper[4754]: I0126 17:07:08.606108 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d4102ba6-b5f4-40a7-8551-300924ddba03-operator-scripts\") pod \"cinder-db-create-899gb\" (UID: \"d4102ba6-b5f4-40a7-8551-300924ddba03\") " pod="openstack/cinder-db-create-899gb" Jan 26 17:07:08 crc kubenswrapper[4754]: I0126 17:07:08.606134 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/83bdc38e-6bbc-4ea8-b57c-9942707b4071-operator-scripts\") pod \"barbican-db-create-dcjt9\" (UID: \"83bdc38e-6bbc-4ea8-b57c-9942707b4071\") " pod="openstack/barbican-db-create-dcjt9" Jan 26 17:07:08 crc kubenswrapper[4754]: I0126 17:07:08.607329 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d4102ba6-b5f4-40a7-8551-300924ddba03-operator-scripts\") pod \"cinder-db-create-899gb\" (UID: \"d4102ba6-b5f4-40a7-8551-300924ddba03\") " pod="openstack/cinder-db-create-899gb" Jan 26 17:07:08 crc kubenswrapper[4754]: I0126 17:07:08.623541 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-b037-account-create-update-sb2vx"] Jan 26 17:07:08 crc kubenswrapper[4754]: I0126 17:07:08.635370 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mf9jt\" (UniqueName: \"kubernetes.io/projected/d4102ba6-b5f4-40a7-8551-300924ddba03-kube-api-access-mf9jt\") pod \"cinder-db-create-899gb\" (UID: \"d4102ba6-b5f4-40a7-8551-300924ddba03\") " pod="openstack/cinder-db-create-899gb" Jan 26 17:07:08 crc kubenswrapper[4754]: I0126 17:07:08.706072 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-92jjf"] Jan 26 17:07:08 crc kubenswrapper[4754]: I0126 17:07:08.706998 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-92jjf" Jan 26 17:07:08 crc kubenswrapper[4754]: I0126 17:07:08.707901 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/22b9c223-a293-4248-a32b-36498fa1a43a-operator-scripts\") pod \"cinder-b037-account-create-update-sb2vx\" (UID: \"22b9c223-a293-4248-a32b-36498fa1a43a\") " pod="openstack/cinder-b037-account-create-update-sb2vx" Jan 26 17:07:08 crc kubenswrapper[4754]: I0126 17:07:08.707982 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xvddn\" (UniqueName: \"kubernetes.io/projected/83bdc38e-6bbc-4ea8-b57c-9942707b4071-kube-api-access-xvddn\") pod \"barbican-db-create-dcjt9\" (UID: \"83bdc38e-6bbc-4ea8-b57c-9942707b4071\") " pod="openstack/barbican-db-create-dcjt9" Jan 26 17:07:08 crc kubenswrapper[4754]: I0126 17:07:08.708005 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/83bdc38e-6bbc-4ea8-b57c-9942707b4071-operator-scripts\") pod \"barbican-db-create-dcjt9\" (UID: \"83bdc38e-6bbc-4ea8-b57c-9942707b4071\") " pod="openstack/barbican-db-create-dcjt9" Jan 26 17:07:08 crc kubenswrapper[4754]: I0126 17:07:08.708045 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x5982\" (UniqueName: \"kubernetes.io/projected/22b9c223-a293-4248-a32b-36498fa1a43a-kube-api-access-x5982\") pod \"cinder-b037-account-create-update-sb2vx\" (UID: \"22b9c223-a293-4248-a32b-36498fa1a43a\") " pod="openstack/cinder-b037-account-create-update-sb2vx" Jan 26 17:07:08 crc kubenswrapper[4754]: I0126 17:07:08.708739 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/83bdc38e-6bbc-4ea8-b57c-9942707b4071-operator-scripts\") pod \"barbican-db-create-dcjt9\" (UID: \"83bdc38e-6bbc-4ea8-b57c-9942707b4071\") " pod="openstack/barbican-db-create-dcjt9" Jan 26 17:07:08 crc kubenswrapper[4754]: I0126 17:07:08.720715 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-92jjf"] Jan 26 17:07:08 crc kubenswrapper[4754]: I0126 17:07:08.729956 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xvddn\" (UniqueName: \"kubernetes.io/projected/83bdc38e-6bbc-4ea8-b57c-9942707b4071-kube-api-access-xvddn\") pod \"barbican-db-create-dcjt9\" (UID: \"83bdc38e-6bbc-4ea8-b57c-9942707b4071\") " pod="openstack/barbican-db-create-dcjt9" Jan 26 17:07:08 crc kubenswrapper[4754]: I0126 17:07:08.804745 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-68nzr"] Jan 26 17:07:08 crc kubenswrapper[4754]: I0126 17:07:08.806376 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-68nzr" Jan 26 17:07:08 crc kubenswrapper[4754]: I0126 17:07:08.810166 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Jan 26 17:07:08 crc kubenswrapper[4754]: I0126 17:07:08.810343 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Jan 26 17:07:08 crc kubenswrapper[4754]: I0126 17:07:08.811029 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/22b9c223-a293-4248-a32b-36498fa1a43a-operator-scripts\") pod \"cinder-b037-account-create-update-sb2vx\" (UID: \"22b9c223-a293-4248-a32b-36498fa1a43a\") " pod="openstack/cinder-b037-account-create-update-sb2vx" Jan 26 17:07:08 crc kubenswrapper[4754]: I0126 17:07:08.811107 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qgf5s\" (UniqueName: \"kubernetes.io/projected/0bc80e48-26f1-47c6-9fc2-95057bfd06b9-kube-api-access-qgf5s\") pod \"neutron-db-create-92jjf\" (UID: \"0bc80e48-26f1-47c6-9fc2-95057bfd06b9\") " pod="openstack/neutron-db-create-92jjf" Jan 26 17:07:08 crc kubenswrapper[4754]: I0126 17:07:08.811283 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0bc80e48-26f1-47c6-9fc2-95057bfd06b9-operator-scripts\") pod \"neutron-db-create-92jjf\" (UID: \"0bc80e48-26f1-47c6-9fc2-95057bfd06b9\") " pod="openstack/neutron-db-create-92jjf" Jan 26 17:07:08 crc kubenswrapper[4754]: I0126 17:07:08.811305 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x5982\" (UniqueName: \"kubernetes.io/projected/22b9c223-a293-4248-a32b-36498fa1a43a-kube-api-access-x5982\") pod \"cinder-b037-account-create-update-sb2vx\" (UID: \"22b9c223-a293-4248-a32b-36498fa1a43a\") " pod="openstack/cinder-b037-account-create-update-sb2vx" Jan 26 17:07:08 crc kubenswrapper[4754]: I0126 17:07:08.812188 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/22b9c223-a293-4248-a32b-36498fa1a43a-operator-scripts\") pod \"cinder-b037-account-create-update-sb2vx\" (UID: \"22b9c223-a293-4248-a32b-36498fa1a43a\") " pod="openstack/cinder-b037-account-create-update-sb2vx" Jan 26 17:07:08 crc kubenswrapper[4754]: I0126 17:07:08.812936 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-r42mh" Jan 26 17:07:08 crc kubenswrapper[4754]: I0126 17:07:08.813723 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-68nzr"] Jan 26 17:07:08 crc kubenswrapper[4754]: I0126 17:07:08.816064 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-899gb" Jan 26 17:07:08 crc kubenswrapper[4754]: I0126 17:07:08.816247 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Jan 26 17:07:08 crc kubenswrapper[4754]: I0126 17:07:08.826561 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-c8dc-account-create-update-w9qwp"] Jan 26 17:07:08 crc kubenswrapper[4754]: I0126 17:07:08.831822 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-c8dc-account-create-update-w9qwp" Jan 26 17:07:08 crc kubenswrapper[4754]: I0126 17:07:08.833953 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-c8dc-account-create-update-w9qwp"] Jan 26 17:07:08 crc kubenswrapper[4754]: I0126 17:07:08.835363 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Jan 26 17:07:08 crc kubenswrapper[4754]: I0126 17:07:08.856831 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x5982\" (UniqueName: \"kubernetes.io/projected/22b9c223-a293-4248-a32b-36498fa1a43a-kube-api-access-x5982\") pod \"cinder-b037-account-create-update-sb2vx\" (UID: \"22b9c223-a293-4248-a32b-36498fa1a43a\") " pod="openstack/cinder-b037-account-create-update-sb2vx" Jan 26 17:07:08 crc kubenswrapper[4754]: I0126 17:07:08.913001 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2cb8378d-e281-4399-a82b-4fe351fcac16-operator-scripts\") pod \"barbican-c8dc-account-create-update-w9qwp\" (UID: \"2cb8378d-e281-4399-a82b-4fe351fcac16\") " pod="openstack/barbican-c8dc-account-create-update-w9qwp" Jan 26 17:07:08 crc kubenswrapper[4754]: I0126 17:07:08.913075 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0bc80e48-26f1-47c6-9fc2-95057bfd06b9-operator-scripts\") pod \"neutron-db-create-92jjf\" (UID: \"0bc80e48-26f1-47c6-9fc2-95057bfd06b9\") " pod="openstack/neutron-db-create-92jjf" Jan 26 17:07:08 crc kubenswrapper[4754]: I0126 17:07:08.913125 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/35a2e341-10fc-48ca-8c99-65f1664cbba5-config-data\") pod \"keystone-db-sync-68nzr\" (UID: \"35a2e341-10fc-48ca-8c99-65f1664cbba5\") " pod="openstack/keystone-db-sync-68nzr" Jan 26 17:07:08 crc kubenswrapper[4754]: I0126 17:07:08.913156 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r2wf8\" (UniqueName: \"kubernetes.io/projected/2cb8378d-e281-4399-a82b-4fe351fcac16-kube-api-access-r2wf8\") pod \"barbican-c8dc-account-create-update-w9qwp\" (UID: \"2cb8378d-e281-4399-a82b-4fe351fcac16\") " pod="openstack/barbican-c8dc-account-create-update-w9qwp" Jan 26 17:07:08 crc kubenswrapper[4754]: I0126 17:07:08.913220 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qgf5s\" (UniqueName: \"kubernetes.io/projected/0bc80e48-26f1-47c6-9fc2-95057bfd06b9-kube-api-access-qgf5s\") pod \"neutron-db-create-92jjf\" (UID: \"0bc80e48-26f1-47c6-9fc2-95057bfd06b9\") " pod="openstack/neutron-db-create-92jjf" Jan 26 17:07:08 crc kubenswrapper[4754]: I0126 17:07:08.913244 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-95d28\" (UniqueName: \"kubernetes.io/projected/35a2e341-10fc-48ca-8c99-65f1664cbba5-kube-api-access-95d28\") pod \"keystone-db-sync-68nzr\" (UID: \"35a2e341-10fc-48ca-8c99-65f1664cbba5\") " pod="openstack/keystone-db-sync-68nzr" Jan 26 17:07:08 crc kubenswrapper[4754]: I0126 17:07:08.913294 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35a2e341-10fc-48ca-8c99-65f1664cbba5-combined-ca-bundle\") pod \"keystone-db-sync-68nzr\" (UID: \"35a2e341-10fc-48ca-8c99-65f1664cbba5\") " pod="openstack/keystone-db-sync-68nzr" Jan 26 17:07:08 crc kubenswrapper[4754]: I0126 17:07:08.914151 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0bc80e48-26f1-47c6-9fc2-95057bfd06b9-operator-scripts\") pod \"neutron-db-create-92jjf\" (UID: \"0bc80e48-26f1-47c6-9fc2-95057bfd06b9\") " pod="openstack/neutron-db-create-92jjf" Jan 26 17:07:08 crc kubenswrapper[4754]: I0126 17:07:08.916753 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-764c5664d7-gmrtk"] Jan 26 17:07:08 crc kubenswrapper[4754]: I0126 17:07:08.923498 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-dcjt9" Jan 26 17:07:08 crc kubenswrapper[4754]: I0126 17:07:08.930737 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qgf5s\" (UniqueName: \"kubernetes.io/projected/0bc80e48-26f1-47c6-9fc2-95057bfd06b9-kube-api-access-qgf5s\") pod \"neutron-db-create-92jjf\" (UID: \"0bc80e48-26f1-47c6-9fc2-95057bfd06b9\") " pod="openstack/neutron-db-create-92jjf" Jan 26 17:07:08 crc kubenswrapper[4754]: I0126 17:07:08.984442 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-b037-account-create-update-sb2vx" Jan 26 17:07:09 crc kubenswrapper[4754]: I0126 17:07:09.005965 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-c7d7-account-create-update-qjftw"] Jan 26 17:07:09 crc kubenswrapper[4754]: I0126 17:07:09.006963 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-c7d7-account-create-update-qjftw" Jan 26 17:07:09 crc kubenswrapper[4754]: I0126 17:07:09.011128 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Jan 26 17:07:09 crc kubenswrapper[4754]: I0126 17:07:09.016569 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35a2e341-10fc-48ca-8c99-65f1664cbba5-combined-ca-bundle\") pod \"keystone-db-sync-68nzr\" (UID: \"35a2e341-10fc-48ca-8c99-65f1664cbba5\") " pod="openstack/keystone-db-sync-68nzr" Jan 26 17:07:09 crc kubenswrapper[4754]: I0126 17:07:09.016641 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d5f8d\" (UniqueName: \"kubernetes.io/projected/fc67dbf2-5b6b-44cb-a5d4-8b7e32fd4a87-kube-api-access-d5f8d\") pod \"neutron-c7d7-account-create-update-qjftw\" (UID: \"fc67dbf2-5b6b-44cb-a5d4-8b7e32fd4a87\") " pod="openstack/neutron-c7d7-account-create-update-qjftw" Jan 26 17:07:09 crc kubenswrapper[4754]: I0126 17:07:09.016696 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2cb8378d-e281-4399-a82b-4fe351fcac16-operator-scripts\") pod \"barbican-c8dc-account-create-update-w9qwp\" (UID: \"2cb8378d-e281-4399-a82b-4fe351fcac16\") " pod="openstack/barbican-c8dc-account-create-update-w9qwp" Jan 26 17:07:09 crc kubenswrapper[4754]: I0126 17:07:09.016754 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/35a2e341-10fc-48ca-8c99-65f1664cbba5-config-data\") pod \"keystone-db-sync-68nzr\" (UID: \"35a2e341-10fc-48ca-8c99-65f1664cbba5\") " pod="openstack/keystone-db-sync-68nzr" Jan 26 17:07:09 crc kubenswrapper[4754]: I0126 17:07:09.016777 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r2wf8\" (UniqueName: \"kubernetes.io/projected/2cb8378d-e281-4399-a82b-4fe351fcac16-kube-api-access-r2wf8\") pod \"barbican-c8dc-account-create-update-w9qwp\" (UID: \"2cb8378d-e281-4399-a82b-4fe351fcac16\") " pod="openstack/barbican-c8dc-account-create-update-w9qwp" Jan 26 17:07:09 crc kubenswrapper[4754]: I0126 17:07:09.016814 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fc67dbf2-5b6b-44cb-a5d4-8b7e32fd4a87-operator-scripts\") pod \"neutron-c7d7-account-create-update-qjftw\" (UID: \"fc67dbf2-5b6b-44cb-a5d4-8b7e32fd4a87\") " pod="openstack/neutron-c7d7-account-create-update-qjftw" Jan 26 17:07:09 crc kubenswrapper[4754]: I0126 17:07:09.016838 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-95d28\" (UniqueName: \"kubernetes.io/projected/35a2e341-10fc-48ca-8c99-65f1664cbba5-kube-api-access-95d28\") pod \"keystone-db-sync-68nzr\" (UID: \"35a2e341-10fc-48ca-8c99-65f1664cbba5\") " pod="openstack/keystone-db-sync-68nzr" Jan 26 17:07:09 crc kubenswrapper[4754]: I0126 17:07:09.017862 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2cb8378d-e281-4399-a82b-4fe351fcac16-operator-scripts\") pod \"barbican-c8dc-account-create-update-w9qwp\" (UID: \"2cb8378d-e281-4399-a82b-4fe351fcac16\") " pod="openstack/barbican-c8dc-account-create-update-w9qwp" Jan 26 17:07:09 crc kubenswrapper[4754]: I0126 17:07:09.021489 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35a2e341-10fc-48ca-8c99-65f1664cbba5-combined-ca-bundle\") pod \"keystone-db-sync-68nzr\" (UID: \"35a2e341-10fc-48ca-8c99-65f1664cbba5\") " pod="openstack/keystone-db-sync-68nzr" Jan 26 17:07:09 crc kubenswrapper[4754]: I0126 17:07:09.024177 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/35a2e341-10fc-48ca-8c99-65f1664cbba5-config-data\") pod \"keystone-db-sync-68nzr\" (UID: \"35a2e341-10fc-48ca-8c99-65f1664cbba5\") " pod="openstack/keystone-db-sync-68nzr" Jan 26 17:07:09 crc kubenswrapper[4754]: I0126 17:07:09.025643 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-c7d7-account-create-update-qjftw"] Jan 26 17:07:09 crc kubenswrapper[4754]: I0126 17:07:09.048548 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r2wf8\" (UniqueName: \"kubernetes.io/projected/2cb8378d-e281-4399-a82b-4fe351fcac16-kube-api-access-r2wf8\") pod \"barbican-c8dc-account-create-update-w9qwp\" (UID: \"2cb8378d-e281-4399-a82b-4fe351fcac16\") " pod="openstack/barbican-c8dc-account-create-update-w9qwp" Jan 26 17:07:09 crc kubenswrapper[4754]: I0126 17:07:09.058130 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-95d28\" (UniqueName: \"kubernetes.io/projected/35a2e341-10fc-48ca-8c99-65f1664cbba5-kube-api-access-95d28\") pod \"keystone-db-sync-68nzr\" (UID: \"35a2e341-10fc-48ca-8c99-65f1664cbba5\") " pod="openstack/keystone-db-sync-68nzr" Jan 26 17:07:09 crc kubenswrapper[4754]: I0126 17:07:09.073554 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-92jjf" Jan 26 17:07:09 crc kubenswrapper[4754]: I0126 17:07:09.118356 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d5f8d\" (UniqueName: \"kubernetes.io/projected/fc67dbf2-5b6b-44cb-a5d4-8b7e32fd4a87-kube-api-access-d5f8d\") pod \"neutron-c7d7-account-create-update-qjftw\" (UID: \"fc67dbf2-5b6b-44cb-a5d4-8b7e32fd4a87\") " pod="openstack/neutron-c7d7-account-create-update-qjftw" Jan 26 17:07:09 crc kubenswrapper[4754]: I0126 17:07:09.118472 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fc67dbf2-5b6b-44cb-a5d4-8b7e32fd4a87-operator-scripts\") pod \"neutron-c7d7-account-create-update-qjftw\" (UID: \"fc67dbf2-5b6b-44cb-a5d4-8b7e32fd4a87\") " pod="openstack/neutron-c7d7-account-create-update-qjftw" Jan 26 17:07:09 crc kubenswrapper[4754]: I0126 17:07:09.119114 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fc67dbf2-5b6b-44cb-a5d4-8b7e32fd4a87-operator-scripts\") pod \"neutron-c7d7-account-create-update-qjftw\" (UID: \"fc67dbf2-5b6b-44cb-a5d4-8b7e32fd4a87\") " pod="openstack/neutron-c7d7-account-create-update-qjftw" Jan 26 17:07:09 crc kubenswrapper[4754]: I0126 17:07:09.144282 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d5f8d\" (UniqueName: \"kubernetes.io/projected/fc67dbf2-5b6b-44cb-a5d4-8b7e32fd4a87-kube-api-access-d5f8d\") pod \"neutron-c7d7-account-create-update-qjftw\" (UID: \"fc67dbf2-5b6b-44cb-a5d4-8b7e32fd4a87\") " pod="openstack/neutron-c7d7-account-create-update-qjftw" Jan 26 17:07:09 crc kubenswrapper[4754]: I0126 17:07:09.181136 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-68nzr" Jan 26 17:07:09 crc kubenswrapper[4754]: I0126 17:07:09.195383 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-c8dc-account-create-update-w9qwp" Jan 26 17:07:09 crc kubenswrapper[4754]: I0126 17:07:09.338678 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-c7d7-account-create-update-qjftw" Jan 26 17:07:09 crc kubenswrapper[4754]: I0126 17:07:09.383968 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-899gb"] Jan 26 17:07:09 crc kubenswrapper[4754]: W0126 17:07:09.405813 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd4102ba6_b5f4_40a7_8551_300924ddba03.slice/crio-49bbcfa99aa730783f9874850536db6520ba77495a68541e0b096be05ca686f0 WatchSource:0}: Error finding container 49bbcfa99aa730783f9874850536db6520ba77495a68541e0b096be05ca686f0: Status 404 returned error can't find the container with id 49bbcfa99aa730783f9874850536db6520ba77495a68541e0b096be05ca686f0 Jan 26 17:07:09 crc kubenswrapper[4754]: I0126 17:07:09.439491 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/root-account-create-update-bp5dl"] Jan 26 17:07:09 crc kubenswrapper[4754]: I0126 17:07:09.445759 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/root-account-create-update-bp5dl"] Jan 26 17:07:09 crc kubenswrapper[4754]: I0126 17:07:09.524652 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-dcjt9"] Jan 26 17:07:09 crc kubenswrapper[4754]: I0126 17:07:09.630705 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-b037-account-create-update-sb2vx"] Jan 26 17:07:09 crc kubenswrapper[4754]: W0126 17:07:09.633969 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod22b9c223_a293_4248_a32b_36498fa1a43a.slice/crio-151cf2767dee7f8025841de1b9e99b834927bed6083f57d05cf63f2462e5148d WatchSource:0}: Error finding container 151cf2767dee7f8025841de1b9e99b834927bed6083f57d05cf63f2462e5148d: Status 404 returned error can't find the container with id 151cf2767dee7f8025841de1b9e99b834927bed6083f57d05cf63f2462e5148d Jan 26 17:07:09 crc kubenswrapper[4754]: I0126 17:07:09.752290 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-92jjf"] Jan 26 17:07:09 crc kubenswrapper[4754]: I0126 17:07:09.792069 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4ad7fbd5-cd5f-4637-b2a1-8c71df037ed6" path="/var/lib/kubelet/pods/4ad7fbd5-cd5f-4637-b2a1-8c71df037ed6/volumes" Jan 26 17:07:09 crc kubenswrapper[4754]: I0126 17:07:09.792916 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="944939e4-26fe-4583-9c20-4092f6f5dab3" path="/var/lib/kubelet/pods/944939e4-26fe-4583-9c20-4092f6f5dab3/volumes" Jan 26 17:07:09 crc kubenswrapper[4754]: I0126 17:07:09.800463 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-dcjt9" event={"ID":"83bdc38e-6bbc-4ea8-b57c-9942707b4071","Type":"ContainerStarted","Data":"fba3e58ff4d0feddeb997a2011412fbfbb2a4105f0fe85354244b9fa3f9160a7"} Jan 26 17:07:09 crc kubenswrapper[4754]: I0126 17:07:09.801574 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-899gb" event={"ID":"d4102ba6-b5f4-40a7-8551-300924ddba03","Type":"ContainerStarted","Data":"49bbcfa99aa730783f9874850536db6520ba77495a68541e0b096be05ca686f0"} Jan 26 17:07:09 crc kubenswrapper[4754]: I0126 17:07:09.804363 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-764c5664d7-gmrtk" event={"ID":"5d5e322d-aa8b-4f36-8257-09d1ccab02c6","Type":"ContainerStarted","Data":"b8754f9eb16721fb33c204af711238e3f31101e26d6e8971d66e4f6bd4e89b48"} Jan 26 17:07:09 crc kubenswrapper[4754]: I0126 17:07:09.806038 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-92jjf" event={"ID":"0bc80e48-26f1-47c6-9fc2-95057bfd06b9","Type":"ContainerStarted","Data":"1205f95491a521c08c8765c70e234af7423e52e6d2982e01ab0df54ce13cfc80"} Jan 26 17:07:09 crc kubenswrapper[4754]: I0126 17:07:09.806848 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-b037-account-create-update-sb2vx" event={"ID":"22b9c223-a293-4248-a32b-36498fa1a43a","Type":"ContainerStarted","Data":"151cf2767dee7f8025841de1b9e99b834927bed6083f57d05cf63f2462e5148d"} Jan 26 17:07:09 crc kubenswrapper[4754]: W0126 17:07:09.830232 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2cb8378d_e281_4399_a82b_4fe351fcac16.slice/crio-ed0c669695f66a9627fb65fd7b1f83737a368d30fbe5d6436428360dee2d9da4 WatchSource:0}: Error finding container ed0c669695f66a9627fb65fd7b1f83737a368d30fbe5d6436428360dee2d9da4: Status 404 returned error can't find the container with id ed0c669695f66a9627fb65fd7b1f83737a368d30fbe5d6436428360dee2d9da4 Jan 26 17:07:09 crc kubenswrapper[4754]: I0126 17:07:09.832518 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-c8dc-account-create-update-w9qwp"] Jan 26 17:07:09 crc kubenswrapper[4754]: I0126 17:07:09.848215 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-68nzr"] Jan 26 17:07:09 crc kubenswrapper[4754]: I0126 17:07:09.941401 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-c7d7-account-create-update-qjftw"] Jan 26 17:07:10 crc kubenswrapper[4754]: I0126 17:07:10.818818 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-dcjt9" event={"ID":"83bdc38e-6bbc-4ea8-b57c-9942707b4071","Type":"ContainerStarted","Data":"528723486d42beba27f3015599fc78a7ff5226695e282daaca10290a51dc0fbe"} Jan 26 17:07:10 crc kubenswrapper[4754]: I0126 17:07:10.820786 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-899gb" event={"ID":"d4102ba6-b5f4-40a7-8551-300924ddba03","Type":"ContainerStarted","Data":"7b2c6f588352f73fbb3612f888d11cd5358bb977bdf45587a044dab856652069"} Jan 26 17:07:10 crc kubenswrapper[4754]: I0126 17:07:10.822435 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-c8dc-account-create-update-w9qwp" event={"ID":"2cb8378d-e281-4399-a82b-4fe351fcac16","Type":"ContainerStarted","Data":"fce952ed99ebf0ec3ce4c1388209744f2a30e0a41cc72ee2c1b854e6dd140583"} Jan 26 17:07:10 crc kubenswrapper[4754]: I0126 17:07:10.822465 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-c8dc-account-create-update-w9qwp" event={"ID":"2cb8378d-e281-4399-a82b-4fe351fcac16","Type":"ContainerStarted","Data":"ed0c669695f66a9627fb65fd7b1f83737a368d30fbe5d6436428360dee2d9da4"} Jan 26 17:07:10 crc kubenswrapper[4754]: I0126 17:07:10.824364 4754 generic.go:334] "Generic (PLEG): container finished" podID="5d5e322d-aa8b-4f36-8257-09d1ccab02c6" containerID="702220854f984609ce86913f04217dceba69fcb60a0a87776849f13e43a5d464" exitCode=0 Jan 26 17:07:10 crc kubenswrapper[4754]: I0126 17:07:10.824438 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-764c5664d7-gmrtk" event={"ID":"5d5e322d-aa8b-4f36-8257-09d1ccab02c6","Type":"ContainerDied","Data":"702220854f984609ce86913f04217dceba69fcb60a0a87776849f13e43a5d464"} Jan 26 17:07:10 crc kubenswrapper[4754]: I0126 17:07:10.825971 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-c7d7-account-create-update-qjftw" event={"ID":"fc67dbf2-5b6b-44cb-a5d4-8b7e32fd4a87","Type":"ContainerStarted","Data":"2b1a9257ec9c280eab069d563f3604849bb5a428f499d0d234618db276339bc8"} Jan 26 17:07:10 crc kubenswrapper[4754]: I0126 17:07:10.825998 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-c7d7-account-create-update-qjftw" event={"ID":"fc67dbf2-5b6b-44cb-a5d4-8b7e32fd4a87","Type":"ContainerStarted","Data":"bd4564da838e319c6a58d8959b606791ba5c9791543c55c9f68cbc9e83533b12"} Jan 26 17:07:10 crc kubenswrapper[4754]: I0126 17:07:10.829237 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-68nzr" event={"ID":"35a2e341-10fc-48ca-8c99-65f1664cbba5","Type":"ContainerStarted","Data":"4044f7375a0edac7261ee9a95007ad80c4c120ca8795b6dbc60d27f19b02a1c9"} Jan 26 17:07:10 crc kubenswrapper[4754]: I0126 17:07:10.832833 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-92jjf" event={"ID":"0bc80e48-26f1-47c6-9fc2-95057bfd06b9","Type":"ContainerStarted","Data":"189d6fed262dc3a4f55ece04db494b95855ed4513e2524c3e6835c1f6bca47b2"} Jan 26 17:07:10 crc kubenswrapper[4754]: I0126 17:07:10.843244 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-b037-account-create-update-sb2vx" event={"ID":"22b9c223-a293-4248-a32b-36498fa1a43a","Type":"ContainerStarted","Data":"c63f8c2db5c0a10f443a10a4a6dca92401a4754f6725760e22c8417fc8de935b"} Jan 26 17:07:10 crc kubenswrapper[4754]: I0126 17:07:10.916693 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-b037-account-create-update-sb2vx" podStartSLOduration=2.916644656 podStartE2EDuration="2.916644656s" podCreationTimestamp="2026-01-26 17:07:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 17:07:10.901791451 +0000 UTC m=+1197.425971915" watchObservedRunningTime="2026-01-26 17:07:10.916644656 +0000 UTC m=+1197.440825090" Jan 26 17:07:11 crc kubenswrapper[4754]: I0126 17:07:11.886603 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-create-899gb" podStartSLOduration=3.886585286 podStartE2EDuration="3.886585286s" podCreationTimestamp="2026-01-26 17:07:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 17:07:11.885740252 +0000 UTC m=+1198.409920696" watchObservedRunningTime="2026-01-26 17:07:11.886585286 +0000 UTC m=+1198.410765720" Jan 26 17:07:11 crc kubenswrapper[4754]: I0126 17:07:11.903302 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-create-92jjf" podStartSLOduration=3.9032698420000003 podStartE2EDuration="3.903269842s" podCreationTimestamp="2026-01-26 17:07:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 17:07:11.900584777 +0000 UTC m=+1198.424765211" watchObservedRunningTime="2026-01-26 17:07:11.903269842 +0000 UTC m=+1198.427450276" Jan 26 17:07:11 crc kubenswrapper[4754]: I0126 17:07:11.940547 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-create-dcjt9" podStartSLOduration=3.9405189529999998 podStartE2EDuration="3.940518953s" podCreationTimestamp="2026-01-26 17:07:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 17:07:11.920326699 +0000 UTC m=+1198.444507133" watchObservedRunningTime="2026-01-26 17:07:11.940518953 +0000 UTC m=+1198.464699387" Jan 26 17:07:11 crc kubenswrapper[4754]: I0126 17:07:11.962947 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-c8dc-account-create-update-w9qwp" podStartSLOduration=3.962910678 podStartE2EDuration="3.962910678s" podCreationTimestamp="2026-01-26 17:07:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 17:07:11.954212315 +0000 UTC m=+1198.478392739" watchObservedRunningTime="2026-01-26 17:07:11.962910678 +0000 UTC m=+1198.487091132" Jan 26 17:07:12 crc kubenswrapper[4754]: I0126 17:07:12.868496 4754 generic.go:334] "Generic (PLEG): container finished" podID="0bc80e48-26f1-47c6-9fc2-95057bfd06b9" containerID="189d6fed262dc3a4f55ece04db494b95855ed4513e2524c3e6835c1f6bca47b2" exitCode=0 Jan 26 17:07:12 crc kubenswrapper[4754]: I0126 17:07:12.868567 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-92jjf" event={"ID":"0bc80e48-26f1-47c6-9fc2-95057bfd06b9","Type":"ContainerDied","Data":"189d6fed262dc3a4f55ece04db494b95855ed4513e2524c3e6835c1f6bca47b2"} Jan 26 17:07:12 crc kubenswrapper[4754]: I0126 17:07:12.870586 4754 generic.go:334] "Generic (PLEG): container finished" podID="83bdc38e-6bbc-4ea8-b57c-9942707b4071" containerID="528723486d42beba27f3015599fc78a7ff5226695e282daaca10290a51dc0fbe" exitCode=0 Jan 26 17:07:12 crc kubenswrapper[4754]: I0126 17:07:12.870717 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-dcjt9" event={"ID":"83bdc38e-6bbc-4ea8-b57c-9942707b4071","Type":"ContainerDied","Data":"528723486d42beba27f3015599fc78a7ff5226695e282daaca10290a51dc0fbe"} Jan 26 17:07:12 crc kubenswrapper[4754]: I0126 17:07:12.872184 4754 generic.go:334] "Generic (PLEG): container finished" podID="d4102ba6-b5f4-40a7-8551-300924ddba03" containerID="7b2c6f588352f73fbb3612f888d11cd5358bb977bdf45587a044dab856652069" exitCode=0 Jan 26 17:07:12 crc kubenswrapper[4754]: I0126 17:07:12.872221 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-899gb" event={"ID":"d4102ba6-b5f4-40a7-8551-300924ddba03","Type":"ContainerDied","Data":"7b2c6f588352f73fbb3612f888d11cd5358bb977bdf45587a044dab856652069"} Jan 26 17:07:13 crc kubenswrapper[4754]: I0126 17:07:13.087843 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/root-account-create-update-bmlgs"] Jan 26 17:07:13 crc kubenswrapper[4754]: I0126 17:07:13.089031 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-bmlgs" Jan 26 17:07:13 crc kubenswrapper[4754]: I0126 17:07:13.091151 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-mariadb-root-db-secret" Jan 26 17:07:13 crc kubenswrapper[4754]: I0126 17:07:13.099558 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-bmlgs"] Jan 26 17:07:13 crc kubenswrapper[4754]: I0126 17:07:13.205859 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f97b956b-11ee-484d-917e-75b5aba780f0-operator-scripts\") pod \"root-account-create-update-bmlgs\" (UID: \"f97b956b-11ee-484d-917e-75b5aba780f0\") " pod="openstack/root-account-create-update-bmlgs" Jan 26 17:07:13 crc kubenswrapper[4754]: I0126 17:07:13.206035 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xbcxr\" (UniqueName: \"kubernetes.io/projected/f97b956b-11ee-484d-917e-75b5aba780f0-kube-api-access-xbcxr\") pod \"root-account-create-update-bmlgs\" (UID: \"f97b956b-11ee-484d-917e-75b5aba780f0\") " pod="openstack/root-account-create-update-bmlgs" Jan 26 17:07:13 crc kubenswrapper[4754]: I0126 17:07:13.307492 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f97b956b-11ee-484d-917e-75b5aba780f0-operator-scripts\") pod \"root-account-create-update-bmlgs\" (UID: \"f97b956b-11ee-484d-917e-75b5aba780f0\") " pod="openstack/root-account-create-update-bmlgs" Jan 26 17:07:13 crc kubenswrapper[4754]: I0126 17:07:13.307595 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xbcxr\" (UniqueName: \"kubernetes.io/projected/f97b956b-11ee-484d-917e-75b5aba780f0-kube-api-access-xbcxr\") pod \"root-account-create-update-bmlgs\" (UID: \"f97b956b-11ee-484d-917e-75b5aba780f0\") " pod="openstack/root-account-create-update-bmlgs" Jan 26 17:07:13 crc kubenswrapper[4754]: I0126 17:07:13.308926 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f97b956b-11ee-484d-917e-75b5aba780f0-operator-scripts\") pod \"root-account-create-update-bmlgs\" (UID: \"f97b956b-11ee-484d-917e-75b5aba780f0\") " pod="openstack/root-account-create-update-bmlgs" Jan 26 17:07:13 crc kubenswrapper[4754]: I0126 17:07:13.338316 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xbcxr\" (UniqueName: \"kubernetes.io/projected/f97b956b-11ee-484d-917e-75b5aba780f0-kube-api-access-xbcxr\") pod \"root-account-create-update-bmlgs\" (UID: \"f97b956b-11ee-484d-917e-75b5aba780f0\") " pod="openstack/root-account-create-update-bmlgs" Jan 26 17:07:13 crc kubenswrapper[4754]: I0126 17:07:13.428778 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-bmlgs" Jan 26 17:07:13 crc kubenswrapper[4754]: I0126 17:07:13.884039 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-764c5664d7-gmrtk" event={"ID":"5d5e322d-aa8b-4f36-8257-09d1ccab02c6","Type":"ContainerStarted","Data":"a9c9a553d41ab84f3ade093e9e2fe01824a18d3f81daba957bbe3e8118796c8e"} Jan 26 17:07:13 crc kubenswrapper[4754]: I0126 17:07:13.887265 4754 generic.go:334] "Generic (PLEG): container finished" podID="fc67dbf2-5b6b-44cb-a5d4-8b7e32fd4a87" containerID="2b1a9257ec9c280eab069d563f3604849bb5a428f499d0d234618db276339bc8" exitCode=0 Jan 26 17:07:13 crc kubenswrapper[4754]: I0126 17:07:13.887313 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-c7d7-account-create-update-qjftw" event={"ID":"fc67dbf2-5b6b-44cb-a5d4-8b7e32fd4a87","Type":"ContainerDied","Data":"2b1a9257ec9c280eab069d563f3604849bb5a428f499d0d234618db276339bc8"} Jan 26 17:07:13 crc kubenswrapper[4754]: I0126 17:07:13.890203 4754 generic.go:334] "Generic (PLEG): container finished" podID="22b9c223-a293-4248-a32b-36498fa1a43a" containerID="c63f8c2db5c0a10f443a10a4a6dca92401a4754f6725760e22c8417fc8de935b" exitCode=0 Jan 26 17:07:13 crc kubenswrapper[4754]: I0126 17:07:13.890330 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-b037-account-create-update-sb2vx" event={"ID":"22b9c223-a293-4248-a32b-36498fa1a43a","Type":"ContainerDied","Data":"c63f8c2db5c0a10f443a10a4a6dca92401a4754f6725760e22c8417fc8de935b"} Jan 26 17:07:13 crc kubenswrapper[4754]: I0126 17:07:13.897555 4754 generic.go:334] "Generic (PLEG): container finished" podID="2cb8378d-e281-4399-a82b-4fe351fcac16" containerID="fce952ed99ebf0ec3ce4c1388209744f2a30e0a41cc72ee2c1b854e6dd140583" exitCode=0 Jan 26 17:07:13 crc kubenswrapper[4754]: I0126 17:07:13.897763 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-c8dc-account-create-update-w9qwp" event={"ID":"2cb8378d-e281-4399-a82b-4fe351fcac16","Type":"ContainerDied","Data":"fce952ed99ebf0ec3ce4c1388209744f2a30e0a41cc72ee2c1b854e6dd140583"} Jan 26 17:07:13 crc kubenswrapper[4754]: I0126 17:07:13.906614 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-bmlgs"] Jan 26 17:07:14 crc kubenswrapper[4754]: I0126 17:07:14.017381 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-mariadb-root-db-secret" Jan 26 17:07:14 crc kubenswrapper[4754]: I0126 17:07:14.298639 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-dcjt9" Jan 26 17:07:14 crc kubenswrapper[4754]: I0126 17:07:14.336045 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xvddn\" (UniqueName: \"kubernetes.io/projected/83bdc38e-6bbc-4ea8-b57c-9942707b4071-kube-api-access-xvddn\") pod \"83bdc38e-6bbc-4ea8-b57c-9942707b4071\" (UID: \"83bdc38e-6bbc-4ea8-b57c-9942707b4071\") " Jan 26 17:07:14 crc kubenswrapper[4754]: I0126 17:07:14.336178 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/83bdc38e-6bbc-4ea8-b57c-9942707b4071-operator-scripts\") pod \"83bdc38e-6bbc-4ea8-b57c-9942707b4071\" (UID: \"83bdc38e-6bbc-4ea8-b57c-9942707b4071\") " Jan 26 17:07:14 crc kubenswrapper[4754]: I0126 17:07:14.337167 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/83bdc38e-6bbc-4ea8-b57c-9942707b4071-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "83bdc38e-6bbc-4ea8-b57c-9942707b4071" (UID: "83bdc38e-6bbc-4ea8-b57c-9942707b4071"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:07:14 crc kubenswrapper[4754]: I0126 17:07:14.349801 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/83bdc38e-6bbc-4ea8-b57c-9942707b4071-kube-api-access-xvddn" (OuterVolumeSpecName: "kube-api-access-xvddn") pod "83bdc38e-6bbc-4ea8-b57c-9942707b4071" (UID: "83bdc38e-6bbc-4ea8-b57c-9942707b4071"). InnerVolumeSpecName "kube-api-access-xvddn". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:07:14 crc kubenswrapper[4754]: I0126 17:07:14.437658 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xvddn\" (UniqueName: \"kubernetes.io/projected/83bdc38e-6bbc-4ea8-b57c-9942707b4071-kube-api-access-xvddn\") on node \"crc\" DevicePath \"\"" Jan 26 17:07:14 crc kubenswrapper[4754]: I0126 17:07:14.437713 4754 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/83bdc38e-6bbc-4ea8-b57c-9942707b4071-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 26 17:07:14 crc kubenswrapper[4754]: I0126 17:07:14.438030 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-92jjf" Jan 26 17:07:14 crc kubenswrapper[4754]: I0126 17:07:14.444839 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-899gb" Jan 26 17:07:14 crc kubenswrapper[4754]: I0126 17:07:14.537973 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qgf5s\" (UniqueName: \"kubernetes.io/projected/0bc80e48-26f1-47c6-9fc2-95057bfd06b9-kube-api-access-qgf5s\") pod \"0bc80e48-26f1-47c6-9fc2-95057bfd06b9\" (UID: \"0bc80e48-26f1-47c6-9fc2-95057bfd06b9\") " Jan 26 17:07:14 crc kubenswrapper[4754]: I0126 17:07:14.538077 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mf9jt\" (UniqueName: \"kubernetes.io/projected/d4102ba6-b5f4-40a7-8551-300924ddba03-kube-api-access-mf9jt\") pod \"d4102ba6-b5f4-40a7-8551-300924ddba03\" (UID: \"d4102ba6-b5f4-40a7-8551-300924ddba03\") " Jan 26 17:07:14 crc kubenswrapper[4754]: I0126 17:07:14.538113 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0bc80e48-26f1-47c6-9fc2-95057bfd06b9-operator-scripts\") pod \"0bc80e48-26f1-47c6-9fc2-95057bfd06b9\" (UID: \"0bc80e48-26f1-47c6-9fc2-95057bfd06b9\") " Jan 26 17:07:14 crc kubenswrapper[4754]: I0126 17:07:14.538152 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d4102ba6-b5f4-40a7-8551-300924ddba03-operator-scripts\") pod \"d4102ba6-b5f4-40a7-8551-300924ddba03\" (UID: \"d4102ba6-b5f4-40a7-8551-300924ddba03\") " Jan 26 17:07:14 crc kubenswrapper[4754]: I0126 17:07:14.538881 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d4102ba6-b5f4-40a7-8551-300924ddba03-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "d4102ba6-b5f4-40a7-8551-300924ddba03" (UID: "d4102ba6-b5f4-40a7-8551-300924ddba03"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:07:14 crc kubenswrapper[4754]: I0126 17:07:14.539881 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0bc80e48-26f1-47c6-9fc2-95057bfd06b9-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "0bc80e48-26f1-47c6-9fc2-95057bfd06b9" (UID: "0bc80e48-26f1-47c6-9fc2-95057bfd06b9"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:07:14 crc kubenswrapper[4754]: I0126 17:07:14.544833 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0bc80e48-26f1-47c6-9fc2-95057bfd06b9-kube-api-access-qgf5s" (OuterVolumeSpecName: "kube-api-access-qgf5s") pod "0bc80e48-26f1-47c6-9fc2-95057bfd06b9" (UID: "0bc80e48-26f1-47c6-9fc2-95057bfd06b9"). InnerVolumeSpecName "kube-api-access-qgf5s". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:07:14 crc kubenswrapper[4754]: I0126 17:07:14.544912 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d4102ba6-b5f4-40a7-8551-300924ddba03-kube-api-access-mf9jt" (OuterVolumeSpecName: "kube-api-access-mf9jt") pod "d4102ba6-b5f4-40a7-8551-300924ddba03" (UID: "d4102ba6-b5f4-40a7-8551-300924ddba03"). InnerVolumeSpecName "kube-api-access-mf9jt". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:07:14 crc kubenswrapper[4754]: I0126 17:07:14.639742 4754 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d4102ba6-b5f4-40a7-8551-300924ddba03-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 26 17:07:14 crc kubenswrapper[4754]: I0126 17:07:14.639785 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qgf5s\" (UniqueName: \"kubernetes.io/projected/0bc80e48-26f1-47c6-9fc2-95057bfd06b9-kube-api-access-qgf5s\") on node \"crc\" DevicePath \"\"" Jan 26 17:07:14 crc kubenswrapper[4754]: I0126 17:07:14.639799 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mf9jt\" (UniqueName: \"kubernetes.io/projected/d4102ba6-b5f4-40a7-8551-300924ddba03-kube-api-access-mf9jt\") on node \"crc\" DevicePath \"\"" Jan 26 17:07:14 crc kubenswrapper[4754]: I0126 17:07:14.639811 4754 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0bc80e48-26f1-47c6-9fc2-95057bfd06b9-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 26 17:07:14 crc kubenswrapper[4754]: I0126 17:07:14.918873 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-899gb" Jan 26 17:07:14 crc kubenswrapper[4754]: I0126 17:07:14.918905 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-899gb" event={"ID":"d4102ba6-b5f4-40a7-8551-300924ddba03","Type":"ContainerDied","Data":"49bbcfa99aa730783f9874850536db6520ba77495a68541e0b096be05ca686f0"} Jan 26 17:07:14 crc kubenswrapper[4754]: I0126 17:07:14.919830 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="49bbcfa99aa730783f9874850536db6520ba77495a68541e0b096be05ca686f0" Jan 26 17:07:14 crc kubenswrapper[4754]: I0126 17:07:14.923066 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-92jjf" event={"ID":"0bc80e48-26f1-47c6-9fc2-95057bfd06b9","Type":"ContainerDied","Data":"1205f95491a521c08c8765c70e234af7423e52e6d2982e01ab0df54ce13cfc80"} Jan 26 17:07:14 crc kubenswrapper[4754]: I0126 17:07:14.923262 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1205f95491a521c08c8765c70e234af7423e52e6d2982e01ab0df54ce13cfc80" Jan 26 17:07:14 crc kubenswrapper[4754]: I0126 17:07:14.923118 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-92jjf" Jan 26 17:07:14 crc kubenswrapper[4754]: I0126 17:07:14.929245 4754 generic.go:334] "Generic (PLEG): container finished" podID="f97b956b-11ee-484d-917e-75b5aba780f0" containerID="ee7fb2afad886aa0086afd0f8654f9b549226222264cb0f25e231475789ed4b7" exitCode=0 Jan 26 17:07:14 crc kubenswrapper[4754]: I0126 17:07:14.929299 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-bmlgs" event={"ID":"f97b956b-11ee-484d-917e-75b5aba780f0","Type":"ContainerDied","Data":"ee7fb2afad886aa0086afd0f8654f9b549226222264cb0f25e231475789ed4b7"} Jan 26 17:07:14 crc kubenswrapper[4754]: I0126 17:07:14.929357 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-bmlgs" event={"ID":"f97b956b-11ee-484d-917e-75b5aba780f0","Type":"ContainerStarted","Data":"27c43f7bf6ef2d9395f1cf3561f02d2143813a0662e4795cba81c858d3b7023a"} Jan 26 17:07:14 crc kubenswrapper[4754]: I0126 17:07:14.933582 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-dcjt9" event={"ID":"83bdc38e-6bbc-4ea8-b57c-9942707b4071","Type":"ContainerDied","Data":"fba3e58ff4d0feddeb997a2011412fbfbb2a4105f0fe85354244b9fa3f9160a7"} Jan 26 17:07:14 crc kubenswrapper[4754]: I0126 17:07:14.933618 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fba3e58ff4d0feddeb997a2011412fbfbb2a4105f0fe85354244b9fa3f9160a7" Jan 26 17:07:14 crc kubenswrapper[4754]: I0126 17:07:14.933699 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-dcjt9" Jan 26 17:07:14 crc kubenswrapper[4754]: I0126 17:07:14.934221 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-764c5664d7-gmrtk" Jan 26 17:07:14 crc kubenswrapper[4754]: I0126 17:07:14.970992 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-764c5664d7-gmrtk" podStartSLOduration=6.970968901 podStartE2EDuration="6.970968901s" podCreationTimestamp="2026-01-26 17:07:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 17:07:14.967316218 +0000 UTC m=+1201.491496682" watchObservedRunningTime="2026-01-26 17:07:14.970968901 +0000 UTC m=+1201.495149335" Jan 26 17:07:15 crc kubenswrapper[4754]: I0126 17:07:15.286793 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-b037-account-create-update-sb2vx" Jan 26 17:07:15 crc kubenswrapper[4754]: I0126 17:07:15.292334 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-c7d7-account-create-update-qjftw" Jan 26 17:07:15 crc kubenswrapper[4754]: I0126 17:07:15.352505 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-c8dc-account-create-update-w9qwp" Jan 26 17:07:15 crc kubenswrapper[4754]: I0126 17:07:15.461909 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r2wf8\" (UniqueName: \"kubernetes.io/projected/2cb8378d-e281-4399-a82b-4fe351fcac16-kube-api-access-r2wf8\") pod \"2cb8378d-e281-4399-a82b-4fe351fcac16\" (UID: \"2cb8378d-e281-4399-a82b-4fe351fcac16\") " Jan 26 17:07:15 crc kubenswrapper[4754]: I0126 17:07:15.462225 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d5f8d\" (UniqueName: \"kubernetes.io/projected/fc67dbf2-5b6b-44cb-a5d4-8b7e32fd4a87-kube-api-access-d5f8d\") pod \"fc67dbf2-5b6b-44cb-a5d4-8b7e32fd4a87\" (UID: \"fc67dbf2-5b6b-44cb-a5d4-8b7e32fd4a87\") " Jan 26 17:07:15 crc kubenswrapper[4754]: I0126 17:07:15.462259 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x5982\" (UniqueName: \"kubernetes.io/projected/22b9c223-a293-4248-a32b-36498fa1a43a-kube-api-access-x5982\") pod \"22b9c223-a293-4248-a32b-36498fa1a43a\" (UID: \"22b9c223-a293-4248-a32b-36498fa1a43a\") " Jan 26 17:07:15 crc kubenswrapper[4754]: I0126 17:07:15.462290 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/22b9c223-a293-4248-a32b-36498fa1a43a-operator-scripts\") pod \"22b9c223-a293-4248-a32b-36498fa1a43a\" (UID: \"22b9c223-a293-4248-a32b-36498fa1a43a\") " Jan 26 17:07:15 crc kubenswrapper[4754]: I0126 17:07:15.462339 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2cb8378d-e281-4399-a82b-4fe351fcac16-operator-scripts\") pod \"2cb8378d-e281-4399-a82b-4fe351fcac16\" (UID: \"2cb8378d-e281-4399-a82b-4fe351fcac16\") " Jan 26 17:07:15 crc kubenswrapper[4754]: I0126 17:07:15.462401 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fc67dbf2-5b6b-44cb-a5d4-8b7e32fd4a87-operator-scripts\") pod \"fc67dbf2-5b6b-44cb-a5d4-8b7e32fd4a87\" (UID: \"fc67dbf2-5b6b-44cb-a5d4-8b7e32fd4a87\") " Jan 26 17:07:15 crc kubenswrapper[4754]: I0126 17:07:15.463184 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22b9c223-a293-4248-a32b-36498fa1a43a-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "22b9c223-a293-4248-a32b-36498fa1a43a" (UID: "22b9c223-a293-4248-a32b-36498fa1a43a"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:07:15 crc kubenswrapper[4754]: I0126 17:07:15.463184 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2cb8378d-e281-4399-a82b-4fe351fcac16-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "2cb8378d-e281-4399-a82b-4fe351fcac16" (UID: "2cb8378d-e281-4399-a82b-4fe351fcac16"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:07:15 crc kubenswrapper[4754]: I0126 17:07:15.463273 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fc67dbf2-5b6b-44cb-a5d4-8b7e32fd4a87-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "fc67dbf2-5b6b-44cb-a5d4-8b7e32fd4a87" (UID: "fc67dbf2-5b6b-44cb-a5d4-8b7e32fd4a87"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:07:15 crc kubenswrapper[4754]: I0126 17:07:15.468220 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22b9c223-a293-4248-a32b-36498fa1a43a-kube-api-access-x5982" (OuterVolumeSpecName: "kube-api-access-x5982") pod "22b9c223-a293-4248-a32b-36498fa1a43a" (UID: "22b9c223-a293-4248-a32b-36498fa1a43a"). InnerVolumeSpecName "kube-api-access-x5982". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:07:15 crc kubenswrapper[4754]: I0126 17:07:15.468238 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fc67dbf2-5b6b-44cb-a5d4-8b7e32fd4a87-kube-api-access-d5f8d" (OuterVolumeSpecName: "kube-api-access-d5f8d") pod "fc67dbf2-5b6b-44cb-a5d4-8b7e32fd4a87" (UID: "fc67dbf2-5b6b-44cb-a5d4-8b7e32fd4a87"). InnerVolumeSpecName "kube-api-access-d5f8d". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:07:15 crc kubenswrapper[4754]: I0126 17:07:15.474949 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2cb8378d-e281-4399-a82b-4fe351fcac16-kube-api-access-r2wf8" (OuterVolumeSpecName: "kube-api-access-r2wf8") pod "2cb8378d-e281-4399-a82b-4fe351fcac16" (UID: "2cb8378d-e281-4399-a82b-4fe351fcac16"). InnerVolumeSpecName "kube-api-access-r2wf8". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:07:15 crc kubenswrapper[4754]: I0126 17:07:15.564477 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r2wf8\" (UniqueName: \"kubernetes.io/projected/2cb8378d-e281-4399-a82b-4fe351fcac16-kube-api-access-r2wf8\") on node \"crc\" DevicePath \"\"" Jan 26 17:07:15 crc kubenswrapper[4754]: I0126 17:07:15.564513 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d5f8d\" (UniqueName: \"kubernetes.io/projected/fc67dbf2-5b6b-44cb-a5d4-8b7e32fd4a87-kube-api-access-d5f8d\") on node \"crc\" DevicePath \"\"" Jan 26 17:07:15 crc kubenswrapper[4754]: I0126 17:07:15.564523 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x5982\" (UniqueName: \"kubernetes.io/projected/22b9c223-a293-4248-a32b-36498fa1a43a-kube-api-access-x5982\") on node \"crc\" DevicePath \"\"" Jan 26 17:07:15 crc kubenswrapper[4754]: I0126 17:07:15.564532 4754 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/22b9c223-a293-4248-a32b-36498fa1a43a-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 26 17:07:15 crc kubenswrapper[4754]: I0126 17:07:15.564540 4754 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2cb8378d-e281-4399-a82b-4fe351fcac16-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 26 17:07:15 crc kubenswrapper[4754]: I0126 17:07:15.564548 4754 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fc67dbf2-5b6b-44cb-a5d4-8b7e32fd4a87-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 26 17:07:15 crc kubenswrapper[4754]: I0126 17:07:15.945289 4754 generic.go:334] "Generic (PLEG): container finished" podID="bae8ecc7-8fac-4a63-9e99-4e7543eae5af" containerID="e330185b9b6205d6f4d57fde1a2bb14997226cc270954bfc37385ba8ca09676b" exitCode=0 Jan 26 17:07:15 crc kubenswrapper[4754]: I0126 17:07:15.945364 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-qbqp5" event={"ID":"bae8ecc7-8fac-4a63-9e99-4e7543eae5af","Type":"ContainerDied","Data":"e330185b9b6205d6f4d57fde1a2bb14997226cc270954bfc37385ba8ca09676b"} Jan 26 17:07:15 crc kubenswrapper[4754]: I0126 17:07:15.947980 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-b037-account-create-update-sb2vx" event={"ID":"22b9c223-a293-4248-a32b-36498fa1a43a","Type":"ContainerDied","Data":"151cf2767dee7f8025841de1b9e99b834927bed6083f57d05cf63f2462e5148d"} Jan 26 17:07:15 crc kubenswrapper[4754]: I0126 17:07:15.948008 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="151cf2767dee7f8025841de1b9e99b834927bed6083f57d05cf63f2462e5148d" Jan 26 17:07:15 crc kubenswrapper[4754]: I0126 17:07:15.948158 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-b037-account-create-update-sb2vx" Jan 26 17:07:15 crc kubenswrapper[4754]: I0126 17:07:15.950336 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-c8dc-account-create-update-w9qwp" event={"ID":"2cb8378d-e281-4399-a82b-4fe351fcac16","Type":"ContainerDied","Data":"ed0c669695f66a9627fb65fd7b1f83737a368d30fbe5d6436428360dee2d9da4"} Jan 26 17:07:15 crc kubenswrapper[4754]: I0126 17:07:15.950360 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ed0c669695f66a9627fb65fd7b1f83737a368d30fbe5d6436428360dee2d9da4" Jan 26 17:07:15 crc kubenswrapper[4754]: I0126 17:07:15.950410 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-c8dc-account-create-update-w9qwp" Jan 26 17:07:15 crc kubenswrapper[4754]: I0126 17:07:15.955039 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-c7d7-account-create-update-qjftw" Jan 26 17:07:15 crc kubenswrapper[4754]: I0126 17:07:15.956349 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-c7d7-account-create-update-qjftw" event={"ID":"fc67dbf2-5b6b-44cb-a5d4-8b7e32fd4a87","Type":"ContainerDied","Data":"bd4564da838e319c6a58d8959b606791ba5c9791543c55c9f68cbc9e83533b12"} Jan 26 17:07:15 crc kubenswrapper[4754]: I0126 17:07:15.956470 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bd4564da838e319c6a58d8959b606791ba5c9791543c55c9f68cbc9e83533b12" Jan 26 17:07:18 crc kubenswrapper[4754]: I0126 17:07:18.502882 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-764c5664d7-gmrtk" Jan 26 17:07:18 crc kubenswrapper[4754]: I0126 17:07:18.575881 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-698758b865-z5pcq"] Jan 26 17:07:18 crc kubenswrapper[4754]: I0126 17:07:18.576206 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-698758b865-z5pcq" podUID="b5317650-1469-4237-8645-6031475c5b37" containerName="dnsmasq-dns" containerID="cri-o://195a1f054f708b9ea9b0b61078c6a54456031c18ec70286ce347d62b42231752" gracePeriod=10 Jan 26 17:07:18 crc kubenswrapper[4754]: I0126 17:07:18.803845 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-bmlgs" Jan 26 17:07:18 crc kubenswrapper[4754]: I0126 17:07:18.824226 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-qbqp5" Jan 26 17:07:18 crc kubenswrapper[4754]: I0126 17:07:18.922894 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f97b956b-11ee-484d-917e-75b5aba780f0-operator-scripts\") pod \"f97b956b-11ee-484d-917e-75b5aba780f0\" (UID: \"f97b956b-11ee-484d-917e-75b5aba780f0\") " Jan 26 17:07:18 crc kubenswrapper[4754]: I0126 17:07:18.923000 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xbcxr\" (UniqueName: \"kubernetes.io/projected/f97b956b-11ee-484d-917e-75b5aba780f0-kube-api-access-xbcxr\") pod \"f97b956b-11ee-484d-917e-75b5aba780f0\" (UID: \"f97b956b-11ee-484d-917e-75b5aba780f0\") " Jan 26 17:07:18 crc kubenswrapper[4754]: I0126 17:07:18.931794 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f97b956b-11ee-484d-917e-75b5aba780f0-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "f97b956b-11ee-484d-917e-75b5aba780f0" (UID: "f97b956b-11ee-484d-917e-75b5aba780f0"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:07:18 crc kubenswrapper[4754]: I0126 17:07:18.933961 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f97b956b-11ee-484d-917e-75b5aba780f0-kube-api-access-xbcxr" (OuterVolumeSpecName: "kube-api-access-xbcxr") pod "f97b956b-11ee-484d-917e-75b5aba780f0" (UID: "f97b956b-11ee-484d-917e-75b5aba780f0"). InnerVolumeSpecName "kube-api-access-xbcxr". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:07:18 crc kubenswrapper[4754]: I0126 17:07:18.994344 4754 generic.go:334] "Generic (PLEG): container finished" podID="b5317650-1469-4237-8645-6031475c5b37" containerID="195a1f054f708b9ea9b0b61078c6a54456031c18ec70286ce347d62b42231752" exitCode=0 Jan 26 17:07:18 crc kubenswrapper[4754]: I0126 17:07:18.994427 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-z5pcq" event={"ID":"b5317650-1469-4237-8645-6031475c5b37","Type":"ContainerDied","Data":"195a1f054f708b9ea9b0b61078c6a54456031c18ec70286ce347d62b42231752"} Jan 26 17:07:18 crc kubenswrapper[4754]: I0126 17:07:18.995949 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-qbqp5" event={"ID":"bae8ecc7-8fac-4a63-9e99-4e7543eae5af","Type":"ContainerDied","Data":"d1d22ae82d685b4b2ac68b19cf28f22b7537f243a37e0d0cee27ebbc64da0f64"} Jan 26 17:07:18 crc kubenswrapper[4754]: I0126 17:07:18.995976 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d1d22ae82d685b4b2ac68b19cf28f22b7537f243a37e0d0cee27ebbc64da0f64" Jan 26 17:07:18 crc kubenswrapper[4754]: I0126 17:07:18.996033 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-qbqp5" Jan 26 17:07:18 crc kubenswrapper[4754]: I0126 17:07:18.997952 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-bmlgs" event={"ID":"f97b956b-11ee-484d-917e-75b5aba780f0","Type":"ContainerDied","Data":"27c43f7bf6ef2d9395f1cf3561f02d2143813a0662e4795cba81c858d3b7023a"} Jan 26 17:07:18 crc kubenswrapper[4754]: I0126 17:07:18.997980 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="27c43f7bf6ef2d9395f1cf3561f02d2143813a0662e4795cba81c858d3b7023a" Jan 26 17:07:18 crc kubenswrapper[4754]: I0126 17:07:18.998039 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-bmlgs" Jan 26 17:07:19 crc kubenswrapper[4754]: I0126 17:07:19.024365 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/bae8ecc7-8fac-4a63-9e99-4e7543eae5af-db-sync-config-data\") pod \"bae8ecc7-8fac-4a63-9e99-4e7543eae5af\" (UID: \"bae8ecc7-8fac-4a63-9e99-4e7543eae5af\") " Jan 26 17:07:19 crc kubenswrapper[4754]: I0126 17:07:19.024775 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bae8ecc7-8fac-4a63-9e99-4e7543eae5af-config-data\") pod \"bae8ecc7-8fac-4a63-9e99-4e7543eae5af\" (UID: \"bae8ecc7-8fac-4a63-9e99-4e7543eae5af\") " Jan 26 17:07:19 crc kubenswrapper[4754]: I0126 17:07:19.024893 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z4sm2\" (UniqueName: \"kubernetes.io/projected/bae8ecc7-8fac-4a63-9e99-4e7543eae5af-kube-api-access-z4sm2\") pod \"bae8ecc7-8fac-4a63-9e99-4e7543eae5af\" (UID: \"bae8ecc7-8fac-4a63-9e99-4e7543eae5af\") " Jan 26 17:07:19 crc kubenswrapper[4754]: I0126 17:07:19.024951 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bae8ecc7-8fac-4a63-9e99-4e7543eae5af-combined-ca-bundle\") pod \"bae8ecc7-8fac-4a63-9e99-4e7543eae5af\" (UID: \"bae8ecc7-8fac-4a63-9e99-4e7543eae5af\") " Jan 26 17:07:19 crc kubenswrapper[4754]: I0126 17:07:19.025317 4754 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f97b956b-11ee-484d-917e-75b5aba780f0-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 26 17:07:19 crc kubenswrapper[4754]: I0126 17:07:19.025332 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xbcxr\" (UniqueName: \"kubernetes.io/projected/f97b956b-11ee-484d-917e-75b5aba780f0-kube-api-access-xbcxr\") on node \"crc\" DevicePath \"\"" Jan 26 17:07:19 crc kubenswrapper[4754]: I0126 17:07:19.030393 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bae8ecc7-8fac-4a63-9e99-4e7543eae5af-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "bae8ecc7-8fac-4a63-9e99-4e7543eae5af" (UID: "bae8ecc7-8fac-4a63-9e99-4e7543eae5af"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:07:19 crc kubenswrapper[4754]: I0126 17:07:19.032518 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bae8ecc7-8fac-4a63-9e99-4e7543eae5af-kube-api-access-z4sm2" (OuterVolumeSpecName: "kube-api-access-z4sm2") pod "bae8ecc7-8fac-4a63-9e99-4e7543eae5af" (UID: "bae8ecc7-8fac-4a63-9e99-4e7543eae5af"). InnerVolumeSpecName "kube-api-access-z4sm2". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:07:19 crc kubenswrapper[4754]: I0126 17:07:19.047982 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-z5pcq" Jan 26 17:07:19 crc kubenswrapper[4754]: I0126 17:07:19.062029 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bae8ecc7-8fac-4a63-9e99-4e7543eae5af-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bae8ecc7-8fac-4a63-9e99-4e7543eae5af" (UID: "bae8ecc7-8fac-4a63-9e99-4e7543eae5af"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:07:19 crc kubenswrapper[4754]: I0126 17:07:19.089893 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bae8ecc7-8fac-4a63-9e99-4e7543eae5af-config-data" (OuterVolumeSpecName: "config-data") pod "bae8ecc7-8fac-4a63-9e99-4e7543eae5af" (UID: "bae8ecc7-8fac-4a63-9e99-4e7543eae5af"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:07:19 crc kubenswrapper[4754]: I0126 17:07:19.127783 4754 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bae8ecc7-8fac-4a63-9e99-4e7543eae5af-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 26 17:07:19 crc kubenswrapper[4754]: I0126 17:07:19.127813 4754 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/bae8ecc7-8fac-4a63-9e99-4e7543eae5af-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Jan 26 17:07:19 crc kubenswrapper[4754]: I0126 17:07:19.127823 4754 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bae8ecc7-8fac-4a63-9e99-4e7543eae5af-config-data\") on node \"crc\" DevicePath \"\"" Jan 26 17:07:19 crc kubenswrapper[4754]: I0126 17:07:19.127833 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z4sm2\" (UniqueName: \"kubernetes.io/projected/bae8ecc7-8fac-4a63-9e99-4e7543eae5af-kube-api-access-z4sm2\") on node \"crc\" DevicePath \"\"" Jan 26 17:07:19 crc kubenswrapper[4754]: I0126 17:07:19.229276 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b5317650-1469-4237-8645-6031475c5b37-dns-svc\") pod \"b5317650-1469-4237-8645-6031475c5b37\" (UID: \"b5317650-1469-4237-8645-6031475c5b37\") " Jan 26 17:07:19 crc kubenswrapper[4754]: I0126 17:07:19.229452 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b5317650-1469-4237-8645-6031475c5b37-ovsdbserver-nb\") pod \"b5317650-1469-4237-8645-6031475c5b37\" (UID: \"b5317650-1469-4237-8645-6031475c5b37\") " Jan 26 17:07:19 crc kubenswrapper[4754]: I0126 17:07:19.229508 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hxqbk\" (UniqueName: \"kubernetes.io/projected/b5317650-1469-4237-8645-6031475c5b37-kube-api-access-hxqbk\") pod \"b5317650-1469-4237-8645-6031475c5b37\" (UID: \"b5317650-1469-4237-8645-6031475c5b37\") " Jan 26 17:07:19 crc kubenswrapper[4754]: I0126 17:07:19.229539 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b5317650-1469-4237-8645-6031475c5b37-ovsdbserver-sb\") pod \"b5317650-1469-4237-8645-6031475c5b37\" (UID: \"b5317650-1469-4237-8645-6031475c5b37\") " Jan 26 17:07:19 crc kubenswrapper[4754]: I0126 17:07:19.229565 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b5317650-1469-4237-8645-6031475c5b37-config\") pod \"b5317650-1469-4237-8645-6031475c5b37\" (UID: \"b5317650-1469-4237-8645-6031475c5b37\") " Jan 26 17:07:19 crc kubenswrapper[4754]: I0126 17:07:19.240650 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b5317650-1469-4237-8645-6031475c5b37-kube-api-access-hxqbk" (OuterVolumeSpecName: "kube-api-access-hxqbk") pod "b5317650-1469-4237-8645-6031475c5b37" (UID: "b5317650-1469-4237-8645-6031475c5b37"). InnerVolumeSpecName "kube-api-access-hxqbk". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:07:19 crc kubenswrapper[4754]: I0126 17:07:19.281891 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b5317650-1469-4237-8645-6031475c5b37-config" (OuterVolumeSpecName: "config") pod "b5317650-1469-4237-8645-6031475c5b37" (UID: "b5317650-1469-4237-8645-6031475c5b37"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:07:19 crc kubenswrapper[4754]: I0126 17:07:19.290923 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b5317650-1469-4237-8645-6031475c5b37-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "b5317650-1469-4237-8645-6031475c5b37" (UID: "b5317650-1469-4237-8645-6031475c5b37"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:07:19 crc kubenswrapper[4754]: I0126 17:07:19.294232 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b5317650-1469-4237-8645-6031475c5b37-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "b5317650-1469-4237-8645-6031475c5b37" (UID: "b5317650-1469-4237-8645-6031475c5b37"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:07:19 crc kubenswrapper[4754]: I0126 17:07:19.298874 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b5317650-1469-4237-8645-6031475c5b37-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "b5317650-1469-4237-8645-6031475c5b37" (UID: "b5317650-1469-4237-8645-6031475c5b37"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:07:19 crc kubenswrapper[4754]: I0126 17:07:19.334936 4754 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b5317650-1469-4237-8645-6031475c5b37-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Jan 26 17:07:19 crc kubenswrapper[4754]: I0126 17:07:19.334969 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hxqbk\" (UniqueName: \"kubernetes.io/projected/b5317650-1469-4237-8645-6031475c5b37-kube-api-access-hxqbk\") on node \"crc\" DevicePath \"\"" Jan 26 17:07:19 crc kubenswrapper[4754]: I0126 17:07:19.334979 4754 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b5317650-1469-4237-8645-6031475c5b37-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Jan 26 17:07:19 crc kubenswrapper[4754]: I0126 17:07:19.334988 4754 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b5317650-1469-4237-8645-6031475c5b37-config\") on node \"crc\" DevicePath \"\"" Jan 26 17:07:19 crc kubenswrapper[4754]: I0126 17:07:19.334996 4754 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b5317650-1469-4237-8645-6031475c5b37-dns-svc\") on node \"crc\" DevicePath \"\"" Jan 26 17:07:20 crc kubenswrapper[4754]: I0126 17:07:20.010775 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-68nzr" event={"ID":"35a2e341-10fc-48ca-8c99-65f1664cbba5","Type":"ContainerStarted","Data":"c553cf4f5c0479e11cee44a2fd1c1c574a2cb4a6536dec6cc0242f7317cbc1fb"} Jan 26 17:07:20 crc kubenswrapper[4754]: I0126 17:07:20.018516 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-z5pcq" event={"ID":"b5317650-1469-4237-8645-6031475c5b37","Type":"ContainerDied","Data":"55eb578a90c8ba0834c5d56e7fdc32dd86c9ce14da617f0b35505ec34e6168cb"} Jan 26 17:07:20 crc kubenswrapper[4754]: I0126 17:07:20.018590 4754 scope.go:117] "RemoveContainer" containerID="195a1f054f708b9ea9b0b61078c6a54456031c18ec70286ce347d62b42231752" Jan 26 17:07:20 crc kubenswrapper[4754]: I0126 17:07:20.018859 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-z5pcq" Jan 26 17:07:20 crc kubenswrapper[4754]: I0126 17:07:20.069201 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-68nzr" podStartSLOduration=3.116294505 podStartE2EDuration="12.06917339s" podCreationTimestamp="2026-01-26 17:07:08 +0000 UTC" firstStartedPulling="2026-01-26 17:07:09.85123432 +0000 UTC m=+1196.375414754" lastFinishedPulling="2026-01-26 17:07:18.804113205 +0000 UTC m=+1205.328293639" observedRunningTime="2026-01-26 17:07:20.036503186 +0000 UTC m=+1206.560683630" watchObservedRunningTime="2026-01-26 17:07:20.06917339 +0000 UTC m=+1206.593353824" Jan 26 17:07:20 crc kubenswrapper[4754]: I0126 17:07:20.123336 4754 scope.go:117] "RemoveContainer" containerID="eb843a44f878f648e2b6b755b6528ad1dcc1f5f1fd6ead5670f2b8320ba9d8e4" Jan 26 17:07:20 crc kubenswrapper[4754]: I0126 17:07:20.131483 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-698758b865-z5pcq"] Jan 26 17:07:20 crc kubenswrapper[4754]: I0126 17:07:20.148169 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-698758b865-z5pcq"] Jan 26 17:07:20 crc kubenswrapper[4754]: I0126 17:07:20.189281 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-q8pwj"] Jan 26 17:07:20 crc kubenswrapper[4754]: E0126 17:07:20.189705 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5317650-1469-4237-8645-6031475c5b37" containerName="init" Jan 26 17:07:20 crc kubenswrapper[4754]: I0126 17:07:20.189723 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5317650-1469-4237-8645-6031475c5b37" containerName="init" Jan 26 17:07:20 crc kubenswrapper[4754]: E0126 17:07:20.189738 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0bc80e48-26f1-47c6-9fc2-95057bfd06b9" containerName="mariadb-database-create" Jan 26 17:07:20 crc kubenswrapper[4754]: I0126 17:07:20.189745 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="0bc80e48-26f1-47c6-9fc2-95057bfd06b9" containerName="mariadb-database-create" Jan 26 17:07:20 crc kubenswrapper[4754]: E0126 17:07:20.189769 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f97b956b-11ee-484d-917e-75b5aba780f0" containerName="mariadb-account-create-update" Jan 26 17:07:20 crc kubenswrapper[4754]: I0126 17:07:20.189780 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="f97b956b-11ee-484d-917e-75b5aba780f0" containerName="mariadb-account-create-update" Jan 26 17:07:20 crc kubenswrapper[4754]: E0126 17:07:20.189796 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc67dbf2-5b6b-44cb-a5d4-8b7e32fd4a87" containerName="mariadb-account-create-update" Jan 26 17:07:20 crc kubenswrapper[4754]: I0126 17:07:20.189803 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc67dbf2-5b6b-44cb-a5d4-8b7e32fd4a87" containerName="mariadb-account-create-update" Jan 26 17:07:20 crc kubenswrapper[4754]: E0126 17:07:20.189819 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bae8ecc7-8fac-4a63-9e99-4e7543eae5af" containerName="glance-db-sync" Jan 26 17:07:20 crc kubenswrapper[4754]: I0126 17:07:20.189826 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="bae8ecc7-8fac-4a63-9e99-4e7543eae5af" containerName="glance-db-sync" Jan 26 17:07:20 crc kubenswrapper[4754]: E0126 17:07:20.189836 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2cb8378d-e281-4399-a82b-4fe351fcac16" containerName="mariadb-account-create-update" Jan 26 17:07:20 crc kubenswrapper[4754]: I0126 17:07:20.189843 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="2cb8378d-e281-4399-a82b-4fe351fcac16" containerName="mariadb-account-create-update" Jan 26 17:07:20 crc kubenswrapper[4754]: E0126 17:07:20.189853 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83bdc38e-6bbc-4ea8-b57c-9942707b4071" containerName="mariadb-database-create" Jan 26 17:07:20 crc kubenswrapper[4754]: I0126 17:07:20.189860 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="83bdc38e-6bbc-4ea8-b57c-9942707b4071" containerName="mariadb-database-create" Jan 26 17:07:20 crc kubenswrapper[4754]: E0126 17:07:20.189869 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5317650-1469-4237-8645-6031475c5b37" containerName="dnsmasq-dns" Jan 26 17:07:20 crc kubenswrapper[4754]: I0126 17:07:20.189876 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5317650-1469-4237-8645-6031475c5b37" containerName="dnsmasq-dns" Jan 26 17:07:20 crc kubenswrapper[4754]: E0126 17:07:20.189889 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4102ba6-b5f4-40a7-8551-300924ddba03" containerName="mariadb-database-create" Jan 26 17:07:20 crc kubenswrapper[4754]: I0126 17:07:20.189897 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4102ba6-b5f4-40a7-8551-300924ddba03" containerName="mariadb-database-create" Jan 26 17:07:20 crc kubenswrapper[4754]: E0126 17:07:20.189909 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="22b9c223-a293-4248-a32b-36498fa1a43a" containerName="mariadb-account-create-update" Jan 26 17:07:20 crc kubenswrapper[4754]: I0126 17:07:20.189916 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="22b9c223-a293-4248-a32b-36498fa1a43a" containerName="mariadb-account-create-update" Jan 26 17:07:20 crc kubenswrapper[4754]: I0126 17:07:20.190115 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="0bc80e48-26f1-47c6-9fc2-95057bfd06b9" containerName="mariadb-database-create" Jan 26 17:07:20 crc kubenswrapper[4754]: I0126 17:07:20.190131 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="d4102ba6-b5f4-40a7-8551-300924ddba03" containerName="mariadb-database-create" Jan 26 17:07:20 crc kubenswrapper[4754]: I0126 17:07:20.190145 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="22b9c223-a293-4248-a32b-36498fa1a43a" containerName="mariadb-account-create-update" Jan 26 17:07:20 crc kubenswrapper[4754]: I0126 17:07:20.190156 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="f97b956b-11ee-484d-917e-75b5aba780f0" containerName="mariadb-account-create-update" Jan 26 17:07:20 crc kubenswrapper[4754]: I0126 17:07:20.190168 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="2cb8378d-e281-4399-a82b-4fe351fcac16" containerName="mariadb-account-create-update" Jan 26 17:07:20 crc kubenswrapper[4754]: I0126 17:07:20.190180 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="b5317650-1469-4237-8645-6031475c5b37" containerName="dnsmasq-dns" Jan 26 17:07:20 crc kubenswrapper[4754]: I0126 17:07:20.190199 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="83bdc38e-6bbc-4ea8-b57c-9942707b4071" containerName="mariadb-database-create" Jan 26 17:07:20 crc kubenswrapper[4754]: I0126 17:07:20.190212 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="bae8ecc7-8fac-4a63-9e99-4e7543eae5af" containerName="glance-db-sync" Jan 26 17:07:20 crc kubenswrapper[4754]: I0126 17:07:20.190226 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="fc67dbf2-5b6b-44cb-a5d4-8b7e32fd4a87" containerName="mariadb-account-create-update" Jan 26 17:07:22 crc kubenswrapper[4754]: I0126 17:07:20.194939 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74f6bcbc87-q8pwj" Jan 26 17:07:22 crc kubenswrapper[4754]: I0126 17:07:20.252740 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-q8pwj"] Jan 26 17:07:22 crc kubenswrapper[4754]: I0126 17:07:20.280785 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a6873959-d867-4a38-a10f-82ff80d0800b-dns-swift-storage-0\") pod \"dnsmasq-dns-74f6bcbc87-q8pwj\" (UID: \"a6873959-d867-4a38-a10f-82ff80d0800b\") " pod="openstack/dnsmasq-dns-74f6bcbc87-q8pwj" Jan 26 17:07:22 crc kubenswrapper[4754]: I0126 17:07:20.280834 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a6873959-d867-4a38-a10f-82ff80d0800b-config\") pod \"dnsmasq-dns-74f6bcbc87-q8pwj\" (UID: \"a6873959-d867-4a38-a10f-82ff80d0800b\") " pod="openstack/dnsmasq-dns-74f6bcbc87-q8pwj" Jan 26 17:07:22 crc kubenswrapper[4754]: I0126 17:07:20.280861 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a6873959-d867-4a38-a10f-82ff80d0800b-dns-svc\") pod \"dnsmasq-dns-74f6bcbc87-q8pwj\" (UID: \"a6873959-d867-4a38-a10f-82ff80d0800b\") " pod="openstack/dnsmasq-dns-74f6bcbc87-q8pwj" Jan 26 17:07:22 crc kubenswrapper[4754]: I0126 17:07:20.280907 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a6873959-d867-4a38-a10f-82ff80d0800b-ovsdbserver-sb\") pod \"dnsmasq-dns-74f6bcbc87-q8pwj\" (UID: \"a6873959-d867-4a38-a10f-82ff80d0800b\") " pod="openstack/dnsmasq-dns-74f6bcbc87-q8pwj" Jan 26 17:07:22 crc kubenswrapper[4754]: I0126 17:07:20.280926 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a6873959-d867-4a38-a10f-82ff80d0800b-ovsdbserver-nb\") pod \"dnsmasq-dns-74f6bcbc87-q8pwj\" (UID: \"a6873959-d867-4a38-a10f-82ff80d0800b\") " pod="openstack/dnsmasq-dns-74f6bcbc87-q8pwj" Jan 26 17:07:22 crc kubenswrapper[4754]: I0126 17:07:20.281279 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-thfkw\" (UniqueName: \"kubernetes.io/projected/a6873959-d867-4a38-a10f-82ff80d0800b-kube-api-access-thfkw\") pod \"dnsmasq-dns-74f6bcbc87-q8pwj\" (UID: \"a6873959-d867-4a38-a10f-82ff80d0800b\") " pod="openstack/dnsmasq-dns-74f6bcbc87-q8pwj" Jan 26 17:07:22 crc kubenswrapper[4754]: I0126 17:07:20.382435 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a6873959-d867-4a38-a10f-82ff80d0800b-dns-svc\") pod \"dnsmasq-dns-74f6bcbc87-q8pwj\" (UID: \"a6873959-d867-4a38-a10f-82ff80d0800b\") " pod="openstack/dnsmasq-dns-74f6bcbc87-q8pwj" Jan 26 17:07:22 crc kubenswrapper[4754]: I0126 17:07:20.382519 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a6873959-d867-4a38-a10f-82ff80d0800b-ovsdbserver-sb\") pod \"dnsmasq-dns-74f6bcbc87-q8pwj\" (UID: \"a6873959-d867-4a38-a10f-82ff80d0800b\") " pod="openstack/dnsmasq-dns-74f6bcbc87-q8pwj" Jan 26 17:07:22 crc kubenswrapper[4754]: I0126 17:07:20.382555 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a6873959-d867-4a38-a10f-82ff80d0800b-ovsdbserver-nb\") pod \"dnsmasq-dns-74f6bcbc87-q8pwj\" (UID: \"a6873959-d867-4a38-a10f-82ff80d0800b\") " pod="openstack/dnsmasq-dns-74f6bcbc87-q8pwj" Jan 26 17:07:22 crc kubenswrapper[4754]: I0126 17:07:20.382595 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-thfkw\" (UniqueName: \"kubernetes.io/projected/a6873959-d867-4a38-a10f-82ff80d0800b-kube-api-access-thfkw\") pod \"dnsmasq-dns-74f6bcbc87-q8pwj\" (UID: \"a6873959-d867-4a38-a10f-82ff80d0800b\") " pod="openstack/dnsmasq-dns-74f6bcbc87-q8pwj" Jan 26 17:07:22 crc kubenswrapper[4754]: I0126 17:07:20.382687 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a6873959-d867-4a38-a10f-82ff80d0800b-dns-swift-storage-0\") pod \"dnsmasq-dns-74f6bcbc87-q8pwj\" (UID: \"a6873959-d867-4a38-a10f-82ff80d0800b\") " pod="openstack/dnsmasq-dns-74f6bcbc87-q8pwj" Jan 26 17:07:22 crc kubenswrapper[4754]: I0126 17:07:20.382709 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a6873959-d867-4a38-a10f-82ff80d0800b-config\") pod \"dnsmasq-dns-74f6bcbc87-q8pwj\" (UID: \"a6873959-d867-4a38-a10f-82ff80d0800b\") " pod="openstack/dnsmasq-dns-74f6bcbc87-q8pwj" Jan 26 17:07:22 crc kubenswrapper[4754]: I0126 17:07:20.383588 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a6873959-d867-4a38-a10f-82ff80d0800b-dns-svc\") pod \"dnsmasq-dns-74f6bcbc87-q8pwj\" (UID: \"a6873959-d867-4a38-a10f-82ff80d0800b\") " pod="openstack/dnsmasq-dns-74f6bcbc87-q8pwj" Jan 26 17:07:22 crc kubenswrapper[4754]: I0126 17:07:20.383616 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a6873959-d867-4a38-a10f-82ff80d0800b-config\") pod \"dnsmasq-dns-74f6bcbc87-q8pwj\" (UID: \"a6873959-d867-4a38-a10f-82ff80d0800b\") " pod="openstack/dnsmasq-dns-74f6bcbc87-q8pwj" Jan 26 17:07:22 crc kubenswrapper[4754]: I0126 17:07:20.384164 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a6873959-d867-4a38-a10f-82ff80d0800b-ovsdbserver-nb\") pod \"dnsmasq-dns-74f6bcbc87-q8pwj\" (UID: \"a6873959-d867-4a38-a10f-82ff80d0800b\") " pod="openstack/dnsmasq-dns-74f6bcbc87-q8pwj" Jan 26 17:07:22 crc kubenswrapper[4754]: I0126 17:07:20.384381 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a6873959-d867-4a38-a10f-82ff80d0800b-dns-swift-storage-0\") pod \"dnsmasq-dns-74f6bcbc87-q8pwj\" (UID: \"a6873959-d867-4a38-a10f-82ff80d0800b\") " pod="openstack/dnsmasq-dns-74f6bcbc87-q8pwj" Jan 26 17:07:22 crc kubenswrapper[4754]: I0126 17:07:20.384415 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a6873959-d867-4a38-a10f-82ff80d0800b-ovsdbserver-sb\") pod \"dnsmasq-dns-74f6bcbc87-q8pwj\" (UID: \"a6873959-d867-4a38-a10f-82ff80d0800b\") " pod="openstack/dnsmasq-dns-74f6bcbc87-q8pwj" Jan 26 17:07:22 crc kubenswrapper[4754]: I0126 17:07:20.407226 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-thfkw\" (UniqueName: \"kubernetes.io/projected/a6873959-d867-4a38-a10f-82ff80d0800b-kube-api-access-thfkw\") pod \"dnsmasq-dns-74f6bcbc87-q8pwj\" (UID: \"a6873959-d867-4a38-a10f-82ff80d0800b\") " pod="openstack/dnsmasq-dns-74f6bcbc87-q8pwj" Jan 26 17:07:22 crc kubenswrapper[4754]: I0126 17:07:20.527083 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74f6bcbc87-q8pwj" Jan 26 17:07:22 crc kubenswrapper[4754]: I0126 17:07:21.777449 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b5317650-1469-4237-8645-6031475c5b37" path="/var/lib/kubelet/pods/b5317650-1469-4237-8645-6031475c5b37/volumes" Jan 26 17:07:22 crc kubenswrapper[4754]: I0126 17:07:22.897747 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-q8pwj"] Jan 26 17:07:22 crc kubenswrapper[4754]: W0126 17:07:22.906232 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda6873959_d867_4a38_a10f_82ff80d0800b.slice/crio-4450de4c2f1d0f92b77707453ea2d45f78ade41dff7f88ba3ef10ce28e271442 WatchSource:0}: Error finding container 4450de4c2f1d0f92b77707453ea2d45f78ade41dff7f88ba3ef10ce28e271442: Status 404 returned error can't find the container with id 4450de4c2f1d0f92b77707453ea2d45f78ade41dff7f88ba3ef10ce28e271442 Jan 26 17:07:23 crc kubenswrapper[4754]: I0126 17:07:23.045382 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74f6bcbc87-q8pwj" event={"ID":"a6873959-d867-4a38-a10f-82ff80d0800b","Type":"ContainerStarted","Data":"4450de4c2f1d0f92b77707453ea2d45f78ade41dff7f88ba3ef10ce28e271442"} Jan 26 17:07:24 crc kubenswrapper[4754]: I0126 17:07:24.054174 4754 generic.go:334] "Generic (PLEG): container finished" podID="a6873959-d867-4a38-a10f-82ff80d0800b" containerID="1375bbc61e6703dac3f4c1f6ad9ded2c448a889672df69d8943c28ce8f1607e4" exitCode=0 Jan 26 17:07:24 crc kubenswrapper[4754]: I0126 17:07:24.054278 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74f6bcbc87-q8pwj" event={"ID":"a6873959-d867-4a38-a10f-82ff80d0800b","Type":"ContainerDied","Data":"1375bbc61e6703dac3f4c1f6ad9ded2c448a889672df69d8943c28ce8f1607e4"} Jan 26 17:07:24 crc kubenswrapper[4754]: I0126 17:07:24.443823 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/root-account-create-update-bmlgs"] Jan 26 17:07:24 crc kubenswrapper[4754]: I0126 17:07:24.449219 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/root-account-create-update-bmlgs"] Jan 26 17:07:25 crc kubenswrapper[4754]: I0126 17:07:25.064263 4754 generic.go:334] "Generic (PLEG): container finished" podID="35a2e341-10fc-48ca-8c99-65f1664cbba5" containerID="c553cf4f5c0479e11cee44a2fd1c1c574a2cb4a6536dec6cc0242f7317cbc1fb" exitCode=0 Jan 26 17:07:25 crc kubenswrapper[4754]: I0126 17:07:25.064339 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-68nzr" event={"ID":"35a2e341-10fc-48ca-8c99-65f1664cbba5","Type":"ContainerDied","Data":"c553cf4f5c0479e11cee44a2fd1c1c574a2cb4a6536dec6cc0242f7317cbc1fb"} Jan 26 17:07:25 crc kubenswrapper[4754]: I0126 17:07:25.066351 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74f6bcbc87-q8pwj" event={"ID":"a6873959-d867-4a38-a10f-82ff80d0800b","Type":"ContainerStarted","Data":"7c2a92f25263e2691e40742a8be4dc87697b019ee912e3a15e8919c7138e41da"} Jan 26 17:07:25 crc kubenswrapper[4754]: I0126 17:07:25.066565 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-74f6bcbc87-q8pwj" Jan 26 17:07:25 crc kubenswrapper[4754]: I0126 17:07:25.104288 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-74f6bcbc87-q8pwj" podStartSLOduration=5.104268975 podStartE2EDuration="5.104268975s" podCreationTimestamp="2026-01-26 17:07:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 17:07:25.101132638 +0000 UTC m=+1211.625313072" watchObservedRunningTime="2026-01-26 17:07:25.104268975 +0000 UTC m=+1211.628449399" Jan 26 17:07:25 crc kubenswrapper[4754]: I0126 17:07:25.783167 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f97b956b-11ee-484d-917e-75b5aba780f0" path="/var/lib/kubelet/pods/f97b956b-11ee-484d-917e-75b5aba780f0/volumes" Jan 26 17:07:26 crc kubenswrapper[4754]: I0126 17:07:26.395961 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-68nzr" Jan 26 17:07:26 crc kubenswrapper[4754]: I0126 17:07:26.478849 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-95d28\" (UniqueName: \"kubernetes.io/projected/35a2e341-10fc-48ca-8c99-65f1664cbba5-kube-api-access-95d28\") pod \"35a2e341-10fc-48ca-8c99-65f1664cbba5\" (UID: \"35a2e341-10fc-48ca-8c99-65f1664cbba5\") " Jan 26 17:07:26 crc kubenswrapper[4754]: I0126 17:07:26.478983 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35a2e341-10fc-48ca-8c99-65f1664cbba5-combined-ca-bundle\") pod \"35a2e341-10fc-48ca-8c99-65f1664cbba5\" (UID: \"35a2e341-10fc-48ca-8c99-65f1664cbba5\") " Jan 26 17:07:26 crc kubenswrapper[4754]: I0126 17:07:26.479016 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/35a2e341-10fc-48ca-8c99-65f1664cbba5-config-data\") pod \"35a2e341-10fc-48ca-8c99-65f1664cbba5\" (UID: \"35a2e341-10fc-48ca-8c99-65f1664cbba5\") " Jan 26 17:07:26 crc kubenswrapper[4754]: I0126 17:07:26.484013 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/35a2e341-10fc-48ca-8c99-65f1664cbba5-kube-api-access-95d28" (OuterVolumeSpecName: "kube-api-access-95d28") pod "35a2e341-10fc-48ca-8c99-65f1664cbba5" (UID: "35a2e341-10fc-48ca-8c99-65f1664cbba5"). InnerVolumeSpecName "kube-api-access-95d28". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:07:26 crc kubenswrapper[4754]: I0126 17:07:26.507781 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/35a2e341-10fc-48ca-8c99-65f1664cbba5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "35a2e341-10fc-48ca-8c99-65f1664cbba5" (UID: "35a2e341-10fc-48ca-8c99-65f1664cbba5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:07:26 crc kubenswrapper[4754]: I0126 17:07:26.521812 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/35a2e341-10fc-48ca-8c99-65f1664cbba5-config-data" (OuterVolumeSpecName: "config-data") pod "35a2e341-10fc-48ca-8c99-65f1664cbba5" (UID: "35a2e341-10fc-48ca-8c99-65f1664cbba5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:07:26 crc kubenswrapper[4754]: I0126 17:07:26.580837 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-95d28\" (UniqueName: \"kubernetes.io/projected/35a2e341-10fc-48ca-8c99-65f1664cbba5-kube-api-access-95d28\") on node \"crc\" DevicePath \"\"" Jan 26 17:07:26 crc kubenswrapper[4754]: I0126 17:07:26.580889 4754 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35a2e341-10fc-48ca-8c99-65f1664cbba5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 26 17:07:26 crc kubenswrapper[4754]: I0126 17:07:26.580903 4754 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/35a2e341-10fc-48ca-8c99-65f1664cbba5-config-data\") on node \"crc\" DevicePath \"\"" Jan 26 17:07:27 crc kubenswrapper[4754]: I0126 17:07:27.089075 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-68nzr" event={"ID":"35a2e341-10fc-48ca-8c99-65f1664cbba5","Type":"ContainerDied","Data":"4044f7375a0edac7261ee9a95007ad80c4c120ca8795b6dbc60d27f19b02a1c9"} Jan 26 17:07:27 crc kubenswrapper[4754]: I0126 17:07:27.089368 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4044f7375a0edac7261ee9a95007ad80c4c120ca8795b6dbc60d27f19b02a1c9" Jan 26 17:07:27 crc kubenswrapper[4754]: I0126 17:07:27.089149 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-68nzr" Jan 26 17:07:27 crc kubenswrapper[4754]: I0126 17:07:27.351625 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-q8pwj"] Jan 26 17:07:27 crc kubenswrapper[4754]: I0126 17:07:27.351894 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-74f6bcbc87-q8pwj" podUID="a6873959-d867-4a38-a10f-82ff80d0800b" containerName="dnsmasq-dns" containerID="cri-o://7c2a92f25263e2691e40742a8be4dc87697b019ee912e3a15e8919c7138e41da" gracePeriod=10 Jan 26 17:07:27 crc kubenswrapper[4754]: I0126 17:07:27.383911 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-6qghh"] Jan 26 17:07:27 crc kubenswrapper[4754]: E0126 17:07:27.385611 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35a2e341-10fc-48ca-8c99-65f1664cbba5" containerName="keystone-db-sync" Jan 26 17:07:27 crc kubenswrapper[4754]: I0126 17:07:27.385647 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="35a2e341-10fc-48ca-8c99-65f1664cbba5" containerName="keystone-db-sync" Jan 26 17:07:27 crc kubenswrapper[4754]: I0126 17:07:27.385945 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="35a2e341-10fc-48ca-8c99-65f1664cbba5" containerName="keystone-db-sync" Jan 26 17:07:27 crc kubenswrapper[4754]: I0126 17:07:27.389918 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-6qghh" Jan 26 17:07:27 crc kubenswrapper[4754]: I0126 17:07:27.400107 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-r42mh" Jan 26 17:07:27 crc kubenswrapper[4754]: I0126 17:07:27.401287 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Jan 26 17:07:27 crc kubenswrapper[4754]: I0126 17:07:27.401630 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Jan 26 17:07:27 crc kubenswrapper[4754]: I0126 17:07:27.401771 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Jan 26 17:07:27 crc kubenswrapper[4754]: I0126 17:07:27.403993 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Jan 26 17:07:27 crc kubenswrapper[4754]: I0126 17:07:27.407508 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-hfr7t"] Jan 26 17:07:27 crc kubenswrapper[4754]: I0126 17:07:27.421293 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-847c4cc679-hfr7t" Jan 26 17:07:27 crc kubenswrapper[4754]: I0126 17:07:27.428461 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-6qghh"] Jan 26 17:07:27 crc kubenswrapper[4754]: I0126 17:07:27.447050 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-hfr7t"] Jan 26 17:07:27 crc kubenswrapper[4754]: I0126 17:07:27.494296 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee429872-adbd-47bb-9ff2-d5d52dba4d12-combined-ca-bundle\") pod \"keystone-bootstrap-6qghh\" (UID: \"ee429872-adbd-47bb-9ff2-d5d52dba4d12\") " pod="openstack/keystone-bootstrap-6qghh" Jan 26 17:07:27 crc kubenswrapper[4754]: I0126 17:07:27.494350 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/ee429872-adbd-47bb-9ff2-d5d52dba4d12-credential-keys\") pod \"keystone-bootstrap-6qghh\" (UID: \"ee429872-adbd-47bb-9ff2-d5d52dba4d12\") " pod="openstack/keystone-bootstrap-6qghh" Jan 26 17:07:27 crc kubenswrapper[4754]: I0126 17:07:27.494377 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/ee429872-adbd-47bb-9ff2-d5d52dba4d12-fernet-keys\") pod \"keystone-bootstrap-6qghh\" (UID: \"ee429872-adbd-47bb-9ff2-d5d52dba4d12\") " pod="openstack/keystone-bootstrap-6qghh" Jan 26 17:07:27 crc kubenswrapper[4754]: I0126 17:07:27.494393 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ee429872-adbd-47bb-9ff2-d5d52dba4d12-scripts\") pod \"keystone-bootstrap-6qghh\" (UID: \"ee429872-adbd-47bb-9ff2-d5d52dba4d12\") " pod="openstack/keystone-bootstrap-6qghh" Jan 26 17:07:27 crc kubenswrapper[4754]: I0126 17:07:27.494448 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ee429872-adbd-47bb-9ff2-d5d52dba4d12-config-data\") pod \"keystone-bootstrap-6qghh\" (UID: \"ee429872-adbd-47bb-9ff2-d5d52dba4d12\") " pod="openstack/keystone-bootstrap-6qghh" Jan 26 17:07:27 crc kubenswrapper[4754]: I0126 17:07:27.494480 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7lrsm\" (UniqueName: \"kubernetes.io/projected/ee429872-adbd-47bb-9ff2-d5d52dba4d12-kube-api-access-7lrsm\") pod \"keystone-bootstrap-6qghh\" (UID: \"ee429872-adbd-47bb-9ff2-d5d52dba4d12\") " pod="openstack/keystone-bootstrap-6qghh" Jan 26 17:07:27 crc kubenswrapper[4754]: I0126 17:07:27.596411 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ee429872-adbd-47bb-9ff2-d5d52dba4d12-scripts\") pod \"keystone-bootstrap-6qghh\" (UID: \"ee429872-adbd-47bb-9ff2-d5d52dba4d12\") " pod="openstack/keystone-bootstrap-6qghh" Jan 26 17:07:27 crc kubenswrapper[4754]: I0126 17:07:27.596472 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xfkfj\" (UniqueName: \"kubernetes.io/projected/83360adb-6875-47a7-9306-c24c92598e29-kube-api-access-xfkfj\") pod \"dnsmasq-dns-847c4cc679-hfr7t\" (UID: \"83360adb-6875-47a7-9306-c24c92598e29\") " pod="openstack/dnsmasq-dns-847c4cc679-hfr7t" Jan 26 17:07:27 crc kubenswrapper[4754]: I0126 17:07:27.596562 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ee429872-adbd-47bb-9ff2-d5d52dba4d12-config-data\") pod \"keystone-bootstrap-6qghh\" (UID: \"ee429872-adbd-47bb-9ff2-d5d52dba4d12\") " pod="openstack/keystone-bootstrap-6qghh" Jan 26 17:07:27 crc kubenswrapper[4754]: I0126 17:07:27.596594 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/83360adb-6875-47a7-9306-c24c92598e29-dns-swift-storage-0\") pod \"dnsmasq-dns-847c4cc679-hfr7t\" (UID: \"83360adb-6875-47a7-9306-c24c92598e29\") " pod="openstack/dnsmasq-dns-847c4cc679-hfr7t" Jan 26 17:07:27 crc kubenswrapper[4754]: I0126 17:07:27.596702 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7lrsm\" (UniqueName: \"kubernetes.io/projected/ee429872-adbd-47bb-9ff2-d5d52dba4d12-kube-api-access-7lrsm\") pod \"keystone-bootstrap-6qghh\" (UID: \"ee429872-adbd-47bb-9ff2-d5d52dba4d12\") " pod="openstack/keystone-bootstrap-6qghh" Jan 26 17:07:27 crc kubenswrapper[4754]: I0126 17:07:27.596719 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/83360adb-6875-47a7-9306-c24c92598e29-ovsdbserver-sb\") pod \"dnsmasq-dns-847c4cc679-hfr7t\" (UID: \"83360adb-6875-47a7-9306-c24c92598e29\") " pod="openstack/dnsmasq-dns-847c4cc679-hfr7t" Jan 26 17:07:27 crc kubenswrapper[4754]: I0126 17:07:27.596742 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/83360adb-6875-47a7-9306-c24c92598e29-dns-svc\") pod \"dnsmasq-dns-847c4cc679-hfr7t\" (UID: \"83360adb-6875-47a7-9306-c24c92598e29\") " pod="openstack/dnsmasq-dns-847c4cc679-hfr7t" Jan 26 17:07:27 crc kubenswrapper[4754]: I0126 17:07:27.596782 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/83360adb-6875-47a7-9306-c24c92598e29-config\") pod \"dnsmasq-dns-847c4cc679-hfr7t\" (UID: \"83360adb-6875-47a7-9306-c24c92598e29\") " pod="openstack/dnsmasq-dns-847c4cc679-hfr7t" Jan 26 17:07:27 crc kubenswrapper[4754]: I0126 17:07:27.596891 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee429872-adbd-47bb-9ff2-d5d52dba4d12-combined-ca-bundle\") pod \"keystone-bootstrap-6qghh\" (UID: \"ee429872-adbd-47bb-9ff2-d5d52dba4d12\") " pod="openstack/keystone-bootstrap-6qghh" Jan 26 17:07:27 crc kubenswrapper[4754]: I0126 17:07:27.596940 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/83360adb-6875-47a7-9306-c24c92598e29-ovsdbserver-nb\") pod \"dnsmasq-dns-847c4cc679-hfr7t\" (UID: \"83360adb-6875-47a7-9306-c24c92598e29\") " pod="openstack/dnsmasq-dns-847c4cc679-hfr7t" Jan 26 17:07:27 crc kubenswrapper[4754]: I0126 17:07:27.597001 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/ee429872-adbd-47bb-9ff2-d5d52dba4d12-credential-keys\") pod \"keystone-bootstrap-6qghh\" (UID: \"ee429872-adbd-47bb-9ff2-d5d52dba4d12\") " pod="openstack/keystone-bootstrap-6qghh" Jan 26 17:07:27 crc kubenswrapper[4754]: I0126 17:07:27.597059 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/ee429872-adbd-47bb-9ff2-d5d52dba4d12-fernet-keys\") pod \"keystone-bootstrap-6qghh\" (UID: \"ee429872-adbd-47bb-9ff2-d5d52dba4d12\") " pod="openstack/keystone-bootstrap-6qghh" Jan 26 17:07:27 crc kubenswrapper[4754]: I0126 17:07:27.609560 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ee429872-adbd-47bb-9ff2-d5d52dba4d12-config-data\") pod \"keystone-bootstrap-6qghh\" (UID: \"ee429872-adbd-47bb-9ff2-d5d52dba4d12\") " pod="openstack/keystone-bootstrap-6qghh" Jan 26 17:07:27 crc kubenswrapper[4754]: I0126 17:07:27.611725 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ee429872-adbd-47bb-9ff2-d5d52dba4d12-scripts\") pod \"keystone-bootstrap-6qghh\" (UID: \"ee429872-adbd-47bb-9ff2-d5d52dba4d12\") " pod="openstack/keystone-bootstrap-6qghh" Jan 26 17:07:27 crc kubenswrapper[4754]: I0126 17:07:27.613039 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee429872-adbd-47bb-9ff2-d5d52dba4d12-combined-ca-bundle\") pod \"keystone-bootstrap-6qghh\" (UID: \"ee429872-adbd-47bb-9ff2-d5d52dba4d12\") " pod="openstack/keystone-bootstrap-6qghh" Jan 26 17:07:27 crc kubenswrapper[4754]: I0126 17:07:27.616207 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/ee429872-adbd-47bb-9ff2-d5d52dba4d12-credential-keys\") pod \"keystone-bootstrap-6qghh\" (UID: \"ee429872-adbd-47bb-9ff2-d5d52dba4d12\") " pod="openstack/keystone-bootstrap-6qghh" Jan 26 17:07:27 crc kubenswrapper[4754]: I0126 17:07:27.623291 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/ee429872-adbd-47bb-9ff2-d5d52dba4d12-fernet-keys\") pod \"keystone-bootstrap-6qghh\" (UID: \"ee429872-adbd-47bb-9ff2-d5d52dba4d12\") " pod="openstack/keystone-bootstrap-6qghh" Jan 26 17:07:27 crc kubenswrapper[4754]: I0126 17:07:27.644386 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7lrsm\" (UniqueName: \"kubernetes.io/projected/ee429872-adbd-47bb-9ff2-d5d52dba4d12-kube-api-access-7lrsm\") pod \"keystone-bootstrap-6qghh\" (UID: \"ee429872-adbd-47bb-9ff2-d5d52dba4d12\") " pod="openstack/keystone-bootstrap-6qghh" Jan 26 17:07:27 crc kubenswrapper[4754]: I0126 17:07:27.709868 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/83360adb-6875-47a7-9306-c24c92598e29-dns-swift-storage-0\") pod \"dnsmasq-dns-847c4cc679-hfr7t\" (UID: \"83360adb-6875-47a7-9306-c24c92598e29\") " pod="openstack/dnsmasq-dns-847c4cc679-hfr7t" Jan 26 17:07:27 crc kubenswrapper[4754]: I0126 17:07:27.710180 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/83360adb-6875-47a7-9306-c24c92598e29-dns-svc\") pod \"dnsmasq-dns-847c4cc679-hfr7t\" (UID: \"83360adb-6875-47a7-9306-c24c92598e29\") " pod="openstack/dnsmasq-dns-847c4cc679-hfr7t" Jan 26 17:07:27 crc kubenswrapper[4754]: I0126 17:07:27.710205 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/83360adb-6875-47a7-9306-c24c92598e29-ovsdbserver-sb\") pod \"dnsmasq-dns-847c4cc679-hfr7t\" (UID: \"83360adb-6875-47a7-9306-c24c92598e29\") " pod="openstack/dnsmasq-dns-847c4cc679-hfr7t" Jan 26 17:07:27 crc kubenswrapper[4754]: I0126 17:07:27.710245 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/83360adb-6875-47a7-9306-c24c92598e29-config\") pod \"dnsmasq-dns-847c4cc679-hfr7t\" (UID: \"83360adb-6875-47a7-9306-c24c92598e29\") " pod="openstack/dnsmasq-dns-847c4cc679-hfr7t" Jan 26 17:07:27 crc kubenswrapper[4754]: I0126 17:07:27.710286 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/83360adb-6875-47a7-9306-c24c92598e29-ovsdbserver-nb\") pod \"dnsmasq-dns-847c4cc679-hfr7t\" (UID: \"83360adb-6875-47a7-9306-c24c92598e29\") " pod="openstack/dnsmasq-dns-847c4cc679-hfr7t" Jan 26 17:07:27 crc kubenswrapper[4754]: I0126 17:07:27.710330 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xfkfj\" (UniqueName: \"kubernetes.io/projected/83360adb-6875-47a7-9306-c24c92598e29-kube-api-access-xfkfj\") pod \"dnsmasq-dns-847c4cc679-hfr7t\" (UID: \"83360adb-6875-47a7-9306-c24c92598e29\") " pod="openstack/dnsmasq-dns-847c4cc679-hfr7t" Jan 26 17:07:27 crc kubenswrapper[4754]: I0126 17:07:27.711455 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/83360adb-6875-47a7-9306-c24c92598e29-dns-swift-storage-0\") pod \"dnsmasq-dns-847c4cc679-hfr7t\" (UID: \"83360adb-6875-47a7-9306-c24c92598e29\") " pod="openstack/dnsmasq-dns-847c4cc679-hfr7t" Jan 26 17:07:27 crc kubenswrapper[4754]: I0126 17:07:27.711989 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/83360adb-6875-47a7-9306-c24c92598e29-dns-svc\") pod \"dnsmasq-dns-847c4cc679-hfr7t\" (UID: \"83360adb-6875-47a7-9306-c24c92598e29\") " pod="openstack/dnsmasq-dns-847c4cc679-hfr7t" Jan 26 17:07:27 crc kubenswrapper[4754]: I0126 17:07:27.712566 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/83360adb-6875-47a7-9306-c24c92598e29-ovsdbserver-sb\") pod \"dnsmasq-dns-847c4cc679-hfr7t\" (UID: \"83360adb-6875-47a7-9306-c24c92598e29\") " pod="openstack/dnsmasq-dns-847c4cc679-hfr7t" Jan 26 17:07:27 crc kubenswrapper[4754]: I0126 17:07:27.713089 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/83360adb-6875-47a7-9306-c24c92598e29-config\") pod \"dnsmasq-dns-847c4cc679-hfr7t\" (UID: \"83360adb-6875-47a7-9306-c24c92598e29\") " pod="openstack/dnsmasq-dns-847c4cc679-hfr7t" Jan 26 17:07:27 crc kubenswrapper[4754]: I0126 17:07:27.713610 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/83360adb-6875-47a7-9306-c24c92598e29-ovsdbserver-nb\") pod \"dnsmasq-dns-847c4cc679-hfr7t\" (UID: \"83360adb-6875-47a7-9306-c24c92598e29\") " pod="openstack/dnsmasq-dns-847c4cc679-hfr7t" Jan 26 17:07:27 crc kubenswrapper[4754]: I0126 17:07:27.731811 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-6db5cbfc7c-nxhnr"] Jan 26 17:07:27 crc kubenswrapper[4754]: I0126 17:07:27.733031 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6db5cbfc7c-nxhnr" Jan 26 17:07:27 crc kubenswrapper[4754]: I0126 17:07:27.735114 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-6qghh" Jan 26 17:07:27 crc kubenswrapper[4754]: I0126 17:07:27.747520 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-config-data" Jan 26 17:07:27 crc kubenswrapper[4754]: I0126 17:07:27.747719 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon" Jan 26 17:07:27 crc kubenswrapper[4754]: I0126 17:07:27.747856 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon-horizon-dockercfg-pxnkq" Jan 26 17:07:27 crc kubenswrapper[4754]: I0126 17:07:27.748046 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-scripts" Jan 26 17:07:27 crc kubenswrapper[4754]: I0126 17:07:27.800294 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-6db5cbfc7c-nxhnr"] Jan 26 17:07:27 crc kubenswrapper[4754]: I0126 17:07:27.832535 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xfkfj\" (UniqueName: \"kubernetes.io/projected/83360adb-6875-47a7-9306-c24c92598e29-kube-api-access-xfkfj\") pod \"dnsmasq-dns-847c4cc679-hfr7t\" (UID: \"83360adb-6875-47a7-9306-c24c92598e29\") " pod="openstack/dnsmasq-dns-847c4cc679-hfr7t" Jan 26 17:07:27 crc kubenswrapper[4754]: I0126 17:07:27.909144 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-7hskr"] Jan 26 17:07:27 crc kubenswrapper[4754]: I0126 17:07:27.910075 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-7hskr" Jan 26 17:07:27 crc kubenswrapper[4754]: I0126 17:07:27.913158 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/743bf5b2-5143-406b-b3d7-fa3a067974fd-logs\") pod \"horizon-6db5cbfc7c-nxhnr\" (UID: \"743bf5b2-5143-406b-b3d7-fa3a067974fd\") " pod="openstack/horizon-6db5cbfc7c-nxhnr" Jan 26 17:07:27 crc kubenswrapper[4754]: I0126 17:07:27.913748 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/743bf5b2-5143-406b-b3d7-fa3a067974fd-config-data\") pod \"horizon-6db5cbfc7c-nxhnr\" (UID: \"743bf5b2-5143-406b-b3d7-fa3a067974fd\") " pod="openstack/horizon-6db5cbfc7c-nxhnr" Jan 26 17:07:27 crc kubenswrapper[4754]: I0126 17:07:27.913783 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2tn9b\" (UniqueName: \"kubernetes.io/projected/743bf5b2-5143-406b-b3d7-fa3a067974fd-kube-api-access-2tn9b\") pod \"horizon-6db5cbfc7c-nxhnr\" (UID: \"743bf5b2-5143-406b-b3d7-fa3a067974fd\") " pod="openstack/horizon-6db5cbfc7c-nxhnr" Jan 26 17:07:27 crc kubenswrapper[4754]: I0126 17:07:27.913831 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/743bf5b2-5143-406b-b3d7-fa3a067974fd-scripts\") pod \"horizon-6db5cbfc7c-nxhnr\" (UID: \"743bf5b2-5143-406b-b3d7-fa3a067974fd\") " pod="openstack/horizon-6db5cbfc7c-nxhnr" Jan 26 17:07:27 crc kubenswrapper[4754]: I0126 17:07:27.913864 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/743bf5b2-5143-406b-b3d7-fa3a067974fd-horizon-secret-key\") pod \"horizon-6db5cbfc7c-nxhnr\" (UID: \"743bf5b2-5143-406b-b3d7-fa3a067974fd\") " pod="openstack/horizon-6db5cbfc7c-nxhnr" Jan 26 17:07:27 crc kubenswrapper[4754]: I0126 17:07:27.945132 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-bp2xk" Jan 26 17:07:27 crc kubenswrapper[4754]: I0126 17:07:27.945359 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Jan 26 17:07:27 crc kubenswrapper[4754]: I0126 17:07:27.945567 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Jan 26 17:07:27 crc kubenswrapper[4754]: I0126 17:07:27.964270 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-dzfm9"] Jan 26 17:07:27 crc kubenswrapper[4754]: I0126 17:07:27.969963 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-dzfm9" Jan 26 17:07:27 crc kubenswrapper[4754]: I0126 17:07:27.976855 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-9z5cf" Jan 26 17:07:27 crc kubenswrapper[4754]: I0126 17:07:27.977207 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Jan 26 17:07:27 crc kubenswrapper[4754]: I0126 17:07:27.977375 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.006422 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-7hskr"] Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.017130 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5gg2w\" (UniqueName: \"kubernetes.io/projected/ab3f008d-78ab-4b73-962d-4447f312c9fa-kube-api-access-5gg2w\") pod \"cinder-db-sync-7hskr\" (UID: \"ab3f008d-78ab-4b73-962d-4447f312c9fa\") " pod="openstack/cinder-db-sync-7hskr" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.017187 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ab3f008d-78ab-4b73-962d-4447f312c9fa-db-sync-config-data\") pod \"cinder-db-sync-7hskr\" (UID: \"ab3f008d-78ab-4b73-962d-4447f312c9fa\") " pod="openstack/cinder-db-sync-7hskr" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.017233 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/743bf5b2-5143-406b-b3d7-fa3a067974fd-config-data\") pod \"horizon-6db5cbfc7c-nxhnr\" (UID: \"743bf5b2-5143-406b-b3d7-fa3a067974fd\") " pod="openstack/horizon-6db5cbfc7c-nxhnr" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.017249 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2tn9b\" (UniqueName: \"kubernetes.io/projected/743bf5b2-5143-406b-b3d7-fa3a067974fd-kube-api-access-2tn9b\") pod \"horizon-6db5cbfc7c-nxhnr\" (UID: \"743bf5b2-5143-406b-b3d7-fa3a067974fd\") " pod="openstack/horizon-6db5cbfc7c-nxhnr" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.017280 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/743bf5b2-5143-406b-b3d7-fa3a067974fd-scripts\") pod \"horizon-6db5cbfc7c-nxhnr\" (UID: \"743bf5b2-5143-406b-b3d7-fa3a067974fd\") " pod="openstack/horizon-6db5cbfc7c-nxhnr" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.017319 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ab3f008d-78ab-4b73-962d-4447f312c9fa-etc-machine-id\") pod \"cinder-db-sync-7hskr\" (UID: \"ab3f008d-78ab-4b73-962d-4447f312c9fa\") " pod="openstack/cinder-db-sync-7hskr" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.017338 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ab3f008d-78ab-4b73-962d-4447f312c9fa-config-data\") pod \"cinder-db-sync-7hskr\" (UID: \"ab3f008d-78ab-4b73-962d-4447f312c9fa\") " pod="openstack/cinder-db-sync-7hskr" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.017364 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/743bf5b2-5143-406b-b3d7-fa3a067974fd-horizon-secret-key\") pod \"horizon-6db5cbfc7c-nxhnr\" (UID: \"743bf5b2-5143-406b-b3d7-fa3a067974fd\") " pod="openstack/horizon-6db5cbfc7c-nxhnr" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.017444 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/743bf5b2-5143-406b-b3d7-fa3a067974fd-logs\") pod \"horizon-6db5cbfc7c-nxhnr\" (UID: \"743bf5b2-5143-406b-b3d7-fa3a067974fd\") " pod="openstack/horizon-6db5cbfc7c-nxhnr" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.017488 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ab3f008d-78ab-4b73-962d-4447f312c9fa-scripts\") pod \"cinder-db-sync-7hskr\" (UID: \"ab3f008d-78ab-4b73-962d-4447f312c9fa\") " pod="openstack/cinder-db-sync-7hskr" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.017551 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab3f008d-78ab-4b73-962d-4447f312c9fa-combined-ca-bundle\") pod \"cinder-db-sync-7hskr\" (UID: \"ab3f008d-78ab-4b73-962d-4447f312c9fa\") " pod="openstack/cinder-db-sync-7hskr" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.020042 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/743bf5b2-5143-406b-b3d7-fa3a067974fd-logs\") pod \"horizon-6db5cbfc7c-nxhnr\" (UID: \"743bf5b2-5143-406b-b3d7-fa3a067974fd\") " pod="openstack/horizon-6db5cbfc7c-nxhnr" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.020979 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/743bf5b2-5143-406b-b3d7-fa3a067974fd-config-data\") pod \"horizon-6db5cbfc7c-nxhnr\" (UID: \"743bf5b2-5143-406b-b3d7-fa3a067974fd\") " pod="openstack/horizon-6db5cbfc7c-nxhnr" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.025988 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/743bf5b2-5143-406b-b3d7-fa3a067974fd-scripts\") pod \"horizon-6db5cbfc7c-nxhnr\" (UID: \"743bf5b2-5143-406b-b3d7-fa3a067974fd\") " pod="openstack/horizon-6db5cbfc7c-nxhnr" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.030505 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-dzfm9"] Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.048350 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/743bf5b2-5143-406b-b3d7-fa3a067974fd-horizon-secret-key\") pod \"horizon-6db5cbfc7c-nxhnr\" (UID: \"743bf5b2-5143-406b-b3d7-fa3a067974fd\") " pod="openstack/horizon-6db5cbfc7c-nxhnr" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.052210 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-847c4cc679-hfr7t" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.080470 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-whnhn"] Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.081315 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2tn9b\" (UniqueName: \"kubernetes.io/projected/743bf5b2-5143-406b-b3d7-fa3a067974fd-kube-api-access-2tn9b\") pod \"horizon-6db5cbfc7c-nxhnr\" (UID: \"743bf5b2-5143-406b-b3d7-fa3a067974fd\") " pod="openstack/horizon-6db5cbfc7c-nxhnr" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.082221 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-whnhn" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.098748 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.098883 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.122254 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ptw8x\" (UniqueName: \"kubernetes.io/projected/7d5a6ee0-79d3-4112-a180-d2211fee09f6-kube-api-access-ptw8x\") pod \"neutron-db-sync-dzfm9\" (UID: \"7d5a6ee0-79d3-4112-a180-d2211fee09f6\") " pod="openstack/neutron-db-sync-dzfm9" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.122327 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d5a6ee0-79d3-4112-a180-d2211fee09f6-combined-ca-bundle\") pod \"neutron-db-sync-dzfm9\" (UID: \"7d5a6ee0-79d3-4112-a180-d2211fee09f6\") " pod="openstack/neutron-db-sync-dzfm9" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.122355 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ab3f008d-78ab-4b73-962d-4447f312c9fa-scripts\") pod \"cinder-db-sync-7hskr\" (UID: \"ab3f008d-78ab-4b73-962d-4447f312c9fa\") " pod="openstack/cinder-db-sync-7hskr" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.122389 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab3f008d-78ab-4b73-962d-4447f312c9fa-combined-ca-bundle\") pod \"cinder-db-sync-7hskr\" (UID: \"ab3f008d-78ab-4b73-962d-4447f312c9fa\") " pod="openstack/cinder-db-sync-7hskr" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.126983 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/7d5a6ee0-79d3-4112-a180-d2211fee09f6-config\") pod \"neutron-db-sync-dzfm9\" (UID: \"7d5a6ee0-79d3-4112-a180-d2211fee09f6\") " pod="openstack/neutron-db-sync-dzfm9" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.128697 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5gg2w\" (UniqueName: \"kubernetes.io/projected/ab3f008d-78ab-4b73-962d-4447f312c9fa-kube-api-access-5gg2w\") pod \"cinder-db-sync-7hskr\" (UID: \"ab3f008d-78ab-4b73-962d-4447f312c9fa\") " pod="openstack/cinder-db-sync-7hskr" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.128748 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ab3f008d-78ab-4b73-962d-4447f312c9fa-db-sync-config-data\") pod \"cinder-db-sync-7hskr\" (UID: \"ab3f008d-78ab-4b73-962d-4447f312c9fa\") " pod="openstack/cinder-db-sync-7hskr" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.128833 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ab3f008d-78ab-4b73-962d-4447f312c9fa-etc-machine-id\") pod \"cinder-db-sync-7hskr\" (UID: \"ab3f008d-78ab-4b73-962d-4447f312c9fa\") " pod="openstack/cinder-db-sync-7hskr" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.128863 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ab3f008d-78ab-4b73-962d-4447f312c9fa-config-data\") pod \"cinder-db-sync-7hskr\" (UID: \"ab3f008d-78ab-4b73-962d-4447f312c9fa\") " pod="openstack/cinder-db-sync-7hskr" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.138237 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ab3f008d-78ab-4b73-962d-4447f312c9fa-config-data\") pod \"cinder-db-sync-7hskr\" (UID: \"ab3f008d-78ab-4b73-962d-4447f312c9fa\") " pod="openstack/cinder-db-sync-7hskr" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.139807 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-92svj" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.141518 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ab3f008d-78ab-4b73-962d-4447f312c9fa-etc-machine-id\") pod \"cinder-db-sync-7hskr\" (UID: \"ab3f008d-78ab-4b73-962d-4447f312c9fa\") " pod="openstack/cinder-db-sync-7hskr" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.144658 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ab3f008d-78ab-4b73-962d-4447f312c9fa-db-sync-config-data\") pod \"cinder-db-sync-7hskr\" (UID: \"ab3f008d-78ab-4b73-962d-4447f312c9fa\") " pod="openstack/cinder-db-sync-7hskr" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.152987 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.162847 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab3f008d-78ab-4b73-962d-4447f312c9fa-combined-ca-bundle\") pod \"cinder-db-sync-7hskr\" (UID: \"ab3f008d-78ab-4b73-962d-4447f312c9fa\") " pod="openstack/cinder-db-sync-7hskr" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.163576 4754 generic.go:334] "Generic (PLEG): container finished" podID="a6873959-d867-4a38-a10f-82ff80d0800b" containerID="7c2a92f25263e2691e40742a8be4dc87697b019ee912e3a15e8919c7138e41da" exitCode=0 Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.163940 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74f6bcbc87-q8pwj" event={"ID":"a6873959-d867-4a38-a10f-82ff80d0800b","Type":"ContainerDied","Data":"7c2a92f25263e2691e40742a8be4dc87697b019ee912e3a15e8919c7138e41da"} Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.164188 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.169903 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ab3f008d-78ab-4b73-962d-4447f312c9fa-scripts\") pod \"cinder-db-sync-7hskr\" (UID: \"ab3f008d-78ab-4b73-962d-4447f312c9fa\") " pod="openstack/cinder-db-sync-7hskr" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.173700 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.173940 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.176294 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5gg2w\" (UniqueName: \"kubernetes.io/projected/ab3f008d-78ab-4b73-962d-4447f312c9fa-kube-api-access-5gg2w\") pod \"cinder-db-sync-7hskr\" (UID: \"ab3f008d-78ab-4b73-962d-4447f312c9fa\") " pod="openstack/cinder-db-sync-7hskr" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.218244 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6db5cbfc7c-nxhnr" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.230743 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-7hskr" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.232470 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01246f69-d6bc-4f70-a50f-4d7c8e4a7620-combined-ca-bundle\") pod \"placement-db-sync-whnhn\" (UID: \"01246f69-d6bc-4f70-a50f-4d7c8e4a7620\") " pod="openstack/placement-db-sync-whnhn" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.238452 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ptw8x\" (UniqueName: \"kubernetes.io/projected/7d5a6ee0-79d3-4112-a180-d2211fee09f6-kube-api-access-ptw8x\") pod \"neutron-db-sync-dzfm9\" (UID: \"7d5a6ee0-79d3-4112-a180-d2211fee09f6\") " pod="openstack/neutron-db-sync-dzfm9" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.238537 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d84d5918-51f5-4131-9caf-9380004a0ca5-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d84d5918-51f5-4131-9caf-9380004a0ca5\") " pod="openstack/ceilometer-0" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.238576 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d84d5918-51f5-4131-9caf-9380004a0ca5-scripts\") pod \"ceilometer-0\" (UID: \"d84d5918-51f5-4131-9caf-9380004a0ca5\") " pod="openstack/ceilometer-0" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.238594 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d84d5918-51f5-4131-9caf-9380004a0ca5-run-httpd\") pod \"ceilometer-0\" (UID: \"d84d5918-51f5-4131-9caf-9380004a0ca5\") " pod="openstack/ceilometer-0" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.238646 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d5a6ee0-79d3-4112-a180-d2211fee09f6-combined-ca-bundle\") pod \"neutron-db-sync-dzfm9\" (UID: \"7d5a6ee0-79d3-4112-a180-d2211fee09f6\") " pod="openstack/neutron-db-sync-dzfm9" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.238728 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d84d5918-51f5-4131-9caf-9380004a0ca5-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d84d5918-51f5-4131-9caf-9380004a0ca5\") " pod="openstack/ceilometer-0" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.238760 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d84d5918-51f5-4131-9caf-9380004a0ca5-log-httpd\") pod \"ceilometer-0\" (UID: \"d84d5918-51f5-4131-9caf-9380004a0ca5\") " pod="openstack/ceilometer-0" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.238817 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/7d5a6ee0-79d3-4112-a180-d2211fee09f6-config\") pod \"neutron-db-sync-dzfm9\" (UID: \"7d5a6ee0-79d3-4112-a180-d2211fee09f6\") " pod="openstack/neutron-db-sync-dzfm9" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.238851 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/01246f69-d6bc-4f70-a50f-4d7c8e4a7620-logs\") pod \"placement-db-sync-whnhn\" (UID: \"01246f69-d6bc-4f70-a50f-4d7c8e4a7620\") " pod="openstack/placement-db-sync-whnhn" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.238879 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2bpjj\" (UniqueName: \"kubernetes.io/projected/d84d5918-51f5-4131-9caf-9380004a0ca5-kube-api-access-2bpjj\") pod \"ceilometer-0\" (UID: \"d84d5918-51f5-4131-9caf-9380004a0ca5\") " pod="openstack/ceilometer-0" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.238902 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/01246f69-d6bc-4f70-a50f-4d7c8e4a7620-config-data\") pod \"placement-db-sync-whnhn\" (UID: \"01246f69-d6bc-4f70-a50f-4d7c8e4a7620\") " pod="openstack/placement-db-sync-whnhn" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.238997 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/01246f69-d6bc-4f70-a50f-4d7c8e4a7620-scripts\") pod \"placement-db-sync-whnhn\" (UID: \"01246f69-d6bc-4f70-a50f-4d7c8e4a7620\") " pod="openstack/placement-db-sync-whnhn" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.239031 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rr5pm\" (UniqueName: \"kubernetes.io/projected/01246f69-d6bc-4f70-a50f-4d7c8e4a7620-kube-api-access-rr5pm\") pod \"placement-db-sync-whnhn\" (UID: \"01246f69-d6bc-4f70-a50f-4d7c8e4a7620\") " pod="openstack/placement-db-sync-whnhn" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.239085 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d84d5918-51f5-4131-9caf-9380004a0ca5-config-data\") pod \"ceilometer-0\" (UID: \"d84d5918-51f5-4131-9caf-9380004a0ca5\") " pod="openstack/ceilometer-0" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.244163 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-hfr7t"] Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.266179 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d5a6ee0-79d3-4112-a180-d2211fee09f6-combined-ca-bundle\") pod \"neutron-db-sync-dzfm9\" (UID: \"7d5a6ee0-79d3-4112-a180-d2211fee09f6\") " pod="openstack/neutron-db-sync-dzfm9" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.270982 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/7d5a6ee0-79d3-4112-a180-d2211fee09f6-config\") pod \"neutron-db-sync-dzfm9\" (UID: \"7d5a6ee0-79d3-4112-a180-d2211fee09f6\") " pod="openstack/neutron-db-sync-dzfm9" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.274723 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-whnhn"] Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.278226 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ptw8x\" (UniqueName: \"kubernetes.io/projected/7d5a6ee0-79d3-4112-a180-d2211fee09f6-kube-api-access-ptw8x\") pod \"neutron-db-sync-dzfm9\" (UID: \"7d5a6ee0-79d3-4112-a180-d2211fee09f6\") " pod="openstack/neutron-db-sync-dzfm9" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.289790 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-585cbdbc67-qwfhd"] Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.291331 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-585cbdbc67-qwfhd" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.294906 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.307005 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-dzfm9" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.307749 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-9spps"] Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.309139 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-9spps" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.312134 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.315999 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-sfxbx" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.340025 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d84d5918-51f5-4131-9caf-9380004a0ca5-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d84d5918-51f5-4131-9caf-9380004a0ca5\") " pod="openstack/ceilometer-0" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.340060 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d84d5918-51f5-4131-9caf-9380004a0ca5-log-httpd\") pod \"ceilometer-0\" (UID: \"d84d5918-51f5-4131-9caf-9380004a0ca5\") " pod="openstack/ceilometer-0" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.340108 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/eabead53-af6d-4405-b76b-a20c5a84f469-logs\") pod \"horizon-585cbdbc67-qwfhd\" (UID: \"eabead53-af6d-4405-b76b-a20c5a84f469\") " pod="openstack/horizon-585cbdbc67-qwfhd" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.340126 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/01246f69-d6bc-4f70-a50f-4d7c8e4a7620-logs\") pod \"placement-db-sync-whnhn\" (UID: \"01246f69-d6bc-4f70-a50f-4d7c8e4a7620\") " pod="openstack/placement-db-sync-whnhn" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.340144 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/01246f69-d6bc-4f70-a50f-4d7c8e4a7620-config-data\") pod \"placement-db-sync-whnhn\" (UID: \"01246f69-d6bc-4f70-a50f-4d7c8e4a7620\") " pod="openstack/placement-db-sync-whnhn" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.340160 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2bpjj\" (UniqueName: \"kubernetes.io/projected/d84d5918-51f5-4131-9caf-9380004a0ca5-kube-api-access-2bpjj\") pod \"ceilometer-0\" (UID: \"d84d5918-51f5-4131-9caf-9380004a0ca5\") " pod="openstack/ceilometer-0" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.340195 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/eabead53-af6d-4405-b76b-a20c5a84f469-scripts\") pod \"horizon-585cbdbc67-qwfhd\" (UID: \"eabead53-af6d-4405-b76b-a20c5a84f469\") " pod="openstack/horizon-585cbdbc67-qwfhd" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.340212 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/01246f69-d6bc-4f70-a50f-4d7c8e4a7620-scripts\") pod \"placement-db-sync-whnhn\" (UID: \"01246f69-d6bc-4f70-a50f-4d7c8e4a7620\") " pod="openstack/placement-db-sync-whnhn" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.340231 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rr5pm\" (UniqueName: \"kubernetes.io/projected/01246f69-d6bc-4f70-a50f-4d7c8e4a7620-kube-api-access-rr5pm\") pod \"placement-db-sync-whnhn\" (UID: \"01246f69-d6bc-4f70-a50f-4d7c8e4a7620\") " pod="openstack/placement-db-sync-whnhn" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.340248 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/eabead53-af6d-4405-b76b-a20c5a84f469-config-data\") pod \"horizon-585cbdbc67-qwfhd\" (UID: \"eabead53-af6d-4405-b76b-a20c5a84f469\") " pod="openstack/horizon-585cbdbc67-qwfhd" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.340270 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d84d5918-51f5-4131-9caf-9380004a0ca5-config-data\") pod \"ceilometer-0\" (UID: \"d84d5918-51f5-4131-9caf-9380004a0ca5\") " pod="openstack/ceilometer-0" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.340289 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/eabead53-af6d-4405-b76b-a20c5a84f469-horizon-secret-key\") pod \"horizon-585cbdbc67-qwfhd\" (UID: \"eabead53-af6d-4405-b76b-a20c5a84f469\") " pod="openstack/horizon-585cbdbc67-qwfhd" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.340311 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ea094395-bb3c-4464-829f-313ab6b9e14c-combined-ca-bundle\") pod \"barbican-db-sync-9spps\" (UID: \"ea094395-bb3c-4464-829f-313ab6b9e14c\") " pod="openstack/barbican-db-sync-9spps" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.340330 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ea094395-bb3c-4464-829f-313ab6b9e14c-db-sync-config-data\") pod \"barbican-db-sync-9spps\" (UID: \"ea094395-bb3c-4464-829f-313ab6b9e14c\") " pod="openstack/barbican-db-sync-9spps" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.340350 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01246f69-d6bc-4f70-a50f-4d7c8e4a7620-combined-ca-bundle\") pod \"placement-db-sync-whnhn\" (UID: \"01246f69-d6bc-4f70-a50f-4d7c8e4a7620\") " pod="openstack/placement-db-sync-whnhn" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.340376 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c6spp\" (UniqueName: \"kubernetes.io/projected/eabead53-af6d-4405-b76b-a20c5a84f469-kube-api-access-c6spp\") pod \"horizon-585cbdbc67-qwfhd\" (UID: \"eabead53-af6d-4405-b76b-a20c5a84f469\") " pod="openstack/horizon-585cbdbc67-qwfhd" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.340409 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d84d5918-51f5-4131-9caf-9380004a0ca5-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d84d5918-51f5-4131-9caf-9380004a0ca5\") " pod="openstack/ceilometer-0" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.340425 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d84d5918-51f5-4131-9caf-9380004a0ca5-scripts\") pod \"ceilometer-0\" (UID: \"d84d5918-51f5-4131-9caf-9380004a0ca5\") " pod="openstack/ceilometer-0" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.340439 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d84d5918-51f5-4131-9caf-9380004a0ca5-run-httpd\") pod \"ceilometer-0\" (UID: \"d84d5918-51f5-4131-9caf-9380004a0ca5\") " pod="openstack/ceilometer-0" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.340455 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wk6gz\" (UniqueName: \"kubernetes.io/projected/ea094395-bb3c-4464-829f-313ab6b9e14c-kube-api-access-wk6gz\") pod \"barbican-db-sync-9spps\" (UID: \"ea094395-bb3c-4464-829f-313ab6b9e14c\") " pod="openstack/barbican-db-sync-9spps" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.340882 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d84d5918-51f5-4131-9caf-9380004a0ca5-log-httpd\") pod \"ceilometer-0\" (UID: \"d84d5918-51f5-4131-9caf-9380004a0ca5\") " pod="openstack/ceilometer-0" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.346891 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/01246f69-d6bc-4f70-a50f-4d7c8e4a7620-logs\") pod \"placement-db-sync-whnhn\" (UID: \"01246f69-d6bc-4f70-a50f-4d7c8e4a7620\") " pod="openstack/placement-db-sync-whnhn" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.353343 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d84d5918-51f5-4131-9caf-9380004a0ca5-run-httpd\") pod \"ceilometer-0\" (UID: \"d84d5918-51f5-4131-9caf-9380004a0ca5\") " pod="openstack/ceilometer-0" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.353388 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-585cbdbc67-qwfhd"] Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.354628 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d84d5918-51f5-4131-9caf-9380004a0ca5-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d84d5918-51f5-4131-9caf-9380004a0ca5\") " pod="openstack/ceilometer-0" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.354882 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d84d5918-51f5-4131-9caf-9380004a0ca5-scripts\") pod \"ceilometer-0\" (UID: \"d84d5918-51f5-4131-9caf-9380004a0ca5\") " pod="openstack/ceilometer-0" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.355959 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/01246f69-d6bc-4f70-a50f-4d7c8e4a7620-scripts\") pod \"placement-db-sync-whnhn\" (UID: \"01246f69-d6bc-4f70-a50f-4d7c8e4a7620\") " pod="openstack/placement-db-sync-whnhn" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.356323 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01246f69-d6bc-4f70-a50f-4d7c8e4a7620-combined-ca-bundle\") pod \"placement-db-sync-whnhn\" (UID: \"01246f69-d6bc-4f70-a50f-4d7c8e4a7620\") " pod="openstack/placement-db-sync-whnhn" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.357782 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/01246f69-d6bc-4f70-a50f-4d7c8e4a7620-config-data\") pod \"placement-db-sync-whnhn\" (UID: \"01246f69-d6bc-4f70-a50f-4d7c8e4a7620\") " pod="openstack/placement-db-sync-whnhn" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.359230 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d84d5918-51f5-4131-9caf-9380004a0ca5-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d84d5918-51f5-4131-9caf-9380004a0ca5\") " pod="openstack/ceilometer-0" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.363222 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d84d5918-51f5-4131-9caf-9380004a0ca5-config-data\") pod \"ceilometer-0\" (UID: \"d84d5918-51f5-4131-9caf-9380004a0ca5\") " pod="openstack/ceilometer-0" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.363279 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-rjqkz"] Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.366975 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-rjqkz" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.369978 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2bpjj\" (UniqueName: \"kubernetes.io/projected/d84d5918-51f5-4131-9caf-9380004a0ca5-kube-api-access-2bpjj\") pod \"ceilometer-0\" (UID: \"d84d5918-51f5-4131-9caf-9380004a0ca5\") " pod="openstack/ceilometer-0" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.370652 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-9spps"] Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.382321 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rr5pm\" (UniqueName: \"kubernetes.io/projected/01246f69-d6bc-4f70-a50f-4d7c8e4a7620-kube-api-access-rr5pm\") pod \"placement-db-sync-whnhn\" (UID: \"01246f69-d6bc-4f70-a50f-4d7c8e4a7620\") " pod="openstack/placement-db-sync-whnhn" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.392954 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-rjqkz"] Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.423242 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.424547 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.430047 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.430253 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.430271 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-hph75" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.430503 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.432210 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.442976 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/eabead53-af6d-4405-b76b-a20c5a84f469-scripts\") pod \"horizon-585cbdbc67-qwfhd\" (UID: \"eabead53-af6d-4405-b76b-a20c5a84f469\") " pod="openstack/horizon-585cbdbc67-qwfhd" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.443027 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/eabead53-af6d-4405-b76b-a20c5a84f469-config-data\") pod \"horizon-585cbdbc67-qwfhd\" (UID: \"eabead53-af6d-4405-b76b-a20c5a84f469\") " pod="openstack/horizon-585cbdbc67-qwfhd" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.443062 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/eabead53-af6d-4405-b76b-a20c5a84f469-horizon-secret-key\") pod \"horizon-585cbdbc67-qwfhd\" (UID: \"eabead53-af6d-4405-b76b-a20c5a84f469\") " pod="openstack/horizon-585cbdbc67-qwfhd" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.443080 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ea094395-bb3c-4464-829f-313ab6b9e14c-combined-ca-bundle\") pod \"barbican-db-sync-9spps\" (UID: \"ea094395-bb3c-4464-829f-313ab6b9e14c\") " pod="openstack/barbican-db-sync-9spps" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.443097 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ea094395-bb3c-4464-829f-313ab6b9e14c-db-sync-config-data\") pod \"barbican-db-sync-9spps\" (UID: \"ea094395-bb3c-4464-829f-313ab6b9e14c\") " pod="openstack/barbican-db-sync-9spps" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.443120 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c6spp\" (UniqueName: \"kubernetes.io/projected/eabead53-af6d-4405-b76b-a20c5a84f469-kube-api-access-c6spp\") pod \"horizon-585cbdbc67-qwfhd\" (UID: \"eabead53-af6d-4405-b76b-a20c5a84f469\") " pod="openstack/horizon-585cbdbc67-qwfhd" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.443158 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wk6gz\" (UniqueName: \"kubernetes.io/projected/ea094395-bb3c-4464-829f-313ab6b9e14c-kube-api-access-wk6gz\") pod \"barbican-db-sync-9spps\" (UID: \"ea094395-bb3c-4464-829f-313ab6b9e14c\") " pod="openstack/barbican-db-sync-9spps" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.443206 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/eabead53-af6d-4405-b76b-a20c5a84f469-logs\") pod \"horizon-585cbdbc67-qwfhd\" (UID: \"eabead53-af6d-4405-b76b-a20c5a84f469\") " pod="openstack/horizon-585cbdbc67-qwfhd" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.443749 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/eabead53-af6d-4405-b76b-a20c5a84f469-logs\") pod \"horizon-585cbdbc67-qwfhd\" (UID: \"eabead53-af6d-4405-b76b-a20c5a84f469\") " pod="openstack/horizon-585cbdbc67-qwfhd" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.444560 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/eabead53-af6d-4405-b76b-a20c5a84f469-config-data\") pod \"horizon-585cbdbc67-qwfhd\" (UID: \"eabead53-af6d-4405-b76b-a20c5a84f469\") " pod="openstack/horizon-585cbdbc67-qwfhd" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.444582 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/eabead53-af6d-4405-b76b-a20c5a84f469-scripts\") pod \"horizon-585cbdbc67-qwfhd\" (UID: \"eabead53-af6d-4405-b76b-a20c5a84f469\") " pod="openstack/horizon-585cbdbc67-qwfhd" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.448829 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/eabead53-af6d-4405-b76b-a20c5a84f469-horizon-secret-key\") pod \"horizon-585cbdbc67-qwfhd\" (UID: \"eabead53-af6d-4405-b76b-a20c5a84f469\") " pod="openstack/horizon-585cbdbc67-qwfhd" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.451144 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ea094395-bb3c-4464-829f-313ab6b9e14c-db-sync-config-data\") pod \"barbican-db-sync-9spps\" (UID: \"ea094395-bb3c-4464-829f-313ab6b9e14c\") " pod="openstack/barbican-db-sync-9spps" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.453537 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ea094395-bb3c-4464-829f-313ab6b9e14c-combined-ca-bundle\") pod \"barbican-db-sync-9spps\" (UID: \"ea094395-bb3c-4464-829f-313ab6b9e14c\") " pod="openstack/barbican-db-sync-9spps" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.454001 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-whnhn" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.471165 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.476650 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.484592 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c6spp\" (UniqueName: \"kubernetes.io/projected/eabead53-af6d-4405-b76b-a20c5a84f469-kube-api-access-c6spp\") pod \"horizon-585cbdbc67-qwfhd\" (UID: \"eabead53-af6d-4405-b76b-a20c5a84f469\") " pod="openstack/horizon-585cbdbc67-qwfhd" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.485203 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.485469 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.493584 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.505744 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.519936 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wk6gz\" (UniqueName: \"kubernetes.io/projected/ea094395-bb3c-4464-829f-313ab6b9e14c-kube-api-access-wk6gz\") pod \"barbican-db-sync-9spps\" (UID: \"ea094395-bb3c-4464-829f-313ab6b9e14c\") " pod="openstack/barbican-db-sync-9spps" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.546503 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nd7x6\" (UniqueName: \"kubernetes.io/projected/bcc405ec-8acd-4fd7-bb9e-f3ab471d03ec-kube-api-access-nd7x6\") pod \"dnsmasq-dns-785d8bcb8c-rjqkz\" (UID: \"bcc405ec-8acd-4fd7-bb9e-f3ab471d03ec\") " pod="openstack/dnsmasq-dns-785d8bcb8c-rjqkz" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.546576 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/522901db-3c0d-4cf5-a534-690eac9ae4db-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"522901db-3c0d-4cf5-a534-690eac9ae4db\") " pod="openstack/glance-default-external-api-0" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.546622 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/522901db-3c0d-4cf5-a534-690eac9ae4db-logs\") pod \"glance-default-external-api-0\" (UID: \"522901db-3c0d-4cf5-a534-690eac9ae4db\") " pod="openstack/glance-default-external-api-0" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.546690 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b646p\" (UniqueName: \"kubernetes.io/projected/522901db-3c0d-4cf5-a534-690eac9ae4db-kube-api-access-b646p\") pod \"glance-default-external-api-0\" (UID: \"522901db-3c0d-4cf5-a534-690eac9ae4db\") " pod="openstack/glance-default-external-api-0" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.546719 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/522901db-3c0d-4cf5-a534-690eac9ae4db-scripts\") pod \"glance-default-external-api-0\" (UID: \"522901db-3c0d-4cf5-a534-690eac9ae4db\") " pod="openstack/glance-default-external-api-0" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.546774 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/522901db-3c0d-4cf5-a534-690eac9ae4db-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"522901db-3c0d-4cf5-a534-690eac9ae4db\") " pod="openstack/glance-default-external-api-0" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.546799 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bcc405ec-8acd-4fd7-bb9e-f3ab471d03ec-config\") pod \"dnsmasq-dns-785d8bcb8c-rjqkz\" (UID: \"bcc405ec-8acd-4fd7-bb9e-f3ab471d03ec\") " pod="openstack/dnsmasq-dns-785d8bcb8c-rjqkz" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.546830 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bcc405ec-8acd-4fd7-bb9e-f3ab471d03ec-ovsdbserver-sb\") pod \"dnsmasq-dns-785d8bcb8c-rjqkz\" (UID: \"bcc405ec-8acd-4fd7-bb9e-f3ab471d03ec\") " pod="openstack/dnsmasq-dns-785d8bcb8c-rjqkz" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.546850 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bcc405ec-8acd-4fd7-bb9e-f3ab471d03ec-dns-svc\") pod \"dnsmasq-dns-785d8bcb8c-rjqkz\" (UID: \"bcc405ec-8acd-4fd7-bb9e-f3ab471d03ec\") " pod="openstack/dnsmasq-dns-785d8bcb8c-rjqkz" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.546883 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bcc405ec-8acd-4fd7-bb9e-f3ab471d03ec-ovsdbserver-nb\") pod \"dnsmasq-dns-785d8bcb8c-rjqkz\" (UID: \"bcc405ec-8acd-4fd7-bb9e-f3ab471d03ec\") " pod="openstack/dnsmasq-dns-785d8bcb8c-rjqkz" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.546908 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"522901db-3c0d-4cf5-a534-690eac9ae4db\") " pod="openstack/glance-default-external-api-0" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.546940 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/522901db-3c0d-4cf5-a534-690eac9ae4db-config-data\") pod \"glance-default-external-api-0\" (UID: \"522901db-3c0d-4cf5-a534-690eac9ae4db\") " pod="openstack/glance-default-external-api-0" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.546966 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/522901db-3c0d-4cf5-a534-690eac9ae4db-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"522901db-3c0d-4cf5-a534-690eac9ae4db\") " pod="openstack/glance-default-external-api-0" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.546986 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bcc405ec-8acd-4fd7-bb9e-f3ab471d03ec-dns-swift-storage-0\") pod \"dnsmasq-dns-785d8bcb8c-rjqkz\" (UID: \"bcc405ec-8acd-4fd7-bb9e-f3ab471d03ec\") " pod="openstack/dnsmasq-dns-785d8bcb8c-rjqkz" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.630468 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-585cbdbc67-qwfhd" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.647885 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/522901db-3c0d-4cf5-a534-690eac9ae4db-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"522901db-3c0d-4cf5-a534-690eac9ae4db\") " pod="openstack/glance-default-external-api-0" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.647934 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bcc405ec-8acd-4fd7-bb9e-f3ab471d03ec-config\") pod \"dnsmasq-dns-785d8bcb8c-rjqkz\" (UID: \"bcc405ec-8acd-4fd7-bb9e-f3ab471d03ec\") " pod="openstack/dnsmasq-dns-785d8bcb8c-rjqkz" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.647959 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bcc405ec-8acd-4fd7-bb9e-f3ab471d03ec-ovsdbserver-sb\") pod \"dnsmasq-dns-785d8bcb8c-rjqkz\" (UID: \"bcc405ec-8acd-4fd7-bb9e-f3ab471d03ec\") " pod="openstack/dnsmasq-dns-785d8bcb8c-rjqkz" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.647979 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bcc405ec-8acd-4fd7-bb9e-f3ab471d03ec-dns-svc\") pod \"dnsmasq-dns-785d8bcb8c-rjqkz\" (UID: \"bcc405ec-8acd-4fd7-bb9e-f3ab471d03ec\") " pod="openstack/dnsmasq-dns-785d8bcb8c-rjqkz" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.648003 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bcc405ec-8acd-4fd7-bb9e-f3ab471d03ec-ovsdbserver-nb\") pod \"dnsmasq-dns-785d8bcb8c-rjqkz\" (UID: \"bcc405ec-8acd-4fd7-bb9e-f3ab471d03ec\") " pod="openstack/dnsmasq-dns-785d8bcb8c-rjqkz" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.648023 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"522901db-3c0d-4cf5-a534-690eac9ae4db\") " pod="openstack/glance-default-external-api-0" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.648049 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-66r5j\" (UniqueName: \"kubernetes.io/projected/15b16489-4c64-4a41-bee1-c590e38a87e5-kube-api-access-66r5j\") pod \"glance-default-internal-api-0\" (UID: \"15b16489-4c64-4a41-bee1-c590e38a87e5\") " pod="openstack/glance-default-internal-api-0" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.648069 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/15b16489-4c64-4a41-bee1-c590e38a87e5-logs\") pod \"glance-default-internal-api-0\" (UID: \"15b16489-4c64-4a41-bee1-c590e38a87e5\") " pod="openstack/glance-default-internal-api-0" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.648092 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/522901db-3c0d-4cf5-a534-690eac9ae4db-config-data\") pod \"glance-default-external-api-0\" (UID: \"522901db-3c0d-4cf5-a534-690eac9ae4db\") " pod="openstack/glance-default-external-api-0" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.648109 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/15b16489-4c64-4a41-bee1-c590e38a87e5-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"15b16489-4c64-4a41-bee1-c590e38a87e5\") " pod="openstack/glance-default-internal-api-0" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.648127 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/15b16489-4c64-4a41-bee1-c590e38a87e5-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"15b16489-4c64-4a41-bee1-c590e38a87e5\") " pod="openstack/glance-default-internal-api-0" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.648145 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bcc405ec-8acd-4fd7-bb9e-f3ab471d03ec-dns-swift-storage-0\") pod \"dnsmasq-dns-785d8bcb8c-rjqkz\" (UID: \"bcc405ec-8acd-4fd7-bb9e-f3ab471d03ec\") " pod="openstack/dnsmasq-dns-785d8bcb8c-rjqkz" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.648162 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/522901db-3c0d-4cf5-a534-690eac9ae4db-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"522901db-3c0d-4cf5-a534-690eac9ae4db\") " pod="openstack/glance-default-external-api-0" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.648182 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/15b16489-4c64-4a41-bee1-c590e38a87e5-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"15b16489-4c64-4a41-bee1-c590e38a87e5\") " pod="openstack/glance-default-internal-api-0" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.648222 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nd7x6\" (UniqueName: \"kubernetes.io/projected/bcc405ec-8acd-4fd7-bb9e-f3ab471d03ec-kube-api-access-nd7x6\") pod \"dnsmasq-dns-785d8bcb8c-rjqkz\" (UID: \"bcc405ec-8acd-4fd7-bb9e-f3ab471d03ec\") " pod="openstack/dnsmasq-dns-785d8bcb8c-rjqkz" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.648244 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/522901db-3c0d-4cf5-a534-690eac9ae4db-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"522901db-3c0d-4cf5-a534-690eac9ae4db\") " pod="openstack/glance-default-external-api-0" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.648262 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/522901db-3c0d-4cf5-a534-690eac9ae4db-logs\") pod \"glance-default-external-api-0\" (UID: \"522901db-3c0d-4cf5-a534-690eac9ae4db\") " pod="openstack/glance-default-external-api-0" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.648277 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/15b16489-4c64-4a41-bee1-c590e38a87e5-config-data\") pod \"glance-default-internal-api-0\" (UID: \"15b16489-4c64-4a41-bee1-c590e38a87e5\") " pod="openstack/glance-default-internal-api-0" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.648302 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b646p\" (UniqueName: \"kubernetes.io/projected/522901db-3c0d-4cf5-a534-690eac9ae4db-kube-api-access-b646p\") pod \"glance-default-external-api-0\" (UID: \"522901db-3c0d-4cf5-a534-690eac9ae4db\") " pod="openstack/glance-default-external-api-0" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.648317 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/522901db-3c0d-4cf5-a534-690eac9ae4db-scripts\") pod \"glance-default-external-api-0\" (UID: \"522901db-3c0d-4cf5-a534-690eac9ae4db\") " pod="openstack/glance-default-external-api-0" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.648317 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/522901db-3c0d-4cf5-a534-690eac9ae4db-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"522901db-3c0d-4cf5-a534-690eac9ae4db\") " pod="openstack/glance-default-external-api-0" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.648337 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"15b16489-4c64-4a41-bee1-c590e38a87e5\") " pod="openstack/glance-default-internal-api-0" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.648353 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/15b16489-4c64-4a41-bee1-c590e38a87e5-scripts\") pod \"glance-default-internal-api-0\" (UID: \"15b16489-4c64-4a41-bee1-c590e38a87e5\") " pod="openstack/glance-default-internal-api-0" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.649055 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bcc405ec-8acd-4fd7-bb9e-f3ab471d03ec-ovsdbserver-sb\") pod \"dnsmasq-dns-785d8bcb8c-rjqkz\" (UID: \"bcc405ec-8acd-4fd7-bb9e-f3ab471d03ec\") " pod="openstack/dnsmasq-dns-785d8bcb8c-rjqkz" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.649091 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bcc405ec-8acd-4fd7-bb9e-f3ab471d03ec-config\") pod \"dnsmasq-dns-785d8bcb8c-rjqkz\" (UID: \"bcc405ec-8acd-4fd7-bb9e-f3ab471d03ec\") " pod="openstack/dnsmasq-dns-785d8bcb8c-rjqkz" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.649198 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bcc405ec-8acd-4fd7-bb9e-f3ab471d03ec-dns-swift-storage-0\") pod \"dnsmasq-dns-785d8bcb8c-rjqkz\" (UID: \"bcc405ec-8acd-4fd7-bb9e-f3ab471d03ec\") " pod="openstack/dnsmasq-dns-785d8bcb8c-rjqkz" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.649491 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/522901db-3c0d-4cf5-a534-690eac9ae4db-logs\") pod \"glance-default-external-api-0\" (UID: \"522901db-3c0d-4cf5-a534-690eac9ae4db\") " pod="openstack/glance-default-external-api-0" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.649886 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bcc405ec-8acd-4fd7-bb9e-f3ab471d03ec-dns-svc\") pod \"dnsmasq-dns-785d8bcb8c-rjqkz\" (UID: \"bcc405ec-8acd-4fd7-bb9e-f3ab471d03ec\") " pod="openstack/dnsmasq-dns-785d8bcb8c-rjqkz" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.650194 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bcc405ec-8acd-4fd7-bb9e-f3ab471d03ec-ovsdbserver-nb\") pod \"dnsmasq-dns-785d8bcb8c-rjqkz\" (UID: \"bcc405ec-8acd-4fd7-bb9e-f3ab471d03ec\") " pod="openstack/dnsmasq-dns-785d8bcb8c-rjqkz" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.650444 4754 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"522901db-3c0d-4cf5-a534-690eac9ae4db\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/glance-default-external-api-0" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.656070 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-9spps" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.662199 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/522901db-3c0d-4cf5-a534-690eac9ae4db-scripts\") pod \"glance-default-external-api-0\" (UID: \"522901db-3c0d-4cf5-a534-690eac9ae4db\") " pod="openstack/glance-default-external-api-0" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.663612 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/522901db-3c0d-4cf5-a534-690eac9ae4db-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"522901db-3c0d-4cf5-a534-690eac9ae4db\") " pod="openstack/glance-default-external-api-0" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.677205 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/522901db-3c0d-4cf5-a534-690eac9ae4db-config-data\") pod \"glance-default-external-api-0\" (UID: \"522901db-3c0d-4cf5-a534-690eac9ae4db\") " pod="openstack/glance-default-external-api-0" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.678180 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/522901db-3c0d-4cf5-a534-690eac9ae4db-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"522901db-3c0d-4cf5-a534-690eac9ae4db\") " pod="openstack/glance-default-external-api-0" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.685624 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nd7x6\" (UniqueName: \"kubernetes.io/projected/bcc405ec-8acd-4fd7-bb9e-f3ab471d03ec-kube-api-access-nd7x6\") pod \"dnsmasq-dns-785d8bcb8c-rjqkz\" (UID: \"bcc405ec-8acd-4fd7-bb9e-f3ab471d03ec\") " pod="openstack/dnsmasq-dns-785d8bcb8c-rjqkz" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.685883 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b646p\" (UniqueName: \"kubernetes.io/projected/522901db-3c0d-4cf5-a534-690eac9ae4db-kube-api-access-b646p\") pod \"glance-default-external-api-0\" (UID: \"522901db-3c0d-4cf5-a534-690eac9ae4db\") " pod="openstack/glance-default-external-api-0" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.697638 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"522901db-3c0d-4cf5-a534-690eac9ae4db\") " pod="openstack/glance-default-external-api-0" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.722507 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-rjqkz" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.750024 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-66r5j\" (UniqueName: \"kubernetes.io/projected/15b16489-4c64-4a41-bee1-c590e38a87e5-kube-api-access-66r5j\") pod \"glance-default-internal-api-0\" (UID: \"15b16489-4c64-4a41-bee1-c590e38a87e5\") " pod="openstack/glance-default-internal-api-0" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.750089 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/15b16489-4c64-4a41-bee1-c590e38a87e5-logs\") pod \"glance-default-internal-api-0\" (UID: \"15b16489-4c64-4a41-bee1-c590e38a87e5\") " pod="openstack/glance-default-internal-api-0" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.750123 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/15b16489-4c64-4a41-bee1-c590e38a87e5-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"15b16489-4c64-4a41-bee1-c590e38a87e5\") " pod="openstack/glance-default-internal-api-0" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.750147 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/15b16489-4c64-4a41-bee1-c590e38a87e5-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"15b16489-4c64-4a41-bee1-c590e38a87e5\") " pod="openstack/glance-default-internal-api-0" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.750186 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/15b16489-4c64-4a41-bee1-c590e38a87e5-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"15b16489-4c64-4a41-bee1-c590e38a87e5\") " pod="openstack/glance-default-internal-api-0" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.750261 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/15b16489-4c64-4a41-bee1-c590e38a87e5-config-data\") pod \"glance-default-internal-api-0\" (UID: \"15b16489-4c64-4a41-bee1-c590e38a87e5\") " pod="openstack/glance-default-internal-api-0" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.750321 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"15b16489-4c64-4a41-bee1-c590e38a87e5\") " pod="openstack/glance-default-internal-api-0" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.750340 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/15b16489-4c64-4a41-bee1-c590e38a87e5-scripts\") pod \"glance-default-internal-api-0\" (UID: \"15b16489-4c64-4a41-bee1-c590e38a87e5\") " pod="openstack/glance-default-internal-api-0" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.751489 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/15b16489-4c64-4a41-bee1-c590e38a87e5-logs\") pod \"glance-default-internal-api-0\" (UID: \"15b16489-4c64-4a41-bee1-c590e38a87e5\") " pod="openstack/glance-default-internal-api-0" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.754157 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/15b16489-4c64-4a41-bee1-c590e38a87e5-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"15b16489-4c64-4a41-bee1-c590e38a87e5\") " pod="openstack/glance-default-internal-api-0" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.755953 4754 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"15b16489-4c64-4a41-bee1-c590e38a87e5\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/glance-default-internal-api-0" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.769698 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-6qghh"] Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.786791 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/15b16489-4c64-4a41-bee1-c590e38a87e5-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"15b16489-4c64-4a41-bee1-c590e38a87e5\") " pod="openstack/glance-default-internal-api-0" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.800854 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/15b16489-4c64-4a41-bee1-c590e38a87e5-config-data\") pod \"glance-default-internal-api-0\" (UID: \"15b16489-4c64-4a41-bee1-c590e38a87e5\") " pod="openstack/glance-default-internal-api-0" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.805653 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/15b16489-4c64-4a41-bee1-c590e38a87e5-scripts\") pod \"glance-default-internal-api-0\" (UID: \"15b16489-4c64-4a41-bee1-c590e38a87e5\") " pod="openstack/glance-default-internal-api-0" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.805657 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-66r5j\" (UniqueName: \"kubernetes.io/projected/15b16489-4c64-4a41-bee1-c590e38a87e5-kube-api-access-66r5j\") pod \"glance-default-internal-api-0\" (UID: \"15b16489-4c64-4a41-bee1-c590e38a87e5\") " pod="openstack/glance-default-internal-api-0" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.806149 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/15b16489-4c64-4a41-bee1-c590e38a87e5-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"15b16489-4c64-4a41-bee1-c590e38a87e5\") " pod="openstack/glance-default-internal-api-0" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.808209 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"15b16489-4c64-4a41-bee1-c590e38a87e5\") " pod="openstack/glance-default-internal-api-0" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.844627 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Jan 26 17:07:28 crc kubenswrapper[4754]: I0126 17:07:28.953273 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Jan 26 17:07:29 crc kubenswrapper[4754]: I0126 17:07:29.050945 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74f6bcbc87-q8pwj" Jan 26 17:07:29 crc kubenswrapper[4754]: I0126 17:07:29.165954 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a6873959-d867-4a38-a10f-82ff80d0800b-ovsdbserver-sb\") pod \"a6873959-d867-4a38-a10f-82ff80d0800b\" (UID: \"a6873959-d867-4a38-a10f-82ff80d0800b\") " Jan 26 17:07:29 crc kubenswrapper[4754]: I0126 17:07:29.166003 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a6873959-d867-4a38-a10f-82ff80d0800b-ovsdbserver-nb\") pod \"a6873959-d867-4a38-a10f-82ff80d0800b\" (UID: \"a6873959-d867-4a38-a10f-82ff80d0800b\") " Jan 26 17:07:29 crc kubenswrapper[4754]: I0126 17:07:29.166040 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a6873959-d867-4a38-a10f-82ff80d0800b-config\") pod \"a6873959-d867-4a38-a10f-82ff80d0800b\" (UID: \"a6873959-d867-4a38-a10f-82ff80d0800b\") " Jan 26 17:07:29 crc kubenswrapper[4754]: I0126 17:07:29.172718 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a6873959-d867-4a38-a10f-82ff80d0800b-kube-api-access-thfkw" (OuterVolumeSpecName: "kube-api-access-thfkw") pod "a6873959-d867-4a38-a10f-82ff80d0800b" (UID: "a6873959-d867-4a38-a10f-82ff80d0800b"). InnerVolumeSpecName "kube-api-access-thfkw". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:07:29 crc kubenswrapper[4754]: I0126 17:07:29.176230 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-thfkw\" (UniqueName: \"kubernetes.io/projected/a6873959-d867-4a38-a10f-82ff80d0800b-kube-api-access-thfkw\") pod \"a6873959-d867-4a38-a10f-82ff80d0800b\" (UID: \"a6873959-d867-4a38-a10f-82ff80d0800b\") " Jan 26 17:07:29 crc kubenswrapper[4754]: I0126 17:07:29.176376 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a6873959-d867-4a38-a10f-82ff80d0800b-dns-swift-storage-0\") pod \"a6873959-d867-4a38-a10f-82ff80d0800b\" (UID: \"a6873959-d867-4a38-a10f-82ff80d0800b\") " Jan 26 17:07:29 crc kubenswrapper[4754]: I0126 17:07:29.176882 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a6873959-d867-4a38-a10f-82ff80d0800b-dns-svc\") pod \"a6873959-d867-4a38-a10f-82ff80d0800b\" (UID: \"a6873959-d867-4a38-a10f-82ff80d0800b\") " Jan 26 17:07:29 crc kubenswrapper[4754]: I0126 17:07:29.177835 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-thfkw\" (UniqueName: \"kubernetes.io/projected/a6873959-d867-4a38-a10f-82ff80d0800b-kube-api-access-thfkw\") on node \"crc\" DevicePath \"\"" Jan 26 17:07:29 crc kubenswrapper[4754]: I0126 17:07:29.183650 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74f6bcbc87-q8pwj" event={"ID":"a6873959-d867-4a38-a10f-82ff80d0800b","Type":"ContainerDied","Data":"4450de4c2f1d0f92b77707453ea2d45f78ade41dff7f88ba3ef10ce28e271442"} Jan 26 17:07:29 crc kubenswrapper[4754]: I0126 17:07:29.183704 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74f6bcbc87-q8pwj" Jan 26 17:07:29 crc kubenswrapper[4754]: I0126 17:07:29.183729 4754 scope.go:117] "RemoveContainer" containerID="7c2a92f25263e2691e40742a8be4dc87697b019ee912e3a15e8919c7138e41da" Jan 26 17:07:29 crc kubenswrapper[4754]: I0126 17:07:29.184899 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-6qghh" event={"ID":"ee429872-adbd-47bb-9ff2-d5d52dba4d12","Type":"ContainerStarted","Data":"1ae7e8779ee6da77f73e73f093d94b2faf10fc02abe6f19fb543bdabfdf5da0a"} Jan 26 17:07:29 crc kubenswrapper[4754]: I0126 17:07:29.217175 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a6873959-d867-4a38-a10f-82ff80d0800b-config" (OuterVolumeSpecName: "config") pod "a6873959-d867-4a38-a10f-82ff80d0800b" (UID: "a6873959-d867-4a38-a10f-82ff80d0800b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:07:29 crc kubenswrapper[4754]: I0126 17:07:29.220438 4754 scope.go:117] "RemoveContainer" containerID="1375bbc61e6703dac3f4c1f6ad9ded2c448a889672df69d8943c28ce8f1607e4" Jan 26 17:07:29 crc kubenswrapper[4754]: I0126 17:07:29.257238 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a6873959-d867-4a38-a10f-82ff80d0800b-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "a6873959-d867-4a38-a10f-82ff80d0800b" (UID: "a6873959-d867-4a38-a10f-82ff80d0800b"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:07:29 crc kubenswrapper[4754]: I0126 17:07:29.282105 4754 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a6873959-d867-4a38-a10f-82ff80d0800b-config\") on node \"crc\" DevicePath \"\"" Jan 26 17:07:29 crc kubenswrapper[4754]: I0126 17:07:29.282134 4754 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a6873959-d867-4a38-a10f-82ff80d0800b-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Jan 26 17:07:29 crc kubenswrapper[4754]: I0126 17:07:29.297271 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a6873959-d867-4a38-a10f-82ff80d0800b-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "a6873959-d867-4a38-a10f-82ff80d0800b" (UID: "a6873959-d867-4a38-a10f-82ff80d0800b"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:07:29 crc kubenswrapper[4754]: I0126 17:07:29.298380 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a6873959-d867-4a38-a10f-82ff80d0800b-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "a6873959-d867-4a38-a10f-82ff80d0800b" (UID: "a6873959-d867-4a38-a10f-82ff80d0800b"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:07:29 crc kubenswrapper[4754]: I0126 17:07:29.299881 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a6873959-d867-4a38-a10f-82ff80d0800b-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "a6873959-d867-4a38-a10f-82ff80d0800b" (UID: "a6873959-d867-4a38-a10f-82ff80d0800b"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:07:29 crc kubenswrapper[4754]: I0126 17:07:29.326636 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-7hskr"] Jan 26 17:07:29 crc kubenswrapper[4754]: I0126 17:07:29.333816 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-hfr7t"] Jan 26 17:07:29 crc kubenswrapper[4754]: I0126 17:07:29.372587 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-6db5cbfc7c-nxhnr"] Jan 26 17:07:29 crc kubenswrapper[4754]: I0126 17:07:29.386931 4754 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a6873959-d867-4a38-a10f-82ff80d0800b-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Jan 26 17:07:29 crc kubenswrapper[4754]: I0126 17:07:29.386968 4754 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a6873959-d867-4a38-a10f-82ff80d0800b-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Jan 26 17:07:29 crc kubenswrapper[4754]: I0126 17:07:29.386977 4754 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a6873959-d867-4a38-a10f-82ff80d0800b-dns-svc\") on node \"crc\" DevicePath \"\"" Jan 26 17:07:29 crc kubenswrapper[4754]: I0126 17:07:29.413020 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Jan 26 17:07:29 crc kubenswrapper[4754]: I0126 17:07:29.417948 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-whnhn"] Jan 26 17:07:29 crc kubenswrapper[4754]: I0126 17:07:29.424910 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-dzfm9"] Jan 26 17:07:29 crc kubenswrapper[4754]: I0126 17:07:29.551806 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/root-account-create-update-hznmv"] Jan 26 17:07:29 crc kubenswrapper[4754]: E0126 17:07:29.552285 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a6873959-d867-4a38-a10f-82ff80d0800b" containerName="dnsmasq-dns" Jan 26 17:07:29 crc kubenswrapper[4754]: I0126 17:07:29.552304 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="a6873959-d867-4a38-a10f-82ff80d0800b" containerName="dnsmasq-dns" Jan 26 17:07:29 crc kubenswrapper[4754]: E0126 17:07:29.552328 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a6873959-d867-4a38-a10f-82ff80d0800b" containerName="init" Jan 26 17:07:29 crc kubenswrapper[4754]: I0126 17:07:29.552338 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="a6873959-d867-4a38-a10f-82ff80d0800b" containerName="init" Jan 26 17:07:29 crc kubenswrapper[4754]: I0126 17:07:29.552567 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="a6873959-d867-4a38-a10f-82ff80d0800b" containerName="dnsmasq-dns" Jan 26 17:07:29 crc kubenswrapper[4754]: I0126 17:07:29.553374 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-hznmv" Jan 26 17:07:29 crc kubenswrapper[4754]: I0126 17:07:29.556301 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-mariadb-root-db-secret" Jan 26 17:07:29 crc kubenswrapper[4754]: I0126 17:07:29.559233 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-hznmv"] Jan 26 17:07:29 crc kubenswrapper[4754]: W0126 17:07:29.612107 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podeabead53_af6d_4405_b76b_a20c5a84f469.slice/crio-732ce722519d0100e33e34175dbac50759bc15630f8b052b15e80f9ca50aa5ad WatchSource:0}: Error finding container 732ce722519d0100e33e34175dbac50759bc15630f8b052b15e80f9ca50aa5ad: Status 404 returned error can't find the container with id 732ce722519d0100e33e34175dbac50759bc15630f8b052b15e80f9ca50aa5ad Jan 26 17:07:29 crc kubenswrapper[4754]: I0126 17:07:29.618000 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-9spps"] Jan 26 17:07:29 crc kubenswrapper[4754]: I0126 17:07:29.633913 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-585cbdbc67-qwfhd"] Jan 26 17:07:29 crc kubenswrapper[4754]: I0126 17:07:29.640635 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-rjqkz"] Jan 26 17:07:29 crc kubenswrapper[4754]: I0126 17:07:29.705647 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-45mq4\" (UniqueName: \"kubernetes.io/projected/4d4b0b97-349d-45e4-884b-04c0dae58ccb-kube-api-access-45mq4\") pod \"root-account-create-update-hznmv\" (UID: \"4d4b0b97-349d-45e4-884b-04c0dae58ccb\") " pod="openstack/root-account-create-update-hznmv" Jan 26 17:07:29 crc kubenswrapper[4754]: I0126 17:07:29.710470 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4d4b0b97-349d-45e4-884b-04c0dae58ccb-operator-scripts\") pod \"root-account-create-update-hznmv\" (UID: \"4d4b0b97-349d-45e4-884b-04c0dae58ccb\") " pod="openstack/root-account-create-update-hznmv" Jan 26 17:07:29 crc kubenswrapper[4754]: I0126 17:07:29.812780 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4d4b0b97-349d-45e4-884b-04c0dae58ccb-operator-scripts\") pod \"root-account-create-update-hznmv\" (UID: \"4d4b0b97-349d-45e4-884b-04c0dae58ccb\") " pod="openstack/root-account-create-update-hznmv" Jan 26 17:07:29 crc kubenswrapper[4754]: I0126 17:07:29.812868 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-45mq4\" (UniqueName: \"kubernetes.io/projected/4d4b0b97-349d-45e4-884b-04c0dae58ccb-kube-api-access-45mq4\") pod \"root-account-create-update-hznmv\" (UID: \"4d4b0b97-349d-45e4-884b-04c0dae58ccb\") " pod="openstack/root-account-create-update-hznmv" Jan 26 17:07:29 crc kubenswrapper[4754]: I0126 17:07:29.814088 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4d4b0b97-349d-45e4-884b-04c0dae58ccb-operator-scripts\") pod \"root-account-create-update-hznmv\" (UID: \"4d4b0b97-349d-45e4-884b-04c0dae58ccb\") " pod="openstack/root-account-create-update-hznmv" Jan 26 17:07:29 crc kubenswrapper[4754]: I0126 17:07:29.840916 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Jan 26 17:07:29 crc kubenswrapper[4754]: I0126 17:07:29.890610 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-45mq4\" (UniqueName: \"kubernetes.io/projected/4d4b0b97-349d-45e4-884b-04c0dae58ccb-kube-api-access-45mq4\") pod \"root-account-create-update-hznmv\" (UID: \"4d4b0b97-349d-45e4-884b-04c0dae58ccb\") " pod="openstack/root-account-create-update-hznmv" Jan 26 17:07:29 crc kubenswrapper[4754]: I0126 17:07:29.997938 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-q8pwj"] Jan 26 17:07:30 crc kubenswrapper[4754]: I0126 17:07:30.015949 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-q8pwj"] Jan 26 17:07:30 crc kubenswrapper[4754]: I0126 17:07:30.035115 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-585cbdbc67-qwfhd"] Jan 26 17:07:30 crc kubenswrapper[4754]: I0126 17:07:30.098734 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Jan 26 17:07:30 crc kubenswrapper[4754]: I0126 17:07:30.144747 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-6759b7f84f-bzcgl"] Jan 26 17:07:30 crc kubenswrapper[4754]: I0126 17:07:30.147032 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6759b7f84f-bzcgl" Jan 26 17:07:30 crc kubenswrapper[4754]: I0126 17:07:30.176262 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-hznmv" Jan 26 17:07:30 crc kubenswrapper[4754]: I0126 17:07:30.243352 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-6759b7f84f-bzcgl"] Jan 26 17:07:30 crc kubenswrapper[4754]: I0126 17:07:30.246832 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/31d93e77-fe09-47ec-acd8-eced4e88a416-horizon-secret-key\") pod \"horizon-6759b7f84f-bzcgl\" (UID: \"31d93e77-fe09-47ec-acd8-eced4e88a416\") " pod="openstack/horizon-6759b7f84f-bzcgl" Jan 26 17:07:30 crc kubenswrapper[4754]: I0126 17:07:30.247056 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c7vcl\" (UniqueName: \"kubernetes.io/projected/31d93e77-fe09-47ec-acd8-eced4e88a416-kube-api-access-c7vcl\") pod \"horizon-6759b7f84f-bzcgl\" (UID: \"31d93e77-fe09-47ec-acd8-eced4e88a416\") " pod="openstack/horizon-6759b7f84f-bzcgl" Jan 26 17:07:30 crc kubenswrapper[4754]: I0126 17:07:30.247090 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/31d93e77-fe09-47ec-acd8-eced4e88a416-scripts\") pod \"horizon-6759b7f84f-bzcgl\" (UID: \"31d93e77-fe09-47ec-acd8-eced4e88a416\") " pod="openstack/horizon-6759b7f84f-bzcgl" Jan 26 17:07:30 crc kubenswrapper[4754]: I0126 17:07:30.247125 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/31d93e77-fe09-47ec-acd8-eced4e88a416-logs\") pod \"horizon-6759b7f84f-bzcgl\" (UID: \"31d93e77-fe09-47ec-acd8-eced4e88a416\") " pod="openstack/horizon-6759b7f84f-bzcgl" Jan 26 17:07:30 crc kubenswrapper[4754]: I0126 17:07:30.247162 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/31d93e77-fe09-47ec-acd8-eced4e88a416-config-data\") pod \"horizon-6759b7f84f-bzcgl\" (UID: \"31d93e77-fe09-47ec-acd8-eced4e88a416\") " pod="openstack/horizon-6759b7f84f-bzcgl" Jan 26 17:07:30 crc kubenswrapper[4754]: I0126 17:07:30.271657 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Jan 26 17:07:30 crc kubenswrapper[4754]: I0126 17:07:30.286721 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Jan 26 17:07:30 crc kubenswrapper[4754]: I0126 17:07:30.340098 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d84d5918-51f5-4131-9caf-9380004a0ca5","Type":"ContainerStarted","Data":"d699729cd44d0e51f597fa6591ec44c18203dc73bc45ce3e89d2c8230052c8c9"} Jan 26 17:07:30 crc kubenswrapper[4754]: I0126 17:07:30.343223 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-rjqkz" event={"ID":"bcc405ec-8acd-4fd7-bb9e-f3ab471d03ec","Type":"ContainerStarted","Data":"7dae99ce358a95b74a1485097aa09927ce38b696a146189337145849541693dd"} Jan 26 17:07:30 crc kubenswrapper[4754]: I0126 17:07:30.345637 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-585cbdbc67-qwfhd" event={"ID":"eabead53-af6d-4405-b76b-a20c5a84f469","Type":"ContainerStarted","Data":"732ce722519d0100e33e34175dbac50759bc15630f8b052b15e80f9ca50aa5ad"} Jan 26 17:07:30 crc kubenswrapper[4754]: I0126 17:07:30.347172 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-dzfm9" event={"ID":"7d5a6ee0-79d3-4112-a180-d2211fee09f6","Type":"ContainerStarted","Data":"64712fda37a9c52e0bb5053258072c3db8b94658f03933e4e85b50bda7af4cac"} Jan 26 17:07:30 crc kubenswrapper[4754]: I0126 17:07:30.349304 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/31d93e77-fe09-47ec-acd8-eced4e88a416-horizon-secret-key\") pod \"horizon-6759b7f84f-bzcgl\" (UID: \"31d93e77-fe09-47ec-acd8-eced4e88a416\") " pod="openstack/horizon-6759b7f84f-bzcgl" Jan 26 17:07:30 crc kubenswrapper[4754]: I0126 17:07:30.349463 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c7vcl\" (UniqueName: \"kubernetes.io/projected/31d93e77-fe09-47ec-acd8-eced4e88a416-kube-api-access-c7vcl\") pod \"horizon-6759b7f84f-bzcgl\" (UID: \"31d93e77-fe09-47ec-acd8-eced4e88a416\") " pod="openstack/horizon-6759b7f84f-bzcgl" Jan 26 17:07:30 crc kubenswrapper[4754]: I0126 17:07:30.349491 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/31d93e77-fe09-47ec-acd8-eced4e88a416-scripts\") pod \"horizon-6759b7f84f-bzcgl\" (UID: \"31d93e77-fe09-47ec-acd8-eced4e88a416\") " pod="openstack/horizon-6759b7f84f-bzcgl" Jan 26 17:07:30 crc kubenswrapper[4754]: I0126 17:07:30.349526 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/31d93e77-fe09-47ec-acd8-eced4e88a416-logs\") pod \"horizon-6759b7f84f-bzcgl\" (UID: \"31d93e77-fe09-47ec-acd8-eced4e88a416\") " pod="openstack/horizon-6759b7f84f-bzcgl" Jan 26 17:07:30 crc kubenswrapper[4754]: I0126 17:07:30.349750 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/31d93e77-fe09-47ec-acd8-eced4e88a416-config-data\") pod \"horizon-6759b7f84f-bzcgl\" (UID: \"31d93e77-fe09-47ec-acd8-eced4e88a416\") " pod="openstack/horizon-6759b7f84f-bzcgl" Jan 26 17:07:30 crc kubenswrapper[4754]: I0126 17:07:30.351944 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/31d93e77-fe09-47ec-acd8-eced4e88a416-config-data\") pod \"horizon-6759b7f84f-bzcgl\" (UID: \"31d93e77-fe09-47ec-acd8-eced4e88a416\") " pod="openstack/horizon-6759b7f84f-bzcgl" Jan 26 17:07:30 crc kubenswrapper[4754]: I0126 17:07:30.352718 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/31d93e77-fe09-47ec-acd8-eced4e88a416-scripts\") pod \"horizon-6759b7f84f-bzcgl\" (UID: \"31d93e77-fe09-47ec-acd8-eced4e88a416\") " pod="openstack/horizon-6759b7f84f-bzcgl" Jan 26 17:07:30 crc kubenswrapper[4754]: I0126 17:07:30.352893 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-847c4cc679-hfr7t" event={"ID":"83360adb-6875-47a7-9306-c24c92598e29","Type":"ContainerStarted","Data":"e0c80c760673af62ec647930199968960c5d42e6613ac18abb09aef6fdb31244"} Jan 26 17:07:30 crc kubenswrapper[4754]: I0126 17:07:30.352987 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/31d93e77-fe09-47ec-acd8-eced4e88a416-logs\") pod \"horizon-6759b7f84f-bzcgl\" (UID: \"31d93e77-fe09-47ec-acd8-eced4e88a416\") " pod="openstack/horizon-6759b7f84f-bzcgl" Jan 26 17:07:30 crc kubenswrapper[4754]: I0126 17:07:30.369160 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-whnhn" event={"ID":"01246f69-d6bc-4f70-a50f-4d7c8e4a7620","Type":"ContainerStarted","Data":"f3ffb4c4669fc103b7394b3bffeca790d3f2c6ffc4578b08a6b116b2fb7efe2c"} Jan 26 17:07:30 crc kubenswrapper[4754]: I0126 17:07:30.369270 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/31d93e77-fe09-47ec-acd8-eced4e88a416-horizon-secret-key\") pod \"horizon-6759b7f84f-bzcgl\" (UID: \"31d93e77-fe09-47ec-acd8-eced4e88a416\") " pod="openstack/horizon-6759b7f84f-bzcgl" Jan 26 17:07:30 crc kubenswrapper[4754]: I0126 17:07:30.376387 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-6qghh" event={"ID":"ee429872-adbd-47bb-9ff2-d5d52dba4d12","Type":"ContainerStarted","Data":"e14f46f7caf5b8f1f4c6a0e93b774795116f970e960a9fda41198bd120576a1b"} Jan 26 17:07:30 crc kubenswrapper[4754]: I0126 17:07:30.382195 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-9spps" event={"ID":"ea094395-bb3c-4464-829f-313ab6b9e14c","Type":"ContainerStarted","Data":"ede0b1d99871904079a1c996a8496a9574c987d4380c1fe223865c6fc4de3cad"} Jan 26 17:07:30 crc kubenswrapper[4754]: I0126 17:07:30.384658 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6db5cbfc7c-nxhnr" event={"ID":"743bf5b2-5143-406b-b3d7-fa3a067974fd","Type":"ContainerStarted","Data":"19399a2650468f91411b2156cba66da58aa77b72b18d53d56e937132b8c7dec0"} Jan 26 17:07:30 crc kubenswrapper[4754]: I0126 17:07:30.393891 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-7hskr" event={"ID":"ab3f008d-78ab-4b73-962d-4447f312c9fa","Type":"ContainerStarted","Data":"b5a4699e6575064c5f4124d122fdc57ab2a2e7cc8cfcf396dd86d91929f75182"} Jan 26 17:07:30 crc kubenswrapper[4754]: I0126 17:07:30.394047 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c7vcl\" (UniqueName: \"kubernetes.io/projected/31d93e77-fe09-47ec-acd8-eced4e88a416-kube-api-access-c7vcl\") pod \"horizon-6759b7f84f-bzcgl\" (UID: \"31d93e77-fe09-47ec-acd8-eced4e88a416\") " pod="openstack/horizon-6759b7f84f-bzcgl" Jan 26 17:07:30 crc kubenswrapper[4754]: I0126 17:07:30.428906 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-6qghh" podStartSLOduration=3.42888038 podStartE2EDuration="3.42888038s" podCreationTimestamp="2026-01-26 17:07:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 17:07:30.424819856 +0000 UTC m=+1216.949000290" watchObservedRunningTime="2026-01-26 17:07:30.42888038 +0000 UTC m=+1216.953060824" Jan 26 17:07:30 crc kubenswrapper[4754]: I0126 17:07:30.574188 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6759b7f84f-bzcgl" Jan 26 17:07:30 crc kubenswrapper[4754]: I0126 17:07:30.907446 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Jan 26 17:07:30 crc kubenswrapper[4754]: I0126 17:07:30.968247 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-hznmv"] Jan 26 17:07:31 crc kubenswrapper[4754]: I0126 17:07:31.282941 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-6759b7f84f-bzcgl"] Jan 26 17:07:31 crc kubenswrapper[4754]: W0126 17:07:31.286160 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod31d93e77_fe09_47ec_acd8_eced4e88a416.slice/crio-b703a498112279a88b93040feea96ce26ac9dd34bae0ea6fdce1d71517f63804 WatchSource:0}: Error finding container b703a498112279a88b93040feea96ce26ac9dd34bae0ea6fdce1d71517f63804: Status 404 returned error can't find the container with id b703a498112279a88b93040feea96ce26ac9dd34bae0ea6fdce1d71517f63804 Jan 26 17:07:31 crc kubenswrapper[4754]: I0126 17:07:31.430127 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"15b16489-4c64-4a41-bee1-c590e38a87e5","Type":"ContainerStarted","Data":"75b5f6d92d820237809deaa313d21aa66f37cbc63c7d94443b9283307f2abc37"} Jan 26 17:07:31 crc kubenswrapper[4754]: I0126 17:07:31.439884 4754 generic.go:334] "Generic (PLEG): container finished" podID="bcc405ec-8acd-4fd7-bb9e-f3ab471d03ec" containerID="ec3ffb21d050781a4c93354e0bc93031a4a8adfdc941863e27e609e89907b997" exitCode=0 Jan 26 17:07:31 crc kubenswrapper[4754]: I0126 17:07:31.439977 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-rjqkz" event={"ID":"bcc405ec-8acd-4fd7-bb9e-f3ab471d03ec","Type":"ContainerDied","Data":"ec3ffb21d050781a4c93354e0bc93031a4a8adfdc941863e27e609e89907b997"} Jan 26 17:07:31 crc kubenswrapper[4754]: I0126 17:07:31.443144 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-dzfm9" event={"ID":"7d5a6ee0-79d3-4112-a180-d2211fee09f6","Type":"ContainerStarted","Data":"d07947f249ed866995ee2e62e47cf9e3718926322c0dbaebe8e36d9013ee73f5"} Jan 26 17:07:31 crc kubenswrapper[4754]: I0126 17:07:31.445441 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6759b7f84f-bzcgl" event={"ID":"31d93e77-fe09-47ec-acd8-eced4e88a416","Type":"ContainerStarted","Data":"b703a498112279a88b93040feea96ce26ac9dd34bae0ea6fdce1d71517f63804"} Jan 26 17:07:31 crc kubenswrapper[4754]: I0126 17:07:31.450581 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"522901db-3c0d-4cf5-a534-690eac9ae4db","Type":"ContainerStarted","Data":"fb19d988b4090d048b29989a30957a4b02bcf3846d45513cfda1a8199d505922"} Jan 26 17:07:31 crc kubenswrapper[4754]: I0126 17:07:31.454623 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-hznmv" event={"ID":"4d4b0b97-349d-45e4-884b-04c0dae58ccb","Type":"ContainerStarted","Data":"95b41b32fb093081378f984b4a590341fb635fd1ed3785ec87ad1db78080ff99"} Jan 26 17:07:31 crc kubenswrapper[4754]: I0126 17:07:31.454890 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-hznmv" event={"ID":"4d4b0b97-349d-45e4-884b-04c0dae58ccb","Type":"ContainerStarted","Data":"52b158c9fbbdab2b1acdde298f73151697273bf81578470e4dea24601ee495b8"} Jan 26 17:07:31 crc kubenswrapper[4754]: I0126 17:07:31.466573 4754 generic.go:334] "Generic (PLEG): container finished" podID="83360adb-6875-47a7-9306-c24c92598e29" containerID="d17d3c415bfe1a3c305f54f3a7f6bd1cad5165980dbaaab62b54e43beee8823c" exitCode=0 Jan 26 17:07:31 crc kubenswrapper[4754]: I0126 17:07:31.467383 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-847c4cc679-hfr7t" event={"ID":"83360adb-6875-47a7-9306-c24c92598e29","Type":"ContainerDied","Data":"d17d3c415bfe1a3c305f54f3a7f6bd1cad5165980dbaaab62b54e43beee8823c"} Jan 26 17:07:31 crc kubenswrapper[4754]: I0126 17:07:31.471007 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-dzfm9" podStartSLOduration=4.470990946 podStartE2EDuration="4.470990946s" podCreationTimestamp="2026-01-26 17:07:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 17:07:31.467568689 +0000 UTC m=+1217.991749123" watchObservedRunningTime="2026-01-26 17:07:31.470990946 +0000 UTC m=+1217.995171380" Jan 26 17:07:31 crc kubenswrapper[4754]: I0126 17:07:31.515924 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/root-account-create-update-hznmv" podStartSLOduration=2.51590584 podStartE2EDuration="2.51590584s" podCreationTimestamp="2026-01-26 17:07:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 17:07:31.495963172 +0000 UTC m=+1218.020143606" watchObservedRunningTime="2026-01-26 17:07:31.51590584 +0000 UTC m=+1218.040086274" Jan 26 17:07:31 crc kubenswrapper[4754]: I0126 17:07:31.796000 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a6873959-d867-4a38-a10f-82ff80d0800b" path="/var/lib/kubelet/pods/a6873959-d867-4a38-a10f-82ff80d0800b/volumes" Jan 26 17:07:32 crc kubenswrapper[4754]: I0126 17:07:32.481742 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"522901db-3c0d-4cf5-a534-690eac9ae4db","Type":"ContainerStarted","Data":"a76f7bb714655e6a4d6ac10a79a4815aefa2f3a2f15c5456bdee219b0b3665e7"} Jan 26 17:07:32 crc kubenswrapper[4754]: I0126 17:07:32.484862 4754 generic.go:334] "Generic (PLEG): container finished" podID="4d4b0b97-349d-45e4-884b-04c0dae58ccb" containerID="95b41b32fb093081378f984b4a590341fb635fd1ed3785ec87ad1db78080ff99" exitCode=0 Jan 26 17:07:32 crc kubenswrapper[4754]: I0126 17:07:32.485608 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-hznmv" event={"ID":"4d4b0b97-349d-45e4-884b-04c0dae58ccb","Type":"ContainerDied","Data":"95b41b32fb093081378f984b4a590341fb635fd1ed3785ec87ad1db78080ff99"} Jan 26 17:07:34 crc kubenswrapper[4754]: I0126 17:07:34.152355 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-847c4cc679-hfr7t" Jan 26 17:07:34 crc kubenswrapper[4754]: I0126 17:07:34.209194 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-hznmv" Jan 26 17:07:34 crc kubenswrapper[4754]: I0126 17:07:34.345148 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-45mq4\" (UniqueName: \"kubernetes.io/projected/4d4b0b97-349d-45e4-884b-04c0dae58ccb-kube-api-access-45mq4\") pod \"4d4b0b97-349d-45e4-884b-04c0dae58ccb\" (UID: \"4d4b0b97-349d-45e4-884b-04c0dae58ccb\") " Jan 26 17:07:34 crc kubenswrapper[4754]: I0126 17:07:34.345220 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xfkfj\" (UniqueName: \"kubernetes.io/projected/83360adb-6875-47a7-9306-c24c92598e29-kube-api-access-xfkfj\") pod \"83360adb-6875-47a7-9306-c24c92598e29\" (UID: \"83360adb-6875-47a7-9306-c24c92598e29\") " Jan 26 17:07:34 crc kubenswrapper[4754]: I0126 17:07:34.345283 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/83360adb-6875-47a7-9306-c24c92598e29-dns-svc\") pod \"83360adb-6875-47a7-9306-c24c92598e29\" (UID: \"83360adb-6875-47a7-9306-c24c92598e29\") " Jan 26 17:07:34 crc kubenswrapper[4754]: I0126 17:07:34.345311 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/83360adb-6875-47a7-9306-c24c92598e29-ovsdbserver-nb\") pod \"83360adb-6875-47a7-9306-c24c92598e29\" (UID: \"83360adb-6875-47a7-9306-c24c92598e29\") " Jan 26 17:07:34 crc kubenswrapper[4754]: I0126 17:07:34.345374 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/83360adb-6875-47a7-9306-c24c92598e29-config\") pod \"83360adb-6875-47a7-9306-c24c92598e29\" (UID: \"83360adb-6875-47a7-9306-c24c92598e29\") " Jan 26 17:07:34 crc kubenswrapper[4754]: I0126 17:07:34.345423 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/83360adb-6875-47a7-9306-c24c92598e29-dns-swift-storage-0\") pod \"83360adb-6875-47a7-9306-c24c92598e29\" (UID: \"83360adb-6875-47a7-9306-c24c92598e29\") " Jan 26 17:07:34 crc kubenswrapper[4754]: I0126 17:07:34.345455 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/83360adb-6875-47a7-9306-c24c92598e29-ovsdbserver-sb\") pod \"83360adb-6875-47a7-9306-c24c92598e29\" (UID: \"83360adb-6875-47a7-9306-c24c92598e29\") " Jan 26 17:07:34 crc kubenswrapper[4754]: I0126 17:07:34.345574 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4d4b0b97-349d-45e4-884b-04c0dae58ccb-operator-scripts\") pod \"4d4b0b97-349d-45e4-884b-04c0dae58ccb\" (UID: \"4d4b0b97-349d-45e4-884b-04c0dae58ccb\") " Jan 26 17:07:34 crc kubenswrapper[4754]: I0126 17:07:34.346466 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4d4b0b97-349d-45e4-884b-04c0dae58ccb-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "4d4b0b97-349d-45e4-884b-04c0dae58ccb" (UID: "4d4b0b97-349d-45e4-884b-04c0dae58ccb"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:07:34 crc kubenswrapper[4754]: I0126 17:07:34.353648 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4d4b0b97-349d-45e4-884b-04c0dae58ccb-kube-api-access-45mq4" (OuterVolumeSpecName: "kube-api-access-45mq4") pod "4d4b0b97-349d-45e4-884b-04c0dae58ccb" (UID: "4d4b0b97-349d-45e4-884b-04c0dae58ccb"). InnerVolumeSpecName "kube-api-access-45mq4". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:07:34 crc kubenswrapper[4754]: I0126 17:07:34.355230 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/83360adb-6875-47a7-9306-c24c92598e29-kube-api-access-xfkfj" (OuterVolumeSpecName: "kube-api-access-xfkfj") pod "83360adb-6875-47a7-9306-c24c92598e29" (UID: "83360adb-6875-47a7-9306-c24c92598e29"). InnerVolumeSpecName "kube-api-access-xfkfj". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:07:34 crc kubenswrapper[4754]: I0126 17:07:34.369912 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/83360adb-6875-47a7-9306-c24c92598e29-config" (OuterVolumeSpecName: "config") pod "83360adb-6875-47a7-9306-c24c92598e29" (UID: "83360adb-6875-47a7-9306-c24c92598e29"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:07:34 crc kubenswrapper[4754]: I0126 17:07:34.372069 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/83360adb-6875-47a7-9306-c24c92598e29-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "83360adb-6875-47a7-9306-c24c92598e29" (UID: "83360adb-6875-47a7-9306-c24c92598e29"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:07:34 crc kubenswrapper[4754]: I0126 17:07:34.374845 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/83360adb-6875-47a7-9306-c24c92598e29-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "83360adb-6875-47a7-9306-c24c92598e29" (UID: "83360adb-6875-47a7-9306-c24c92598e29"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:07:34 crc kubenswrapper[4754]: I0126 17:07:34.379516 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/83360adb-6875-47a7-9306-c24c92598e29-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "83360adb-6875-47a7-9306-c24c92598e29" (UID: "83360adb-6875-47a7-9306-c24c92598e29"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:07:34 crc kubenswrapper[4754]: I0126 17:07:34.382600 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/83360adb-6875-47a7-9306-c24c92598e29-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "83360adb-6875-47a7-9306-c24c92598e29" (UID: "83360adb-6875-47a7-9306-c24c92598e29"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:07:34 crc kubenswrapper[4754]: I0126 17:07:34.448241 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-45mq4\" (UniqueName: \"kubernetes.io/projected/4d4b0b97-349d-45e4-884b-04c0dae58ccb-kube-api-access-45mq4\") on node \"crc\" DevicePath \"\"" Jan 26 17:07:34 crc kubenswrapper[4754]: I0126 17:07:34.448283 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xfkfj\" (UniqueName: \"kubernetes.io/projected/83360adb-6875-47a7-9306-c24c92598e29-kube-api-access-xfkfj\") on node \"crc\" DevicePath \"\"" Jan 26 17:07:34 crc kubenswrapper[4754]: I0126 17:07:34.448297 4754 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/83360adb-6875-47a7-9306-c24c92598e29-dns-svc\") on node \"crc\" DevicePath \"\"" Jan 26 17:07:34 crc kubenswrapper[4754]: I0126 17:07:34.448481 4754 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/83360adb-6875-47a7-9306-c24c92598e29-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Jan 26 17:07:34 crc kubenswrapper[4754]: I0126 17:07:34.449169 4754 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/83360adb-6875-47a7-9306-c24c92598e29-config\") on node \"crc\" DevicePath \"\"" Jan 26 17:07:34 crc kubenswrapper[4754]: I0126 17:07:34.449309 4754 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/83360adb-6875-47a7-9306-c24c92598e29-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Jan 26 17:07:34 crc kubenswrapper[4754]: I0126 17:07:34.449324 4754 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/83360adb-6875-47a7-9306-c24c92598e29-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Jan 26 17:07:34 crc kubenswrapper[4754]: I0126 17:07:34.449334 4754 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4d4b0b97-349d-45e4-884b-04c0dae58ccb-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 26 17:07:34 crc kubenswrapper[4754]: I0126 17:07:34.508168 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-hznmv" event={"ID":"4d4b0b97-349d-45e4-884b-04c0dae58ccb","Type":"ContainerDied","Data":"52b158c9fbbdab2b1acdde298f73151697273bf81578470e4dea24601ee495b8"} Jan 26 17:07:34 crc kubenswrapper[4754]: I0126 17:07:34.508223 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="52b158c9fbbdab2b1acdde298f73151697273bf81578470e4dea24601ee495b8" Jan 26 17:07:34 crc kubenswrapper[4754]: I0126 17:07:34.508308 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-hznmv" Jan 26 17:07:34 crc kubenswrapper[4754]: I0126 17:07:34.512347 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-847c4cc679-hfr7t" event={"ID":"83360adb-6875-47a7-9306-c24c92598e29","Type":"ContainerDied","Data":"e0c80c760673af62ec647930199968960c5d42e6613ac18abb09aef6fdb31244"} Jan 26 17:07:34 crc kubenswrapper[4754]: I0126 17:07:34.512397 4754 scope.go:117] "RemoveContainer" containerID="d17d3c415bfe1a3c305f54f3a7f6bd1cad5165980dbaaab62b54e43beee8823c" Jan 26 17:07:34 crc kubenswrapper[4754]: I0126 17:07:34.512501 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-847c4cc679-hfr7t" Jan 26 17:07:34 crc kubenswrapper[4754]: I0126 17:07:34.524214 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"15b16489-4c64-4a41-bee1-c590e38a87e5","Type":"ContainerStarted","Data":"a91594b456e2ccd5046b61f457d11d7b6b1cb86757c28c205d84f6b0c2974d26"} Jan 26 17:07:34 crc kubenswrapper[4754]: I0126 17:07:34.661248 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-hfr7t"] Jan 26 17:07:34 crc kubenswrapper[4754]: I0126 17:07:34.671274 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-hfr7t"] Jan 26 17:07:35 crc kubenswrapper[4754]: I0126 17:07:35.537834 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-rjqkz" event={"ID":"bcc405ec-8acd-4fd7-bb9e-f3ab471d03ec","Type":"ContainerStarted","Data":"a08ed9215a80b3818664f82516ebea1b8ce063efca461fceaef02f735167f849"} Jan 26 17:07:35 crc kubenswrapper[4754]: I0126 17:07:35.538858 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-785d8bcb8c-rjqkz" Jan 26 17:07:35 crc kubenswrapper[4754]: I0126 17:07:35.549173 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"522901db-3c0d-4cf5-a534-690eac9ae4db","Type":"ContainerStarted","Data":"a13d104a7ebe2620a1317cf97b99c8d04a5eb85c6e61a296b59b613e50de535b"} Jan 26 17:07:35 crc kubenswrapper[4754]: I0126 17:07:35.549341 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="522901db-3c0d-4cf5-a534-690eac9ae4db" containerName="glance-log" containerID="cri-o://a76f7bb714655e6a4d6ac10a79a4815aefa2f3a2f15c5456bdee219b0b3665e7" gracePeriod=30 Jan 26 17:07:35 crc kubenswrapper[4754]: I0126 17:07:35.549497 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="522901db-3c0d-4cf5-a534-690eac9ae4db" containerName="glance-httpd" containerID="cri-o://a13d104a7ebe2620a1317cf97b99c8d04a5eb85c6e61a296b59b613e50de535b" gracePeriod=30 Jan 26 17:07:35 crc kubenswrapper[4754]: I0126 17:07:35.555530 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"15b16489-4c64-4a41-bee1-c590e38a87e5","Type":"ContainerStarted","Data":"99a5dd7c00c68c3f7c97d0ebfb18d28de56956cebe9ba02924aa4aa127d5e23a"} Jan 26 17:07:35 crc kubenswrapper[4754]: I0126 17:07:35.556136 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="15b16489-4c64-4a41-bee1-c590e38a87e5" containerName="glance-log" containerID="cri-o://a91594b456e2ccd5046b61f457d11d7b6b1cb86757c28c205d84f6b0c2974d26" gracePeriod=30 Jan 26 17:07:35 crc kubenswrapper[4754]: I0126 17:07:35.556398 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="15b16489-4c64-4a41-bee1-c590e38a87e5" containerName="glance-httpd" containerID="cri-o://99a5dd7c00c68c3f7c97d0ebfb18d28de56956cebe9ba02924aa4aa127d5e23a" gracePeriod=30 Jan 26 17:07:35 crc kubenswrapper[4754]: I0126 17:07:35.556975 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-785d8bcb8c-rjqkz" podStartSLOduration=7.556954673 podStartE2EDuration="7.556954673s" podCreationTimestamp="2026-01-26 17:07:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 17:07:35.555309097 +0000 UTC m=+1222.079489541" watchObservedRunningTime="2026-01-26 17:07:35.556954673 +0000 UTC m=+1222.081135097" Jan 26 17:07:35 crc kubenswrapper[4754]: I0126 17:07:35.584526 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=7.584503473 podStartE2EDuration="7.584503473s" podCreationTimestamp="2026-01-26 17:07:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 17:07:35.577403474 +0000 UTC m=+1222.101583998" watchObservedRunningTime="2026-01-26 17:07:35.584503473 +0000 UTC m=+1222.108683907" Jan 26 17:07:35 crc kubenswrapper[4754]: I0126 17:07:35.616810 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=7.616784454 podStartE2EDuration="7.616784454s" podCreationTimestamp="2026-01-26 17:07:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 17:07:35.598885324 +0000 UTC m=+1222.123065758" watchObservedRunningTime="2026-01-26 17:07:35.616784454 +0000 UTC m=+1222.140964898" Jan 26 17:07:35 crc kubenswrapper[4754]: I0126 17:07:35.777955 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="83360adb-6875-47a7-9306-c24c92598e29" path="/var/lib/kubelet/pods/83360adb-6875-47a7-9306-c24c92598e29/volumes" Jan 26 17:07:36 crc kubenswrapper[4754]: I0126 17:07:36.570115 4754 generic.go:334] "Generic (PLEG): container finished" podID="15b16489-4c64-4a41-bee1-c590e38a87e5" containerID="99a5dd7c00c68c3f7c97d0ebfb18d28de56956cebe9ba02924aa4aa127d5e23a" exitCode=143 Jan 26 17:07:36 crc kubenswrapper[4754]: I0126 17:07:36.570146 4754 generic.go:334] "Generic (PLEG): container finished" podID="15b16489-4c64-4a41-bee1-c590e38a87e5" containerID="a91594b456e2ccd5046b61f457d11d7b6b1cb86757c28c205d84f6b0c2974d26" exitCode=143 Jan 26 17:07:36 crc kubenswrapper[4754]: I0126 17:07:36.570185 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"15b16489-4c64-4a41-bee1-c590e38a87e5","Type":"ContainerDied","Data":"99a5dd7c00c68c3f7c97d0ebfb18d28de56956cebe9ba02924aa4aa127d5e23a"} Jan 26 17:07:36 crc kubenswrapper[4754]: I0126 17:07:36.570208 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"15b16489-4c64-4a41-bee1-c590e38a87e5","Type":"ContainerDied","Data":"a91594b456e2ccd5046b61f457d11d7b6b1cb86757c28c205d84f6b0c2974d26"} Jan 26 17:07:36 crc kubenswrapper[4754]: I0126 17:07:36.574442 4754 generic.go:334] "Generic (PLEG): container finished" podID="522901db-3c0d-4cf5-a534-690eac9ae4db" containerID="a13d104a7ebe2620a1317cf97b99c8d04a5eb85c6e61a296b59b613e50de535b" exitCode=0 Jan 26 17:07:36 crc kubenswrapper[4754]: I0126 17:07:36.574727 4754 generic.go:334] "Generic (PLEG): container finished" podID="522901db-3c0d-4cf5-a534-690eac9ae4db" containerID="a76f7bb714655e6a4d6ac10a79a4815aefa2f3a2f15c5456bdee219b0b3665e7" exitCode=143 Jan 26 17:07:36 crc kubenswrapper[4754]: I0126 17:07:36.574528 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"522901db-3c0d-4cf5-a534-690eac9ae4db","Type":"ContainerDied","Data":"a13d104a7ebe2620a1317cf97b99c8d04a5eb85c6e61a296b59b613e50de535b"} Jan 26 17:07:36 crc kubenswrapper[4754]: I0126 17:07:36.574839 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"522901db-3c0d-4cf5-a534-690eac9ae4db","Type":"ContainerDied","Data":"a76f7bb714655e6a4d6ac10a79a4815aefa2f3a2f15c5456bdee219b0b3665e7"} Jan 26 17:07:36 crc kubenswrapper[4754]: I0126 17:07:36.694958 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-6db5cbfc7c-nxhnr"] Jan 26 17:07:36 crc kubenswrapper[4754]: I0126 17:07:36.735105 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-85f95cb57b-99z67"] Jan 26 17:07:36 crc kubenswrapper[4754]: E0126 17:07:36.735521 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83360adb-6875-47a7-9306-c24c92598e29" containerName="init" Jan 26 17:07:36 crc kubenswrapper[4754]: I0126 17:07:36.735542 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="83360adb-6875-47a7-9306-c24c92598e29" containerName="init" Jan 26 17:07:36 crc kubenswrapper[4754]: E0126 17:07:36.735558 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d4b0b97-349d-45e4-884b-04c0dae58ccb" containerName="mariadb-account-create-update" Jan 26 17:07:36 crc kubenswrapper[4754]: I0126 17:07:36.735566 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d4b0b97-349d-45e4-884b-04c0dae58ccb" containerName="mariadb-account-create-update" Jan 26 17:07:36 crc kubenswrapper[4754]: I0126 17:07:36.735759 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="83360adb-6875-47a7-9306-c24c92598e29" containerName="init" Jan 26 17:07:36 crc kubenswrapper[4754]: I0126 17:07:36.735779 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="4d4b0b97-349d-45e4-884b-04c0dae58ccb" containerName="mariadb-account-create-update" Jan 26 17:07:36 crc kubenswrapper[4754]: I0126 17:07:36.736866 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-85f95cb57b-99z67" Jan 26 17:07:36 crc kubenswrapper[4754]: I0126 17:07:36.740620 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-horizon-svc" Jan 26 17:07:36 crc kubenswrapper[4754]: I0126 17:07:36.767643 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-85f95cb57b-99z67"] Jan 26 17:07:36 crc kubenswrapper[4754]: I0126 17:07:36.796563 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-6759b7f84f-bzcgl"] Jan 26 17:07:36 crc kubenswrapper[4754]: I0126 17:07:36.892824 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-6dff6969b8-5bklb"] Jan 26 17:07:36 crc kubenswrapper[4754]: I0126 17:07:36.894460 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6dff6969b8-5bklb" Jan 26 17:07:36 crc kubenswrapper[4754]: I0126 17:07:36.904761 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-6dff6969b8-5bklb"] Jan 26 17:07:36 crc kubenswrapper[4754]: I0126 17:07:36.907692 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tl5tm\" (UniqueName: \"kubernetes.io/projected/ea911a5c-f353-4da9-9900-bfcbb234a377-kube-api-access-tl5tm\") pod \"horizon-85f95cb57b-99z67\" (UID: \"ea911a5c-f353-4da9-9900-bfcbb234a377\") " pod="openstack/horizon-85f95cb57b-99z67" Jan 26 17:07:36 crc kubenswrapper[4754]: I0126 17:07:36.907757 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/ea911a5c-f353-4da9-9900-bfcbb234a377-horizon-tls-certs\") pod \"horizon-85f95cb57b-99z67\" (UID: \"ea911a5c-f353-4da9-9900-bfcbb234a377\") " pod="openstack/horizon-85f95cb57b-99z67" Jan 26 17:07:36 crc kubenswrapper[4754]: I0126 17:07:36.907778 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/ea911a5c-f353-4da9-9900-bfcbb234a377-horizon-secret-key\") pod \"horizon-85f95cb57b-99z67\" (UID: \"ea911a5c-f353-4da9-9900-bfcbb234a377\") " pod="openstack/horizon-85f95cb57b-99z67" Jan 26 17:07:36 crc kubenswrapper[4754]: I0126 17:07:36.907838 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ea911a5c-f353-4da9-9900-bfcbb234a377-scripts\") pod \"horizon-85f95cb57b-99z67\" (UID: \"ea911a5c-f353-4da9-9900-bfcbb234a377\") " pod="openstack/horizon-85f95cb57b-99z67" Jan 26 17:07:36 crc kubenswrapper[4754]: I0126 17:07:36.907864 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ea911a5c-f353-4da9-9900-bfcbb234a377-combined-ca-bundle\") pod \"horizon-85f95cb57b-99z67\" (UID: \"ea911a5c-f353-4da9-9900-bfcbb234a377\") " pod="openstack/horizon-85f95cb57b-99z67" Jan 26 17:07:36 crc kubenswrapper[4754]: I0126 17:07:36.907907 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ea911a5c-f353-4da9-9900-bfcbb234a377-config-data\") pod \"horizon-85f95cb57b-99z67\" (UID: \"ea911a5c-f353-4da9-9900-bfcbb234a377\") " pod="openstack/horizon-85f95cb57b-99z67" Jan 26 17:07:36 crc kubenswrapper[4754]: I0126 17:07:36.908477 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ea911a5c-f353-4da9-9900-bfcbb234a377-logs\") pod \"horizon-85f95cb57b-99z67\" (UID: \"ea911a5c-f353-4da9-9900-bfcbb234a377\") " pod="openstack/horizon-85f95cb57b-99z67" Jan 26 17:07:37 crc kubenswrapper[4754]: I0126 17:07:37.010529 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0313cdfe-e309-41aa-a5ab-83d7713628f3-config-data\") pod \"horizon-6dff6969b8-5bklb\" (UID: \"0313cdfe-e309-41aa-a5ab-83d7713628f3\") " pod="openstack/horizon-6dff6969b8-5bklb" Jan 26 17:07:37 crc kubenswrapper[4754]: I0126 17:07:37.010589 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ea911a5c-f353-4da9-9900-bfcbb234a377-logs\") pod \"horizon-85f95cb57b-99z67\" (UID: \"ea911a5c-f353-4da9-9900-bfcbb234a377\") " pod="openstack/horizon-85f95cb57b-99z67" Jan 26 17:07:37 crc kubenswrapper[4754]: I0126 17:07:37.010616 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/0313cdfe-e309-41aa-a5ab-83d7713628f3-horizon-secret-key\") pod \"horizon-6dff6969b8-5bklb\" (UID: \"0313cdfe-e309-41aa-a5ab-83d7713628f3\") " pod="openstack/horizon-6dff6969b8-5bklb" Jan 26 17:07:37 crc kubenswrapper[4754]: I0126 17:07:37.010633 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0313cdfe-e309-41aa-a5ab-83d7713628f3-scripts\") pod \"horizon-6dff6969b8-5bklb\" (UID: \"0313cdfe-e309-41aa-a5ab-83d7713628f3\") " pod="openstack/horizon-6dff6969b8-5bklb" Jan 26 17:07:37 crc kubenswrapper[4754]: I0126 17:07:37.010650 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/0313cdfe-e309-41aa-a5ab-83d7713628f3-horizon-tls-certs\") pod \"horizon-6dff6969b8-5bklb\" (UID: \"0313cdfe-e309-41aa-a5ab-83d7713628f3\") " pod="openstack/horizon-6dff6969b8-5bklb" Jan 26 17:07:37 crc kubenswrapper[4754]: I0126 17:07:37.010714 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0313cdfe-e309-41aa-a5ab-83d7713628f3-combined-ca-bundle\") pod \"horizon-6dff6969b8-5bklb\" (UID: \"0313cdfe-e309-41aa-a5ab-83d7713628f3\") " pod="openstack/horizon-6dff6969b8-5bklb" Jan 26 17:07:37 crc kubenswrapper[4754]: I0126 17:07:37.010748 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tl5tm\" (UniqueName: \"kubernetes.io/projected/ea911a5c-f353-4da9-9900-bfcbb234a377-kube-api-access-tl5tm\") pod \"horizon-85f95cb57b-99z67\" (UID: \"ea911a5c-f353-4da9-9900-bfcbb234a377\") " pod="openstack/horizon-85f95cb57b-99z67" Jan 26 17:07:37 crc kubenswrapper[4754]: I0126 17:07:37.010779 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0313cdfe-e309-41aa-a5ab-83d7713628f3-logs\") pod \"horizon-6dff6969b8-5bklb\" (UID: \"0313cdfe-e309-41aa-a5ab-83d7713628f3\") " pod="openstack/horizon-6dff6969b8-5bklb" Jan 26 17:07:37 crc kubenswrapper[4754]: I0126 17:07:37.010823 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/ea911a5c-f353-4da9-9900-bfcbb234a377-horizon-tls-certs\") pod \"horizon-85f95cb57b-99z67\" (UID: \"ea911a5c-f353-4da9-9900-bfcbb234a377\") " pod="openstack/horizon-85f95cb57b-99z67" Jan 26 17:07:37 crc kubenswrapper[4754]: I0126 17:07:37.010851 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/ea911a5c-f353-4da9-9900-bfcbb234a377-horizon-secret-key\") pod \"horizon-85f95cb57b-99z67\" (UID: \"ea911a5c-f353-4da9-9900-bfcbb234a377\") " pod="openstack/horizon-85f95cb57b-99z67" Jan 26 17:07:37 crc kubenswrapper[4754]: I0126 17:07:37.010874 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g75rk\" (UniqueName: \"kubernetes.io/projected/0313cdfe-e309-41aa-a5ab-83d7713628f3-kube-api-access-g75rk\") pod \"horizon-6dff6969b8-5bklb\" (UID: \"0313cdfe-e309-41aa-a5ab-83d7713628f3\") " pod="openstack/horizon-6dff6969b8-5bklb" Jan 26 17:07:37 crc kubenswrapper[4754]: I0126 17:07:37.010912 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ea911a5c-f353-4da9-9900-bfcbb234a377-scripts\") pod \"horizon-85f95cb57b-99z67\" (UID: \"ea911a5c-f353-4da9-9900-bfcbb234a377\") " pod="openstack/horizon-85f95cb57b-99z67" Jan 26 17:07:37 crc kubenswrapper[4754]: I0126 17:07:37.010943 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ea911a5c-f353-4da9-9900-bfcbb234a377-combined-ca-bundle\") pod \"horizon-85f95cb57b-99z67\" (UID: \"ea911a5c-f353-4da9-9900-bfcbb234a377\") " pod="openstack/horizon-85f95cb57b-99z67" Jan 26 17:07:37 crc kubenswrapper[4754]: I0126 17:07:37.010982 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ea911a5c-f353-4da9-9900-bfcbb234a377-config-data\") pod \"horizon-85f95cb57b-99z67\" (UID: \"ea911a5c-f353-4da9-9900-bfcbb234a377\") " pod="openstack/horizon-85f95cb57b-99z67" Jan 26 17:07:37 crc kubenswrapper[4754]: I0126 17:07:37.012336 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ea911a5c-f353-4da9-9900-bfcbb234a377-logs\") pod \"horizon-85f95cb57b-99z67\" (UID: \"ea911a5c-f353-4da9-9900-bfcbb234a377\") " pod="openstack/horizon-85f95cb57b-99z67" Jan 26 17:07:37 crc kubenswrapper[4754]: I0126 17:07:37.012878 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ea911a5c-f353-4da9-9900-bfcbb234a377-config-data\") pod \"horizon-85f95cb57b-99z67\" (UID: \"ea911a5c-f353-4da9-9900-bfcbb234a377\") " pod="openstack/horizon-85f95cb57b-99z67" Jan 26 17:07:37 crc kubenswrapper[4754]: I0126 17:07:37.013401 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ea911a5c-f353-4da9-9900-bfcbb234a377-scripts\") pod \"horizon-85f95cb57b-99z67\" (UID: \"ea911a5c-f353-4da9-9900-bfcbb234a377\") " pod="openstack/horizon-85f95cb57b-99z67" Jan 26 17:07:37 crc kubenswrapper[4754]: I0126 17:07:37.022655 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ea911a5c-f353-4da9-9900-bfcbb234a377-combined-ca-bundle\") pod \"horizon-85f95cb57b-99z67\" (UID: \"ea911a5c-f353-4da9-9900-bfcbb234a377\") " pod="openstack/horizon-85f95cb57b-99z67" Jan 26 17:07:37 crc kubenswrapper[4754]: I0126 17:07:37.023213 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/ea911a5c-f353-4da9-9900-bfcbb234a377-horizon-tls-certs\") pod \"horizon-85f95cb57b-99z67\" (UID: \"ea911a5c-f353-4da9-9900-bfcbb234a377\") " pod="openstack/horizon-85f95cb57b-99z67" Jan 26 17:07:37 crc kubenswrapper[4754]: I0126 17:07:37.025453 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/ea911a5c-f353-4da9-9900-bfcbb234a377-horizon-secret-key\") pod \"horizon-85f95cb57b-99z67\" (UID: \"ea911a5c-f353-4da9-9900-bfcbb234a377\") " pod="openstack/horizon-85f95cb57b-99z67" Jan 26 17:07:37 crc kubenswrapper[4754]: I0126 17:07:37.043618 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tl5tm\" (UniqueName: \"kubernetes.io/projected/ea911a5c-f353-4da9-9900-bfcbb234a377-kube-api-access-tl5tm\") pod \"horizon-85f95cb57b-99z67\" (UID: \"ea911a5c-f353-4da9-9900-bfcbb234a377\") " pod="openstack/horizon-85f95cb57b-99z67" Jan 26 17:07:37 crc kubenswrapper[4754]: I0126 17:07:37.077662 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-85f95cb57b-99z67" Jan 26 17:07:37 crc kubenswrapper[4754]: I0126 17:07:37.112270 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/0313cdfe-e309-41aa-a5ab-83d7713628f3-horizon-secret-key\") pod \"horizon-6dff6969b8-5bklb\" (UID: \"0313cdfe-e309-41aa-a5ab-83d7713628f3\") " pod="openstack/horizon-6dff6969b8-5bklb" Jan 26 17:07:37 crc kubenswrapper[4754]: I0126 17:07:37.112320 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0313cdfe-e309-41aa-a5ab-83d7713628f3-scripts\") pod \"horizon-6dff6969b8-5bklb\" (UID: \"0313cdfe-e309-41aa-a5ab-83d7713628f3\") " pod="openstack/horizon-6dff6969b8-5bklb" Jan 26 17:07:37 crc kubenswrapper[4754]: I0126 17:07:37.112347 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/0313cdfe-e309-41aa-a5ab-83d7713628f3-horizon-tls-certs\") pod \"horizon-6dff6969b8-5bklb\" (UID: \"0313cdfe-e309-41aa-a5ab-83d7713628f3\") " pod="openstack/horizon-6dff6969b8-5bklb" Jan 26 17:07:37 crc kubenswrapper[4754]: I0126 17:07:37.112415 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0313cdfe-e309-41aa-a5ab-83d7713628f3-combined-ca-bundle\") pod \"horizon-6dff6969b8-5bklb\" (UID: \"0313cdfe-e309-41aa-a5ab-83d7713628f3\") " pod="openstack/horizon-6dff6969b8-5bklb" Jan 26 17:07:37 crc kubenswrapper[4754]: I0126 17:07:37.112461 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0313cdfe-e309-41aa-a5ab-83d7713628f3-logs\") pod \"horizon-6dff6969b8-5bklb\" (UID: \"0313cdfe-e309-41aa-a5ab-83d7713628f3\") " pod="openstack/horizon-6dff6969b8-5bklb" Jan 26 17:07:37 crc kubenswrapper[4754]: I0126 17:07:37.112513 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g75rk\" (UniqueName: \"kubernetes.io/projected/0313cdfe-e309-41aa-a5ab-83d7713628f3-kube-api-access-g75rk\") pod \"horizon-6dff6969b8-5bklb\" (UID: \"0313cdfe-e309-41aa-a5ab-83d7713628f3\") " pod="openstack/horizon-6dff6969b8-5bklb" Jan 26 17:07:37 crc kubenswrapper[4754]: I0126 17:07:37.112606 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0313cdfe-e309-41aa-a5ab-83d7713628f3-config-data\") pod \"horizon-6dff6969b8-5bklb\" (UID: \"0313cdfe-e309-41aa-a5ab-83d7713628f3\") " pod="openstack/horizon-6dff6969b8-5bklb" Jan 26 17:07:37 crc kubenswrapper[4754]: I0126 17:07:37.113987 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0313cdfe-e309-41aa-a5ab-83d7713628f3-config-data\") pod \"horizon-6dff6969b8-5bklb\" (UID: \"0313cdfe-e309-41aa-a5ab-83d7713628f3\") " pod="openstack/horizon-6dff6969b8-5bklb" Jan 26 17:07:37 crc kubenswrapper[4754]: I0126 17:07:37.115163 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0313cdfe-e309-41aa-a5ab-83d7713628f3-logs\") pod \"horizon-6dff6969b8-5bklb\" (UID: \"0313cdfe-e309-41aa-a5ab-83d7713628f3\") " pod="openstack/horizon-6dff6969b8-5bklb" Jan 26 17:07:37 crc kubenswrapper[4754]: I0126 17:07:37.115864 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0313cdfe-e309-41aa-a5ab-83d7713628f3-scripts\") pod \"horizon-6dff6969b8-5bklb\" (UID: \"0313cdfe-e309-41aa-a5ab-83d7713628f3\") " pod="openstack/horizon-6dff6969b8-5bklb" Jan 26 17:07:37 crc kubenswrapper[4754]: I0126 17:07:37.118438 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0313cdfe-e309-41aa-a5ab-83d7713628f3-combined-ca-bundle\") pod \"horizon-6dff6969b8-5bklb\" (UID: \"0313cdfe-e309-41aa-a5ab-83d7713628f3\") " pod="openstack/horizon-6dff6969b8-5bklb" Jan 26 17:07:37 crc kubenswrapper[4754]: I0126 17:07:37.125317 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/0313cdfe-e309-41aa-a5ab-83d7713628f3-horizon-secret-key\") pod \"horizon-6dff6969b8-5bklb\" (UID: \"0313cdfe-e309-41aa-a5ab-83d7713628f3\") " pod="openstack/horizon-6dff6969b8-5bklb" Jan 26 17:07:37 crc kubenswrapper[4754]: I0126 17:07:37.135424 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/0313cdfe-e309-41aa-a5ab-83d7713628f3-horizon-tls-certs\") pod \"horizon-6dff6969b8-5bklb\" (UID: \"0313cdfe-e309-41aa-a5ab-83d7713628f3\") " pod="openstack/horizon-6dff6969b8-5bklb" Jan 26 17:07:37 crc kubenswrapper[4754]: I0126 17:07:37.135428 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g75rk\" (UniqueName: \"kubernetes.io/projected/0313cdfe-e309-41aa-a5ab-83d7713628f3-kube-api-access-g75rk\") pod \"horizon-6dff6969b8-5bklb\" (UID: \"0313cdfe-e309-41aa-a5ab-83d7713628f3\") " pod="openstack/horizon-6dff6969b8-5bklb" Jan 26 17:07:37 crc kubenswrapper[4754]: I0126 17:07:37.215734 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6dff6969b8-5bklb" Jan 26 17:07:38 crc kubenswrapper[4754]: I0126 17:07:38.607451 4754 generic.go:334] "Generic (PLEG): container finished" podID="ee429872-adbd-47bb-9ff2-d5d52dba4d12" containerID="e14f46f7caf5b8f1f4c6a0e93b774795116f970e960a9fda41198bd120576a1b" exitCode=0 Jan 26 17:07:38 crc kubenswrapper[4754]: I0126 17:07:38.607645 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-6qghh" event={"ID":"ee429872-adbd-47bb-9ff2-d5d52dba4d12","Type":"ContainerDied","Data":"e14f46f7caf5b8f1f4c6a0e93b774795116f970e960a9fda41198bd120576a1b"} Jan 26 17:07:43 crc kubenswrapper[4754]: I0126 17:07:43.723844 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-785d8bcb8c-rjqkz" Jan 26 17:07:43 crc kubenswrapper[4754]: I0126 17:07:43.806405 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-764c5664d7-gmrtk"] Jan 26 17:07:43 crc kubenswrapper[4754]: I0126 17:07:43.806684 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-764c5664d7-gmrtk" podUID="5d5e322d-aa8b-4f36-8257-09d1ccab02c6" containerName="dnsmasq-dns" containerID="cri-o://a9c9a553d41ab84f3ade093e9e2fe01824a18d3f81daba957bbe3e8118796c8e" gracePeriod=10 Jan 26 17:07:44 crc kubenswrapper[4754]: I0126 17:07:44.672180 4754 generic.go:334] "Generic (PLEG): container finished" podID="5d5e322d-aa8b-4f36-8257-09d1ccab02c6" containerID="a9c9a553d41ab84f3ade093e9e2fe01824a18d3f81daba957bbe3e8118796c8e" exitCode=0 Jan 26 17:07:44 crc kubenswrapper[4754]: I0126 17:07:44.672272 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-764c5664d7-gmrtk" event={"ID":"5d5e322d-aa8b-4f36-8257-09d1ccab02c6","Type":"ContainerDied","Data":"a9c9a553d41ab84f3ade093e9e2fe01824a18d3f81daba957bbe3e8118796c8e"} Jan 26 17:07:48 crc kubenswrapper[4754]: I0126 17:07:48.502940 4754 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-764c5664d7-gmrtk" podUID="5d5e322d-aa8b-4f36-8257-09d1ccab02c6" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.130:5353: connect: connection refused" Jan 26 17:07:50 crc kubenswrapper[4754]: I0126 17:07:50.615561 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-6qghh" Jan 26 17:07:50 crc kubenswrapper[4754]: I0126 17:07:50.742244 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-6qghh" event={"ID":"ee429872-adbd-47bb-9ff2-d5d52dba4d12","Type":"ContainerDied","Data":"1ae7e8779ee6da77f73e73f093d94b2faf10fc02abe6f19fb543bdabfdf5da0a"} Jan 26 17:07:50 crc kubenswrapper[4754]: I0126 17:07:50.742293 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1ae7e8779ee6da77f73e73f093d94b2faf10fc02abe6f19fb543bdabfdf5da0a" Jan 26 17:07:50 crc kubenswrapper[4754]: I0126 17:07:50.742305 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-6qghh" Jan 26 17:07:50 crc kubenswrapper[4754]: I0126 17:07:50.786509 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee429872-adbd-47bb-9ff2-d5d52dba4d12-combined-ca-bundle\") pod \"ee429872-adbd-47bb-9ff2-d5d52dba4d12\" (UID: \"ee429872-adbd-47bb-9ff2-d5d52dba4d12\") " Jan 26 17:07:50 crc kubenswrapper[4754]: I0126 17:07:50.786601 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ee429872-adbd-47bb-9ff2-d5d52dba4d12-scripts\") pod \"ee429872-adbd-47bb-9ff2-d5d52dba4d12\" (UID: \"ee429872-adbd-47bb-9ff2-d5d52dba4d12\") " Jan 26 17:07:50 crc kubenswrapper[4754]: I0126 17:07:50.786630 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7lrsm\" (UniqueName: \"kubernetes.io/projected/ee429872-adbd-47bb-9ff2-d5d52dba4d12-kube-api-access-7lrsm\") pod \"ee429872-adbd-47bb-9ff2-d5d52dba4d12\" (UID: \"ee429872-adbd-47bb-9ff2-d5d52dba4d12\") " Jan 26 17:07:50 crc kubenswrapper[4754]: I0126 17:07:50.786727 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/ee429872-adbd-47bb-9ff2-d5d52dba4d12-credential-keys\") pod \"ee429872-adbd-47bb-9ff2-d5d52dba4d12\" (UID: \"ee429872-adbd-47bb-9ff2-d5d52dba4d12\") " Jan 26 17:07:50 crc kubenswrapper[4754]: I0126 17:07:50.786758 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ee429872-adbd-47bb-9ff2-d5d52dba4d12-config-data\") pod \"ee429872-adbd-47bb-9ff2-d5d52dba4d12\" (UID: \"ee429872-adbd-47bb-9ff2-d5d52dba4d12\") " Jan 26 17:07:50 crc kubenswrapper[4754]: I0126 17:07:50.786782 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/ee429872-adbd-47bb-9ff2-d5d52dba4d12-fernet-keys\") pod \"ee429872-adbd-47bb-9ff2-d5d52dba4d12\" (UID: \"ee429872-adbd-47bb-9ff2-d5d52dba4d12\") " Jan 26 17:07:50 crc kubenswrapper[4754]: I0126 17:07:50.801387 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee429872-adbd-47bb-9ff2-d5d52dba4d12-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "ee429872-adbd-47bb-9ff2-d5d52dba4d12" (UID: "ee429872-adbd-47bb-9ff2-d5d52dba4d12"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:07:50 crc kubenswrapper[4754]: I0126 17:07:50.801721 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee429872-adbd-47bb-9ff2-d5d52dba4d12-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "ee429872-adbd-47bb-9ff2-d5d52dba4d12" (UID: "ee429872-adbd-47bb-9ff2-d5d52dba4d12"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:07:50 crc kubenswrapper[4754]: I0126 17:07:50.801797 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee429872-adbd-47bb-9ff2-d5d52dba4d12-scripts" (OuterVolumeSpecName: "scripts") pod "ee429872-adbd-47bb-9ff2-d5d52dba4d12" (UID: "ee429872-adbd-47bb-9ff2-d5d52dba4d12"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:07:50 crc kubenswrapper[4754]: I0126 17:07:50.801856 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ee429872-adbd-47bb-9ff2-d5d52dba4d12-kube-api-access-7lrsm" (OuterVolumeSpecName: "kube-api-access-7lrsm") pod "ee429872-adbd-47bb-9ff2-d5d52dba4d12" (UID: "ee429872-adbd-47bb-9ff2-d5d52dba4d12"). InnerVolumeSpecName "kube-api-access-7lrsm". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:07:50 crc kubenswrapper[4754]: I0126 17:07:50.821891 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee429872-adbd-47bb-9ff2-d5d52dba4d12-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ee429872-adbd-47bb-9ff2-d5d52dba4d12" (UID: "ee429872-adbd-47bb-9ff2-d5d52dba4d12"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:07:50 crc kubenswrapper[4754]: I0126 17:07:50.824616 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee429872-adbd-47bb-9ff2-d5d52dba4d12-config-data" (OuterVolumeSpecName: "config-data") pod "ee429872-adbd-47bb-9ff2-d5d52dba4d12" (UID: "ee429872-adbd-47bb-9ff2-d5d52dba4d12"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:07:50 crc kubenswrapper[4754]: I0126 17:07:50.889343 4754 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/ee429872-adbd-47bb-9ff2-d5d52dba4d12-credential-keys\") on node \"crc\" DevicePath \"\"" Jan 26 17:07:50 crc kubenswrapper[4754]: I0126 17:07:50.889396 4754 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ee429872-adbd-47bb-9ff2-d5d52dba4d12-config-data\") on node \"crc\" DevicePath \"\"" Jan 26 17:07:50 crc kubenswrapper[4754]: I0126 17:07:50.889408 4754 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/ee429872-adbd-47bb-9ff2-d5d52dba4d12-fernet-keys\") on node \"crc\" DevicePath \"\"" Jan 26 17:07:50 crc kubenswrapper[4754]: I0126 17:07:50.889418 4754 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee429872-adbd-47bb-9ff2-d5d52dba4d12-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 26 17:07:50 crc kubenswrapper[4754]: I0126 17:07:50.889429 4754 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ee429872-adbd-47bb-9ff2-d5d52dba4d12-scripts\") on node \"crc\" DevicePath \"\"" Jan 26 17:07:50 crc kubenswrapper[4754]: I0126 17:07:50.889441 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7lrsm\" (UniqueName: \"kubernetes.io/projected/ee429872-adbd-47bb-9ff2-d5d52dba4d12-kube-api-access-7lrsm\") on node \"crc\" DevicePath \"\"" Jan 26 17:07:51 crc kubenswrapper[4754]: E0126 17:07:51.153314 4754 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podee429872_adbd_47bb_9ff2_d5d52dba4d12.slice/crio-1ae7e8779ee6da77f73e73f093d94b2faf10fc02abe6f19fb543bdabfdf5da0a\": RecentStats: unable to find data in memory cache]" Jan 26 17:07:51 crc kubenswrapper[4754]: I0126 17:07:51.702183 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-6qghh"] Jan 26 17:07:51 crc kubenswrapper[4754]: I0126 17:07:51.708708 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-6qghh"] Jan 26 17:07:51 crc kubenswrapper[4754]: I0126 17:07:51.778486 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ee429872-adbd-47bb-9ff2-d5d52dba4d12" path="/var/lib/kubelet/pods/ee429872-adbd-47bb-9ff2-d5d52dba4d12/volumes" Jan 26 17:07:51 crc kubenswrapper[4754]: I0126 17:07:51.801448 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-5xmfz"] Jan 26 17:07:51 crc kubenswrapper[4754]: E0126 17:07:51.801896 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee429872-adbd-47bb-9ff2-d5d52dba4d12" containerName="keystone-bootstrap" Jan 26 17:07:51 crc kubenswrapper[4754]: I0126 17:07:51.801922 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee429872-adbd-47bb-9ff2-d5d52dba4d12" containerName="keystone-bootstrap" Jan 26 17:07:51 crc kubenswrapper[4754]: I0126 17:07:51.802137 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee429872-adbd-47bb-9ff2-d5d52dba4d12" containerName="keystone-bootstrap" Jan 26 17:07:51 crc kubenswrapper[4754]: I0126 17:07:51.802678 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-5xmfz" Jan 26 17:07:51 crc kubenswrapper[4754]: I0126 17:07:51.805070 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Jan 26 17:07:51 crc kubenswrapper[4754]: I0126 17:07:51.805284 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Jan 26 17:07:51 crc kubenswrapper[4754]: I0126 17:07:51.805425 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Jan 26 17:07:51 crc kubenswrapper[4754]: I0126 17:07:51.805761 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Jan 26 17:07:51 crc kubenswrapper[4754]: I0126 17:07:51.810909 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-5xmfz"] Jan 26 17:07:51 crc kubenswrapper[4754]: I0126 17:07:51.811254 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-r42mh" Jan 26 17:07:51 crc kubenswrapper[4754]: I0126 17:07:51.908061 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/e934228b-3d4e-4c15-baa6-5c03e0fceb23-credential-keys\") pod \"keystone-bootstrap-5xmfz\" (UID: \"e934228b-3d4e-4c15-baa6-5c03e0fceb23\") " pod="openstack/keystone-bootstrap-5xmfz" Jan 26 17:07:51 crc kubenswrapper[4754]: I0126 17:07:51.908209 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e934228b-3d4e-4c15-baa6-5c03e0fceb23-combined-ca-bundle\") pod \"keystone-bootstrap-5xmfz\" (UID: \"e934228b-3d4e-4c15-baa6-5c03e0fceb23\") " pod="openstack/keystone-bootstrap-5xmfz" Jan 26 17:07:51 crc kubenswrapper[4754]: I0126 17:07:51.908279 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/e934228b-3d4e-4c15-baa6-5c03e0fceb23-fernet-keys\") pod \"keystone-bootstrap-5xmfz\" (UID: \"e934228b-3d4e-4c15-baa6-5c03e0fceb23\") " pod="openstack/keystone-bootstrap-5xmfz" Jan 26 17:07:51 crc kubenswrapper[4754]: I0126 17:07:51.908373 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e934228b-3d4e-4c15-baa6-5c03e0fceb23-config-data\") pod \"keystone-bootstrap-5xmfz\" (UID: \"e934228b-3d4e-4c15-baa6-5c03e0fceb23\") " pod="openstack/keystone-bootstrap-5xmfz" Jan 26 17:07:51 crc kubenswrapper[4754]: I0126 17:07:51.908401 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e934228b-3d4e-4c15-baa6-5c03e0fceb23-scripts\") pod \"keystone-bootstrap-5xmfz\" (UID: \"e934228b-3d4e-4c15-baa6-5c03e0fceb23\") " pod="openstack/keystone-bootstrap-5xmfz" Jan 26 17:07:51 crc kubenswrapper[4754]: I0126 17:07:51.908442 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fwqwc\" (UniqueName: \"kubernetes.io/projected/e934228b-3d4e-4c15-baa6-5c03e0fceb23-kube-api-access-fwqwc\") pod \"keystone-bootstrap-5xmfz\" (UID: \"e934228b-3d4e-4c15-baa6-5c03e0fceb23\") " pod="openstack/keystone-bootstrap-5xmfz" Jan 26 17:07:52 crc kubenswrapper[4754]: I0126 17:07:52.010469 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e934228b-3d4e-4c15-baa6-5c03e0fceb23-config-data\") pod \"keystone-bootstrap-5xmfz\" (UID: \"e934228b-3d4e-4c15-baa6-5c03e0fceb23\") " pod="openstack/keystone-bootstrap-5xmfz" Jan 26 17:07:52 crc kubenswrapper[4754]: I0126 17:07:52.010523 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e934228b-3d4e-4c15-baa6-5c03e0fceb23-scripts\") pod \"keystone-bootstrap-5xmfz\" (UID: \"e934228b-3d4e-4c15-baa6-5c03e0fceb23\") " pod="openstack/keystone-bootstrap-5xmfz" Jan 26 17:07:52 crc kubenswrapper[4754]: I0126 17:07:52.010558 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fwqwc\" (UniqueName: \"kubernetes.io/projected/e934228b-3d4e-4c15-baa6-5c03e0fceb23-kube-api-access-fwqwc\") pod \"keystone-bootstrap-5xmfz\" (UID: \"e934228b-3d4e-4c15-baa6-5c03e0fceb23\") " pod="openstack/keystone-bootstrap-5xmfz" Jan 26 17:07:52 crc kubenswrapper[4754]: I0126 17:07:52.010607 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/e934228b-3d4e-4c15-baa6-5c03e0fceb23-credential-keys\") pod \"keystone-bootstrap-5xmfz\" (UID: \"e934228b-3d4e-4c15-baa6-5c03e0fceb23\") " pod="openstack/keystone-bootstrap-5xmfz" Jan 26 17:07:52 crc kubenswrapper[4754]: I0126 17:07:52.010646 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e934228b-3d4e-4c15-baa6-5c03e0fceb23-combined-ca-bundle\") pod \"keystone-bootstrap-5xmfz\" (UID: \"e934228b-3d4e-4c15-baa6-5c03e0fceb23\") " pod="openstack/keystone-bootstrap-5xmfz" Jan 26 17:07:52 crc kubenswrapper[4754]: I0126 17:07:52.010678 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/e934228b-3d4e-4c15-baa6-5c03e0fceb23-fernet-keys\") pod \"keystone-bootstrap-5xmfz\" (UID: \"e934228b-3d4e-4c15-baa6-5c03e0fceb23\") " pod="openstack/keystone-bootstrap-5xmfz" Jan 26 17:07:52 crc kubenswrapper[4754]: I0126 17:07:52.017062 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/e934228b-3d4e-4c15-baa6-5c03e0fceb23-credential-keys\") pod \"keystone-bootstrap-5xmfz\" (UID: \"e934228b-3d4e-4c15-baa6-5c03e0fceb23\") " pod="openstack/keystone-bootstrap-5xmfz" Jan 26 17:07:52 crc kubenswrapper[4754]: I0126 17:07:52.017277 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e934228b-3d4e-4c15-baa6-5c03e0fceb23-config-data\") pod \"keystone-bootstrap-5xmfz\" (UID: \"e934228b-3d4e-4c15-baa6-5c03e0fceb23\") " pod="openstack/keystone-bootstrap-5xmfz" Jan 26 17:07:52 crc kubenswrapper[4754]: I0126 17:07:52.017310 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/e934228b-3d4e-4c15-baa6-5c03e0fceb23-fernet-keys\") pod \"keystone-bootstrap-5xmfz\" (UID: \"e934228b-3d4e-4c15-baa6-5c03e0fceb23\") " pod="openstack/keystone-bootstrap-5xmfz" Jan 26 17:07:52 crc kubenswrapper[4754]: I0126 17:07:52.018384 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e934228b-3d4e-4c15-baa6-5c03e0fceb23-combined-ca-bundle\") pod \"keystone-bootstrap-5xmfz\" (UID: \"e934228b-3d4e-4c15-baa6-5c03e0fceb23\") " pod="openstack/keystone-bootstrap-5xmfz" Jan 26 17:07:52 crc kubenswrapper[4754]: I0126 17:07:52.031901 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fwqwc\" (UniqueName: \"kubernetes.io/projected/e934228b-3d4e-4c15-baa6-5c03e0fceb23-kube-api-access-fwqwc\") pod \"keystone-bootstrap-5xmfz\" (UID: \"e934228b-3d4e-4c15-baa6-5c03e0fceb23\") " pod="openstack/keystone-bootstrap-5xmfz" Jan 26 17:07:52 crc kubenswrapper[4754]: I0126 17:07:52.039787 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e934228b-3d4e-4c15-baa6-5c03e0fceb23-scripts\") pod \"keystone-bootstrap-5xmfz\" (UID: \"e934228b-3d4e-4c15-baa6-5c03e0fceb23\") " pod="openstack/keystone-bootstrap-5xmfz" Jan 26 17:07:52 crc kubenswrapper[4754]: I0126 17:07:52.119913 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-5xmfz" Jan 26 17:07:53 crc kubenswrapper[4754]: I0126 17:07:53.503074 4754 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-764c5664d7-gmrtk" podUID="5d5e322d-aa8b-4f36-8257-09d1ccab02c6" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.130:5353: connect: connection refused" Jan 26 17:07:58 crc kubenswrapper[4754]: I0126 17:07:58.502351 4754 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-764c5664d7-gmrtk" podUID="5d5e322d-aa8b-4f36-8257-09d1ccab02c6" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.130:5353: connect: connection refused" Jan 26 17:07:58 crc kubenswrapper[4754]: I0126 17:07:58.502984 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-764c5664d7-gmrtk" Jan 26 17:07:58 crc kubenswrapper[4754]: I0126 17:07:58.846115 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Jan 26 17:07:58 crc kubenswrapper[4754]: I0126 17:07:58.846155 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Jan 26 17:07:58 crc kubenswrapper[4754]: I0126 17:07:58.954566 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Jan 26 17:07:58 crc kubenswrapper[4754]: I0126 17:07:58.954691 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Jan 26 17:07:59 crc kubenswrapper[4754]: E0126 17:07:59.915399 4754 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-placement-api:current-podified" Jan 26 17:07:59 crc kubenswrapper[4754]: E0126 17:07:59.916339 4754 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:placement-db-sync,Image:quay.io/podified-antelope-centos9/openstack-placement-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/placement,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:placement-dbsync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rr5pm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42482,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-db-sync-whnhn_openstack(01246f69-d6bc-4f70-a50f-4d7c8e4a7620): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Jan 26 17:07:59 crc kubenswrapper[4754]: E0126 17:07:59.917998 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"placement-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/placement-db-sync-whnhn" podUID="01246f69-d6bc-4f70-a50f-4d7c8e4a7620" Jan 26 17:07:59 crc kubenswrapper[4754]: E0126 17:07:59.918343 4754 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizon:current-podified" Jan 26 17:07:59 crc kubenswrapper[4754]: E0126 17:07:59.918564 4754 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n567h59dh586h59ch69h556h5b9h9dh566hd9h645h5b4h75h5ffh5b6h576h66hf9hf7h686h684h5b8h59bhdbh545hfbh67fh66bh55fh588h5bh594q,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:no,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-c7vcl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-6759b7f84f-bzcgl_openstack(31d93e77-fe09-47ec-acd8-eced4e88a416): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Jan 26 17:07:59 crc kubenswrapper[4754]: E0126 17:07:59.920807 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon:current-podified\\\"\"]" pod="openstack/horizon-6759b7f84f-bzcgl" podUID="31d93e77-fe09-47ec-acd8-eced4e88a416" Jan 26 17:08:00 crc kubenswrapper[4754]: E0126 17:08:00.822237 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"placement-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-placement-api:current-podified\\\"\"" pod="openstack/placement-db-sync-whnhn" podUID="01246f69-d6bc-4f70-a50f-4d7c8e4a7620" Jan 26 17:08:01 crc kubenswrapper[4754]: E0126 17:08:01.797570 4754 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified" Jan 26 17:08:01 crc kubenswrapper[4754]: E0126 17:08:01.798023 4754 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-5gg2w,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-7hskr_openstack(ab3f008d-78ab-4b73-962d-4447f312c9fa): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Jan 26 17:08:01 crc kubenswrapper[4754]: E0126 17:08:01.799218 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-7hskr" podUID="ab3f008d-78ab-4b73-962d-4447f312c9fa" Jan 26 17:08:01 crc kubenswrapper[4754]: E0126 17:08:01.830735 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified\\\"\"" pod="openstack/cinder-db-sync-7hskr" podUID="ab3f008d-78ab-4b73-962d-4447f312c9fa" Jan 26 17:08:02 crc kubenswrapper[4754]: E0126 17:08:02.325901 4754 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified" Jan 26 17:08:02 crc kubenswrapper[4754]: E0126 17:08:02.326100 4754 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ceilometer-central-agent,Image:quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n66h9bh666h644h587h5ch655h68h676h8fh575h577h66ch67dh5fbh549h5ddh675h96h5f6h89h5h684hc8h55fh676h5c9hd8hd5hch5b7hd5q,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/var/lib/openstack/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:ceilometer-central-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-2bpjj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/python3 /var/lib/openstack/bin/centralhealth.py],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:300,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(d84d5918-51f5-4131-9caf-9380004a0ca5): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Jan 26 17:08:03 crc kubenswrapper[4754]: I0126 17:08:03.135097 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-764c5664d7-gmrtk" Jan 26 17:08:03 crc kubenswrapper[4754]: I0126 17:08:03.141192 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Jan 26 17:08:03 crc kubenswrapper[4754]: I0126 17:08:03.145039 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6759b7f84f-bzcgl" Jan 26 17:08:03 crc kubenswrapper[4754]: I0126 17:08:03.157340 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Jan 26 17:08:03 crc kubenswrapper[4754]: E0126 17:08:03.271908 4754 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified" Jan 26 17:08:03 crc kubenswrapper[4754]: E0126 17:08:03.272071 4754 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:barbican-db-sync,Image:quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified,Command:[/bin/bash],Args:[-c barbican-manage db upgrade],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/barbican/barbican.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-wk6gz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42403,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42403,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-db-sync-9spps_openstack(ea094395-bb3c-4464-829f-313ab6b9e14c): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Jan 26 17:08:03 crc kubenswrapper[4754]: E0126 17:08:03.275148 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/barbican-db-sync-9spps" podUID="ea094395-bb3c-4464-829f-313ab6b9e14c" Jan 26 17:08:03 crc kubenswrapper[4754]: I0126 17:08:03.332442 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/15b16489-4c64-4a41-bee1-c590e38a87e5-logs\") pod \"15b16489-4c64-4a41-bee1-c590e38a87e5\" (UID: \"15b16489-4c64-4a41-bee1-c590e38a87e5\") " Jan 26 17:08:03 crc kubenswrapper[4754]: I0126 17:08:03.332505 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/522901db-3c0d-4cf5-a534-690eac9ae4db-combined-ca-bundle\") pod \"522901db-3c0d-4cf5-a534-690eac9ae4db\" (UID: \"522901db-3c0d-4cf5-a534-690eac9ae4db\") " Jan 26 17:08:03 crc kubenswrapper[4754]: I0126 17:08:03.332570 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b646p\" (UniqueName: \"kubernetes.io/projected/522901db-3c0d-4cf5-a534-690eac9ae4db-kube-api-access-b646p\") pod \"522901db-3c0d-4cf5-a534-690eac9ae4db\" (UID: \"522901db-3c0d-4cf5-a534-690eac9ae4db\") " Jan 26 17:08:03 crc kubenswrapper[4754]: I0126 17:08:03.332607 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/15b16489-4c64-4a41-bee1-c590e38a87e5-config-data\") pod \"15b16489-4c64-4a41-bee1-c590e38a87e5\" (UID: \"15b16489-4c64-4a41-bee1-c590e38a87e5\") " Jan 26 17:08:03 crc kubenswrapper[4754]: I0126 17:08:03.332632 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6pczz\" (UniqueName: \"kubernetes.io/projected/5d5e322d-aa8b-4f36-8257-09d1ccab02c6-kube-api-access-6pczz\") pod \"5d5e322d-aa8b-4f36-8257-09d1ccab02c6\" (UID: \"5d5e322d-aa8b-4f36-8257-09d1ccab02c6\") " Jan 26 17:08:03 crc kubenswrapper[4754]: I0126 17:08:03.332693 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/31d93e77-fe09-47ec-acd8-eced4e88a416-logs\") pod \"31d93e77-fe09-47ec-acd8-eced4e88a416\" (UID: \"31d93e77-fe09-47ec-acd8-eced4e88a416\") " Jan 26 17:08:03 crc kubenswrapper[4754]: I0126 17:08:03.332719 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/31d93e77-fe09-47ec-acd8-eced4e88a416-scripts\") pod \"31d93e77-fe09-47ec-acd8-eced4e88a416\" (UID: \"31d93e77-fe09-47ec-acd8-eced4e88a416\") " Jan 26 17:08:03 crc kubenswrapper[4754]: I0126 17:08:03.332745 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/522901db-3c0d-4cf5-a534-690eac9ae4db-scripts\") pod \"522901db-3c0d-4cf5-a534-690eac9ae4db\" (UID: \"522901db-3c0d-4cf5-a534-690eac9ae4db\") " Jan 26 17:08:03 crc kubenswrapper[4754]: I0126 17:08:03.334114 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c7vcl\" (UniqueName: \"kubernetes.io/projected/31d93e77-fe09-47ec-acd8-eced4e88a416-kube-api-access-c7vcl\") pod \"31d93e77-fe09-47ec-acd8-eced4e88a416\" (UID: \"31d93e77-fe09-47ec-acd8-eced4e88a416\") " Jan 26 17:08:03 crc kubenswrapper[4754]: I0126 17:08:03.334152 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5d5e322d-aa8b-4f36-8257-09d1ccab02c6-dns-svc\") pod \"5d5e322d-aa8b-4f36-8257-09d1ccab02c6\" (UID: \"5d5e322d-aa8b-4f36-8257-09d1ccab02c6\") " Jan 26 17:08:03 crc kubenswrapper[4754]: I0126 17:08:03.334188 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/15b16489-4c64-4a41-bee1-c590e38a87e5-scripts\") pod \"15b16489-4c64-4a41-bee1-c590e38a87e5\" (UID: \"15b16489-4c64-4a41-bee1-c590e38a87e5\") " Jan 26 17:08:03 crc kubenswrapper[4754]: I0126 17:08:03.334219 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/31d93e77-fe09-47ec-acd8-eced4e88a416-config-data\") pod \"31d93e77-fe09-47ec-acd8-eced4e88a416\" (UID: \"31d93e77-fe09-47ec-acd8-eced4e88a416\") " Jan 26 17:08:03 crc kubenswrapper[4754]: I0126 17:08:03.334252 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-66r5j\" (UniqueName: \"kubernetes.io/projected/15b16489-4c64-4a41-bee1-c590e38a87e5-kube-api-access-66r5j\") pod \"15b16489-4c64-4a41-bee1-c590e38a87e5\" (UID: \"15b16489-4c64-4a41-bee1-c590e38a87e5\") " Jan 26 17:08:03 crc kubenswrapper[4754]: I0126 17:08:03.333294 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/15b16489-4c64-4a41-bee1-c590e38a87e5-logs" (OuterVolumeSpecName: "logs") pod "15b16489-4c64-4a41-bee1-c590e38a87e5" (UID: "15b16489-4c64-4a41-bee1-c590e38a87e5"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:08:03 crc kubenswrapper[4754]: I0126 17:08:03.333326 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/31d93e77-fe09-47ec-acd8-eced4e88a416-logs" (OuterVolumeSpecName: "logs") pod "31d93e77-fe09-47ec-acd8-eced4e88a416" (UID: "31d93e77-fe09-47ec-acd8-eced4e88a416"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:08:03 crc kubenswrapper[4754]: I0126 17:08:03.333740 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d93e77-fe09-47ec-acd8-eced4e88a416-scripts" (OuterVolumeSpecName: "scripts") pod "31d93e77-fe09-47ec-acd8-eced4e88a416" (UID: "31d93e77-fe09-47ec-acd8-eced4e88a416"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:08:03 crc kubenswrapper[4754]: I0126 17:08:03.335862 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"15b16489-4c64-4a41-bee1-c590e38a87e5\" (UID: \"15b16489-4c64-4a41-bee1-c590e38a87e5\") " Jan 26 17:08:03 crc kubenswrapper[4754]: I0126 17:08:03.336299 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d93e77-fe09-47ec-acd8-eced4e88a416-config-data" (OuterVolumeSpecName: "config-data") pod "31d93e77-fe09-47ec-acd8-eced4e88a416" (UID: "31d93e77-fe09-47ec-acd8-eced4e88a416"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:08:03 crc kubenswrapper[4754]: I0126 17:08:03.336802 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/15b16489-4c64-4a41-bee1-c590e38a87e5-httpd-run\") pod \"15b16489-4c64-4a41-bee1-c590e38a87e5\" (UID: \"15b16489-4c64-4a41-bee1-c590e38a87e5\") " Jan 26 17:08:03 crc kubenswrapper[4754]: I0126 17:08:03.336831 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/15b16489-4c64-4a41-bee1-c590e38a87e5-combined-ca-bundle\") pod \"15b16489-4c64-4a41-bee1-c590e38a87e5\" (UID: \"15b16489-4c64-4a41-bee1-c590e38a87e5\") " Jan 26 17:08:03 crc kubenswrapper[4754]: I0126 17:08:03.336849 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5d5e322d-aa8b-4f36-8257-09d1ccab02c6-config\") pod \"5d5e322d-aa8b-4f36-8257-09d1ccab02c6\" (UID: \"5d5e322d-aa8b-4f36-8257-09d1ccab02c6\") " Jan 26 17:08:03 crc kubenswrapper[4754]: I0126 17:08:03.336875 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/522901db-3c0d-4cf5-a534-690eac9ae4db-public-tls-certs\") pod \"522901db-3c0d-4cf5-a534-690eac9ae4db\" (UID: \"522901db-3c0d-4cf5-a534-690eac9ae4db\") " Jan 26 17:08:03 crc kubenswrapper[4754]: I0126 17:08:03.336898 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/522901db-3c0d-4cf5-a534-690eac9ae4db-config-data\") pod \"522901db-3c0d-4cf5-a534-690eac9ae4db\" (UID: \"522901db-3c0d-4cf5-a534-690eac9ae4db\") " Jan 26 17:08:03 crc kubenswrapper[4754]: I0126 17:08:03.336919 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5d5e322d-aa8b-4f36-8257-09d1ccab02c6-ovsdbserver-nb\") pod \"5d5e322d-aa8b-4f36-8257-09d1ccab02c6\" (UID: \"5d5e322d-aa8b-4f36-8257-09d1ccab02c6\") " Jan 26 17:08:03 crc kubenswrapper[4754]: I0126 17:08:03.336940 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5d5e322d-aa8b-4f36-8257-09d1ccab02c6-dns-swift-storage-0\") pod \"5d5e322d-aa8b-4f36-8257-09d1ccab02c6\" (UID: \"5d5e322d-aa8b-4f36-8257-09d1ccab02c6\") " Jan 26 17:08:03 crc kubenswrapper[4754]: I0126 17:08:03.336958 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/522901db-3c0d-4cf5-a534-690eac9ae4db-httpd-run\") pod \"522901db-3c0d-4cf5-a534-690eac9ae4db\" (UID: \"522901db-3c0d-4cf5-a534-690eac9ae4db\") " Jan 26 17:08:03 crc kubenswrapper[4754]: I0126 17:08:03.336976 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5d5e322d-aa8b-4f36-8257-09d1ccab02c6-ovsdbserver-sb\") pod \"5d5e322d-aa8b-4f36-8257-09d1ccab02c6\" (UID: \"5d5e322d-aa8b-4f36-8257-09d1ccab02c6\") " Jan 26 17:08:03 crc kubenswrapper[4754]: I0126 17:08:03.336992 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/31d93e77-fe09-47ec-acd8-eced4e88a416-horizon-secret-key\") pod \"31d93e77-fe09-47ec-acd8-eced4e88a416\" (UID: \"31d93e77-fe09-47ec-acd8-eced4e88a416\") " Jan 26 17:08:03 crc kubenswrapper[4754]: I0126 17:08:03.337024 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/15b16489-4c64-4a41-bee1-c590e38a87e5-internal-tls-certs\") pod \"15b16489-4c64-4a41-bee1-c590e38a87e5\" (UID: \"15b16489-4c64-4a41-bee1-c590e38a87e5\") " Jan 26 17:08:03 crc kubenswrapper[4754]: I0126 17:08:03.337053 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"522901db-3c0d-4cf5-a534-690eac9ae4db\" (UID: \"522901db-3c0d-4cf5-a534-690eac9ae4db\") " Jan 26 17:08:03 crc kubenswrapper[4754]: I0126 17:08:03.337087 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/522901db-3c0d-4cf5-a534-690eac9ae4db-logs\") pod \"522901db-3c0d-4cf5-a534-690eac9ae4db\" (UID: \"522901db-3c0d-4cf5-a534-690eac9ae4db\") " Jan 26 17:08:03 crc kubenswrapper[4754]: I0126 17:08:03.337568 4754 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/31d93e77-fe09-47ec-acd8-eced4e88a416-config-data\") on node \"crc\" DevicePath \"\"" Jan 26 17:08:03 crc kubenswrapper[4754]: I0126 17:08:03.337580 4754 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/15b16489-4c64-4a41-bee1-c590e38a87e5-logs\") on node \"crc\" DevicePath \"\"" Jan 26 17:08:03 crc kubenswrapper[4754]: I0126 17:08:03.337589 4754 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/31d93e77-fe09-47ec-acd8-eced4e88a416-logs\") on node \"crc\" DevicePath \"\"" Jan 26 17:08:03 crc kubenswrapper[4754]: I0126 17:08:03.337597 4754 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/31d93e77-fe09-47ec-acd8-eced4e88a416-scripts\") on node \"crc\" DevicePath \"\"" Jan 26 17:08:03 crc kubenswrapper[4754]: I0126 17:08:03.338560 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/522901db-3c0d-4cf5-a534-690eac9ae4db-logs" (OuterVolumeSpecName: "logs") pod "522901db-3c0d-4cf5-a534-690eac9ae4db" (UID: "522901db-3c0d-4cf5-a534-690eac9ae4db"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:08:03 crc kubenswrapper[4754]: I0126 17:08:03.339795 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/15b16489-4c64-4a41-bee1-c590e38a87e5-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "15b16489-4c64-4a41-bee1-c590e38a87e5" (UID: "15b16489-4c64-4a41-bee1-c590e38a87e5"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:08:03 crc kubenswrapper[4754]: I0126 17:08:03.343821 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5d5e322d-aa8b-4f36-8257-09d1ccab02c6-kube-api-access-6pczz" (OuterVolumeSpecName: "kube-api-access-6pczz") pod "5d5e322d-aa8b-4f36-8257-09d1ccab02c6" (UID: "5d5e322d-aa8b-4f36-8257-09d1ccab02c6"). InnerVolumeSpecName "kube-api-access-6pczz". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:08:03 crc kubenswrapper[4754]: I0126 17:08:03.343913 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/522901db-3c0d-4cf5-a534-690eac9ae4db-scripts" (OuterVolumeSpecName: "scripts") pod "522901db-3c0d-4cf5-a534-690eac9ae4db" (UID: "522901db-3c0d-4cf5-a534-690eac9ae4db"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:08:03 crc kubenswrapper[4754]: I0126 17:08:03.344953 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/522901db-3c0d-4cf5-a534-690eac9ae4db-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "522901db-3c0d-4cf5-a534-690eac9ae4db" (UID: "522901db-3c0d-4cf5-a534-690eac9ae4db"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:08:03 crc kubenswrapper[4754]: I0126 17:08:03.350471 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/15b16489-4c64-4a41-bee1-c590e38a87e5-scripts" (OuterVolumeSpecName: "scripts") pod "15b16489-4c64-4a41-bee1-c590e38a87e5" (UID: "15b16489-4c64-4a41-bee1-c590e38a87e5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:08:03 crc kubenswrapper[4754]: I0126 17:08:03.350757 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/522901db-3c0d-4cf5-a534-690eac9ae4db-kube-api-access-b646p" (OuterVolumeSpecName: "kube-api-access-b646p") pod "522901db-3c0d-4cf5-a534-690eac9ae4db" (UID: "522901db-3c0d-4cf5-a534-690eac9ae4db"). InnerVolumeSpecName "kube-api-access-b646p". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:08:03 crc kubenswrapper[4754]: I0126 17:08:03.350639 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d93e77-fe09-47ec-acd8-eced4e88a416-kube-api-access-c7vcl" (OuterVolumeSpecName: "kube-api-access-c7vcl") pod "31d93e77-fe09-47ec-acd8-eced4e88a416" (UID: "31d93e77-fe09-47ec-acd8-eced4e88a416"). InnerVolumeSpecName "kube-api-access-c7vcl". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:08:03 crc kubenswrapper[4754]: I0126 17:08:03.354101 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "glance") pod "15b16489-4c64-4a41-bee1-c590e38a87e5" (UID: "15b16489-4c64-4a41-bee1-c590e38a87e5"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Jan 26 17:08:03 crc kubenswrapper[4754]: I0126 17:08:03.356355 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/15b16489-4c64-4a41-bee1-c590e38a87e5-kube-api-access-66r5j" (OuterVolumeSpecName: "kube-api-access-66r5j") pod "15b16489-4c64-4a41-bee1-c590e38a87e5" (UID: "15b16489-4c64-4a41-bee1-c590e38a87e5"). InnerVolumeSpecName "kube-api-access-66r5j". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:08:03 crc kubenswrapper[4754]: I0126 17:08:03.361317 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d93e77-fe09-47ec-acd8-eced4e88a416-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "31d93e77-fe09-47ec-acd8-eced4e88a416" (UID: "31d93e77-fe09-47ec-acd8-eced4e88a416"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:08:03 crc kubenswrapper[4754]: I0126 17:08:03.363452 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage11-crc" (OuterVolumeSpecName: "glance") pod "522901db-3c0d-4cf5-a534-690eac9ae4db" (UID: "522901db-3c0d-4cf5-a534-690eac9ae4db"). InnerVolumeSpecName "local-storage11-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Jan 26 17:08:03 crc kubenswrapper[4754]: I0126 17:08:03.393284 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/522901db-3c0d-4cf5-a534-690eac9ae4db-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "522901db-3c0d-4cf5-a534-690eac9ae4db" (UID: "522901db-3c0d-4cf5-a534-690eac9ae4db"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:08:03 crc kubenswrapper[4754]: I0126 17:08:03.406625 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/15b16489-4c64-4a41-bee1-c590e38a87e5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "15b16489-4c64-4a41-bee1-c590e38a87e5" (UID: "15b16489-4c64-4a41-bee1-c590e38a87e5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:08:03 crc kubenswrapper[4754]: I0126 17:08:03.434203 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/522901db-3c0d-4cf5-a534-690eac9ae4db-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "522901db-3c0d-4cf5-a534-690eac9ae4db" (UID: "522901db-3c0d-4cf5-a534-690eac9ae4db"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:08:03 crc kubenswrapper[4754]: I0126 17:08:03.438356 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5d5e322d-aa8b-4f36-8257-09d1ccab02c6-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "5d5e322d-aa8b-4f36-8257-09d1ccab02c6" (UID: "5d5e322d-aa8b-4f36-8257-09d1ccab02c6"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:08:03 crc kubenswrapper[4754]: I0126 17:08:03.439655 4754 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5d5e322d-aa8b-4f36-8257-09d1ccab02c6-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Jan 26 17:08:03 crc kubenswrapper[4754]: I0126 17:08:03.439697 4754 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/522901db-3c0d-4cf5-a534-690eac9ae4db-httpd-run\") on node \"crc\" DevicePath \"\"" Jan 26 17:08:03 crc kubenswrapper[4754]: I0126 17:08:03.439711 4754 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/31d93e77-fe09-47ec-acd8-eced4e88a416-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Jan 26 17:08:03 crc kubenswrapper[4754]: I0126 17:08:03.439741 4754 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" " Jan 26 17:08:03 crc kubenswrapper[4754]: I0126 17:08:03.439752 4754 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/522901db-3c0d-4cf5-a534-690eac9ae4db-logs\") on node \"crc\" DevicePath \"\"" Jan 26 17:08:03 crc kubenswrapper[4754]: I0126 17:08:03.439776 4754 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/522901db-3c0d-4cf5-a534-690eac9ae4db-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 26 17:08:03 crc kubenswrapper[4754]: I0126 17:08:03.439785 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b646p\" (UniqueName: \"kubernetes.io/projected/522901db-3c0d-4cf5-a534-690eac9ae4db-kube-api-access-b646p\") on node \"crc\" DevicePath \"\"" Jan 26 17:08:03 crc kubenswrapper[4754]: I0126 17:08:03.439794 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6pczz\" (UniqueName: \"kubernetes.io/projected/5d5e322d-aa8b-4f36-8257-09d1ccab02c6-kube-api-access-6pczz\") on node \"crc\" DevicePath \"\"" Jan 26 17:08:03 crc kubenswrapper[4754]: I0126 17:08:03.439802 4754 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/522901db-3c0d-4cf5-a534-690eac9ae4db-scripts\") on node \"crc\" DevicePath \"\"" Jan 26 17:08:03 crc kubenswrapper[4754]: I0126 17:08:03.439809 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c7vcl\" (UniqueName: \"kubernetes.io/projected/31d93e77-fe09-47ec-acd8-eced4e88a416-kube-api-access-c7vcl\") on node \"crc\" DevicePath \"\"" Jan 26 17:08:03 crc kubenswrapper[4754]: I0126 17:08:03.439817 4754 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/15b16489-4c64-4a41-bee1-c590e38a87e5-scripts\") on node \"crc\" DevicePath \"\"" Jan 26 17:08:03 crc kubenswrapper[4754]: I0126 17:08:03.439826 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-66r5j\" (UniqueName: \"kubernetes.io/projected/15b16489-4c64-4a41-bee1-c590e38a87e5-kube-api-access-66r5j\") on node \"crc\" DevicePath \"\"" Jan 26 17:08:03 crc kubenswrapper[4754]: I0126 17:08:03.439839 4754 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Jan 26 17:08:03 crc kubenswrapper[4754]: I0126 17:08:03.439847 4754 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/15b16489-4c64-4a41-bee1-c590e38a87e5-httpd-run\") on node \"crc\" DevicePath \"\"" Jan 26 17:08:03 crc kubenswrapper[4754]: I0126 17:08:03.439856 4754 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/15b16489-4c64-4a41-bee1-c590e38a87e5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 26 17:08:03 crc kubenswrapper[4754]: I0126 17:08:03.439864 4754 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/522901db-3c0d-4cf5-a534-690eac9ae4db-public-tls-certs\") on node \"crc\" DevicePath \"\"" Jan 26 17:08:03 crc kubenswrapper[4754]: I0126 17:08:03.445443 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/15b16489-4c64-4a41-bee1-c590e38a87e5-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "15b16489-4c64-4a41-bee1-c590e38a87e5" (UID: "15b16489-4c64-4a41-bee1-c590e38a87e5"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:08:03 crc kubenswrapper[4754]: I0126 17:08:03.451953 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5d5e322d-aa8b-4f36-8257-09d1ccab02c6-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "5d5e322d-aa8b-4f36-8257-09d1ccab02c6" (UID: "5d5e322d-aa8b-4f36-8257-09d1ccab02c6"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:08:03 crc kubenswrapper[4754]: I0126 17:08:03.453300 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5d5e322d-aa8b-4f36-8257-09d1ccab02c6-config" (OuterVolumeSpecName: "config") pod "5d5e322d-aa8b-4f36-8257-09d1ccab02c6" (UID: "5d5e322d-aa8b-4f36-8257-09d1ccab02c6"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:08:03 crc kubenswrapper[4754]: I0126 17:08:03.453342 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5d5e322d-aa8b-4f36-8257-09d1ccab02c6-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "5d5e322d-aa8b-4f36-8257-09d1ccab02c6" (UID: "5d5e322d-aa8b-4f36-8257-09d1ccab02c6"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:08:03 crc kubenswrapper[4754]: I0126 17:08:03.463556 4754 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Jan 26 17:08:03 crc kubenswrapper[4754]: I0126 17:08:03.466583 4754 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage11-crc" (UniqueName: "kubernetes.io/local-volume/local-storage11-crc") on node "crc" Jan 26 17:08:03 crc kubenswrapper[4754]: I0126 17:08:03.470444 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5d5e322d-aa8b-4f36-8257-09d1ccab02c6-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "5d5e322d-aa8b-4f36-8257-09d1ccab02c6" (UID: "5d5e322d-aa8b-4f36-8257-09d1ccab02c6"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:08:03 crc kubenswrapper[4754]: I0126 17:08:03.471133 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/15b16489-4c64-4a41-bee1-c590e38a87e5-config-data" (OuterVolumeSpecName: "config-data") pod "15b16489-4c64-4a41-bee1-c590e38a87e5" (UID: "15b16489-4c64-4a41-bee1-c590e38a87e5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:08:03 crc kubenswrapper[4754]: I0126 17:08:03.495882 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/522901db-3c0d-4cf5-a534-690eac9ae4db-config-data" (OuterVolumeSpecName: "config-data") pod "522901db-3c0d-4cf5-a534-690eac9ae4db" (UID: "522901db-3c0d-4cf5-a534-690eac9ae4db"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:08:03 crc kubenswrapper[4754]: I0126 17:08:03.541588 4754 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/522901db-3c0d-4cf5-a534-690eac9ae4db-config-data\") on node \"crc\" DevicePath \"\"" Jan 26 17:08:03 crc kubenswrapper[4754]: I0126 17:08:03.541851 4754 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5d5e322d-aa8b-4f36-8257-09d1ccab02c6-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Jan 26 17:08:03 crc kubenswrapper[4754]: I0126 17:08:03.541861 4754 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5d5e322d-aa8b-4f36-8257-09d1ccab02c6-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Jan 26 17:08:03 crc kubenswrapper[4754]: I0126 17:08:03.541869 4754 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/15b16489-4c64-4a41-bee1-c590e38a87e5-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Jan 26 17:08:03 crc kubenswrapper[4754]: I0126 17:08:03.541879 4754 reconciler_common.go:293] "Volume detached for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" DevicePath \"\"" Jan 26 17:08:03 crc kubenswrapper[4754]: I0126 17:08:03.541887 4754 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/15b16489-4c64-4a41-bee1-c590e38a87e5-config-data\") on node \"crc\" DevicePath \"\"" Jan 26 17:08:03 crc kubenswrapper[4754]: I0126 17:08:03.541898 4754 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5d5e322d-aa8b-4f36-8257-09d1ccab02c6-dns-svc\") on node \"crc\" DevicePath \"\"" Jan 26 17:08:03 crc kubenswrapper[4754]: I0126 17:08:03.541906 4754 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Jan 26 17:08:03 crc kubenswrapper[4754]: I0126 17:08:03.541913 4754 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5d5e322d-aa8b-4f36-8257-09d1ccab02c6-config\") on node \"crc\" DevicePath \"\"" Jan 26 17:08:03 crc kubenswrapper[4754]: I0126 17:08:03.756431 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-85f95cb57b-99z67"] Jan 26 17:08:03 crc kubenswrapper[4754]: W0126 17:08:03.759000 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podea911a5c_f353_4da9_9900_bfcbb234a377.slice/crio-f0b91df38c4a01b5f49a3da68f73e9163c1c9e925d047455d811efa5231a8507 WatchSource:0}: Error finding container f0b91df38c4a01b5f49a3da68f73e9163c1c9e925d047455d811efa5231a8507: Status 404 returned error can't find the container with id f0b91df38c4a01b5f49a3da68f73e9163c1c9e925d047455d811efa5231a8507 Jan 26 17:08:03 crc kubenswrapper[4754]: I0126 17:08:03.845715 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-5xmfz"] Jan 26 17:08:03 crc kubenswrapper[4754]: W0126 17:08:03.851408 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode934228b_3d4e_4c15_baa6_5c03e0fceb23.slice/crio-bb542a6a9a29f8c1e21fb86297b004fcdf04ae3fc10defb5b4300eb4c1211473 WatchSource:0}: Error finding container bb542a6a9a29f8c1e21fb86297b004fcdf04ae3fc10defb5b4300eb4c1211473: Status 404 returned error can't find the container with id bb542a6a9a29f8c1e21fb86297b004fcdf04ae3fc10defb5b4300eb4c1211473 Jan 26 17:08:03 crc kubenswrapper[4754]: W0126 17:08:03.854720 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0313cdfe_e309_41aa_a5ab_83d7713628f3.slice/crio-4b0cc607a3b447ef731c1de736bbd1af5156aa1beb5462a0b64f5ebed2291561 WatchSource:0}: Error finding container 4b0cc607a3b447ef731c1de736bbd1af5156aa1beb5462a0b64f5ebed2291561: Status 404 returned error can't find the container with id 4b0cc607a3b447ef731c1de736bbd1af5156aa1beb5462a0b64f5ebed2291561 Jan 26 17:08:03 crc kubenswrapper[4754]: I0126 17:08:03.859997 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-6dff6969b8-5bklb"] Jan 26 17:08:03 crc kubenswrapper[4754]: I0126 17:08:03.862118 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6759b7f84f-bzcgl" event={"ID":"31d93e77-fe09-47ec-acd8-eced4e88a416","Type":"ContainerDied","Data":"b703a498112279a88b93040feea96ce26ac9dd34bae0ea6fdce1d71517f63804"} Jan 26 17:08:03 crc kubenswrapper[4754]: I0126 17:08:03.862602 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6759b7f84f-bzcgl" Jan 26 17:08:03 crc kubenswrapper[4754]: I0126 17:08:03.866465 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-85f95cb57b-99z67" event={"ID":"ea911a5c-f353-4da9-9900-bfcbb234a377","Type":"ContainerStarted","Data":"f0b91df38c4a01b5f49a3da68f73e9163c1c9e925d047455d811efa5231a8507"} Jan 26 17:08:03 crc kubenswrapper[4754]: I0126 17:08:03.871908 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"522901db-3c0d-4cf5-a534-690eac9ae4db","Type":"ContainerDied","Data":"fb19d988b4090d048b29989a30957a4b02bcf3846d45513cfda1a8199d505922"} Jan 26 17:08:03 crc kubenswrapper[4754]: I0126 17:08:03.871943 4754 scope.go:117] "RemoveContainer" containerID="a13d104a7ebe2620a1317cf97b99c8d04a5eb85c6e61a296b59b613e50de535b" Jan 26 17:08:03 crc kubenswrapper[4754]: I0126 17:08:03.872058 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Jan 26 17:08:03 crc kubenswrapper[4754]: I0126 17:08:03.878797 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-764c5664d7-gmrtk" event={"ID":"5d5e322d-aa8b-4f36-8257-09d1ccab02c6","Type":"ContainerDied","Data":"b8754f9eb16721fb33c204af711238e3f31101e26d6e8971d66e4f6bd4e89b48"} Jan 26 17:08:03 crc kubenswrapper[4754]: I0126 17:08:03.878899 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-764c5664d7-gmrtk" Jan 26 17:08:03 crc kubenswrapper[4754]: I0126 17:08:03.900516 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Jan 26 17:08:03 crc kubenswrapper[4754]: I0126 17:08:03.900510 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"15b16489-4c64-4a41-bee1-c590e38a87e5","Type":"ContainerDied","Data":"75b5f6d92d820237809deaa313d21aa66f37cbc63c7d94443b9283307f2abc37"} Jan 26 17:08:03 crc kubenswrapper[4754]: I0126 17:08:03.910724 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6db5cbfc7c-nxhnr" event={"ID":"743bf5b2-5143-406b-b3d7-fa3a067974fd","Type":"ContainerStarted","Data":"2634e3c695e0a29fa8d18c8ece6dbbd81e5f69b92a0b1c6ad95ae9182c0884c7"} Jan 26 17:08:03 crc kubenswrapper[4754]: I0126 17:08:03.959035 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-6759b7f84f-bzcgl"] Jan 26 17:08:03 crc kubenswrapper[4754]: I0126 17:08:03.964459 4754 scope.go:117] "RemoveContainer" containerID="a76f7bb714655e6a4d6ac10a79a4815aefa2f3a2f15c5456bdee219b0b3665e7" Jan 26 17:08:03 crc kubenswrapper[4754]: E0126 17:08:03.964525 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified\\\"\"" pod="openstack/barbican-db-sync-9spps" podUID="ea094395-bb3c-4464-829f-313ab6b9e14c" Jan 26 17:08:03 crc kubenswrapper[4754]: I0126 17:08:03.990181 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-6759b7f84f-bzcgl"] Jan 26 17:08:04 crc kubenswrapper[4754]: I0126 17:08:04.002509 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Jan 26 17:08:04 crc kubenswrapper[4754]: I0126 17:08:04.014821 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Jan 26 17:08:04 crc kubenswrapper[4754]: I0126 17:08:04.018714 4754 scope.go:117] "RemoveContainer" containerID="a9c9a553d41ab84f3ade093e9e2fe01824a18d3f81daba957bbe3e8118796c8e" Jan 26 17:08:04 crc kubenswrapper[4754]: I0126 17:08:04.025650 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-764c5664d7-gmrtk"] Jan 26 17:08:04 crc kubenswrapper[4754]: I0126 17:08:04.038448 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-764c5664d7-gmrtk"] Jan 26 17:08:04 crc kubenswrapper[4754]: I0126 17:08:04.048903 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Jan 26 17:08:04 crc kubenswrapper[4754]: E0126 17:08:04.049375 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5d5e322d-aa8b-4f36-8257-09d1ccab02c6" containerName="dnsmasq-dns" Jan 26 17:08:04 crc kubenswrapper[4754]: I0126 17:08:04.049401 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="5d5e322d-aa8b-4f36-8257-09d1ccab02c6" containerName="dnsmasq-dns" Jan 26 17:08:04 crc kubenswrapper[4754]: E0126 17:08:04.049420 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15b16489-4c64-4a41-bee1-c590e38a87e5" containerName="glance-log" Jan 26 17:08:04 crc kubenswrapper[4754]: I0126 17:08:04.049430 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="15b16489-4c64-4a41-bee1-c590e38a87e5" containerName="glance-log" Jan 26 17:08:04 crc kubenswrapper[4754]: E0126 17:08:04.049455 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5d5e322d-aa8b-4f36-8257-09d1ccab02c6" containerName="init" Jan 26 17:08:04 crc kubenswrapper[4754]: I0126 17:08:04.049463 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="5d5e322d-aa8b-4f36-8257-09d1ccab02c6" containerName="init" Jan 26 17:08:04 crc kubenswrapper[4754]: E0126 17:08:04.049509 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15b16489-4c64-4a41-bee1-c590e38a87e5" containerName="glance-httpd" Jan 26 17:08:04 crc kubenswrapper[4754]: I0126 17:08:04.049517 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="15b16489-4c64-4a41-bee1-c590e38a87e5" containerName="glance-httpd" Jan 26 17:08:04 crc kubenswrapper[4754]: E0126 17:08:04.049535 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="522901db-3c0d-4cf5-a534-690eac9ae4db" containerName="glance-httpd" Jan 26 17:08:04 crc kubenswrapper[4754]: I0126 17:08:04.049543 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="522901db-3c0d-4cf5-a534-690eac9ae4db" containerName="glance-httpd" Jan 26 17:08:04 crc kubenswrapper[4754]: E0126 17:08:04.049555 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="522901db-3c0d-4cf5-a534-690eac9ae4db" containerName="glance-log" Jan 26 17:08:04 crc kubenswrapper[4754]: I0126 17:08:04.049566 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="522901db-3c0d-4cf5-a534-690eac9ae4db" containerName="glance-log" Jan 26 17:08:04 crc kubenswrapper[4754]: I0126 17:08:04.049802 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="522901db-3c0d-4cf5-a534-690eac9ae4db" containerName="glance-log" Jan 26 17:08:04 crc kubenswrapper[4754]: I0126 17:08:04.049832 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="5d5e322d-aa8b-4f36-8257-09d1ccab02c6" containerName="dnsmasq-dns" Jan 26 17:08:04 crc kubenswrapper[4754]: I0126 17:08:04.049848 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="15b16489-4c64-4a41-bee1-c590e38a87e5" containerName="glance-httpd" Jan 26 17:08:04 crc kubenswrapper[4754]: I0126 17:08:04.049870 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="522901db-3c0d-4cf5-a534-690eac9ae4db" containerName="glance-httpd" Jan 26 17:08:04 crc kubenswrapper[4754]: I0126 17:08:04.049886 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="15b16489-4c64-4a41-bee1-c590e38a87e5" containerName="glance-log" Jan 26 17:08:04 crc kubenswrapper[4754]: I0126 17:08:04.051324 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Jan 26 17:08:04 crc kubenswrapper[4754]: I0126 17:08:04.053533 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Jan 26 17:08:04 crc kubenswrapper[4754]: I0126 17:08:04.053785 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-hph75" Jan 26 17:08:04 crc kubenswrapper[4754]: I0126 17:08:04.054740 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Jan 26 17:08:04 crc kubenswrapper[4754]: I0126 17:08:04.058306 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Jan 26 17:08:04 crc kubenswrapper[4754]: I0126 17:08:04.079832 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Jan 26 17:08:04 crc kubenswrapper[4754]: I0126 17:08:04.087077 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Jan 26 17:08:04 crc kubenswrapper[4754]: I0126 17:08:04.094825 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Jan 26 17:08:04 crc kubenswrapper[4754]: I0126 17:08:04.101175 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Jan 26 17:08:04 crc kubenswrapper[4754]: I0126 17:08:04.103137 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Jan 26 17:08:04 crc kubenswrapper[4754]: I0126 17:08:04.105661 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Jan 26 17:08:04 crc kubenswrapper[4754]: I0126 17:08:04.106458 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Jan 26 17:08:04 crc kubenswrapper[4754]: I0126 17:08:04.108428 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Jan 26 17:08:04 crc kubenswrapper[4754]: I0126 17:08:04.124168 4754 scope.go:117] "RemoveContainer" containerID="702220854f984609ce86913f04217dceba69fcb60a0a87776849f13e43a5d464" Jan 26 17:08:04 crc kubenswrapper[4754]: I0126 17:08:04.172950 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4a8fafa1-3d82-4a16-85e1-fd3116bc998a-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"4a8fafa1-3d82-4a16-85e1-fd3116bc998a\") " pod="openstack/glance-default-external-api-0" Jan 26 17:08:04 crc kubenswrapper[4754]: I0126 17:08:04.173024 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4a8fafa1-3d82-4a16-85e1-fd3116bc998a-scripts\") pod \"glance-default-external-api-0\" (UID: \"4a8fafa1-3d82-4a16-85e1-fd3116bc998a\") " pod="openstack/glance-default-external-api-0" Jan 26 17:08:04 crc kubenswrapper[4754]: I0126 17:08:04.173059 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4a8fafa1-3d82-4a16-85e1-fd3116bc998a-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"4a8fafa1-3d82-4a16-85e1-fd3116bc998a\") " pod="openstack/glance-default-external-api-0" Jan 26 17:08:04 crc kubenswrapper[4754]: I0126 17:08:04.173131 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4a8fafa1-3d82-4a16-85e1-fd3116bc998a-config-data\") pod \"glance-default-external-api-0\" (UID: \"4a8fafa1-3d82-4a16-85e1-fd3116bc998a\") " pod="openstack/glance-default-external-api-0" Jan 26 17:08:04 crc kubenswrapper[4754]: I0126 17:08:04.173283 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4a8fafa1-3d82-4a16-85e1-fd3116bc998a-logs\") pod \"glance-default-external-api-0\" (UID: \"4a8fafa1-3d82-4a16-85e1-fd3116bc998a\") " pod="openstack/glance-default-external-api-0" Jan 26 17:08:04 crc kubenswrapper[4754]: I0126 17:08:04.173379 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"4a8fafa1-3d82-4a16-85e1-fd3116bc998a\") " pod="openstack/glance-default-external-api-0" Jan 26 17:08:04 crc kubenswrapper[4754]: I0126 17:08:04.173460 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g62ms\" (UniqueName: \"kubernetes.io/projected/4a8fafa1-3d82-4a16-85e1-fd3116bc998a-kube-api-access-g62ms\") pod \"glance-default-external-api-0\" (UID: \"4a8fafa1-3d82-4a16-85e1-fd3116bc998a\") " pod="openstack/glance-default-external-api-0" Jan 26 17:08:04 crc kubenswrapper[4754]: I0126 17:08:04.173518 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a8fafa1-3d82-4a16-85e1-fd3116bc998a-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"4a8fafa1-3d82-4a16-85e1-fd3116bc998a\") " pod="openstack/glance-default-external-api-0" Jan 26 17:08:04 crc kubenswrapper[4754]: I0126 17:08:04.186384 4754 scope.go:117] "RemoveContainer" containerID="99a5dd7c00c68c3f7c97d0ebfb18d28de56956cebe9ba02924aa4aa127d5e23a" Jan 26 17:08:04 crc kubenswrapper[4754]: I0126 17:08:04.246157 4754 scope.go:117] "RemoveContainer" containerID="a91594b456e2ccd5046b61f457d11d7b6b1cb86757c28c205d84f6b0c2974d26" Jan 26 17:08:04 crc kubenswrapper[4754]: I0126 17:08:04.276071 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4a8fafa1-3d82-4a16-85e1-fd3116bc998a-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"4a8fafa1-3d82-4a16-85e1-fd3116bc998a\") " pod="openstack/glance-default-external-api-0" Jan 26 17:08:04 crc kubenswrapper[4754]: I0126 17:08:04.276126 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4a8fafa1-3d82-4a16-85e1-fd3116bc998a-scripts\") pod \"glance-default-external-api-0\" (UID: \"4a8fafa1-3d82-4a16-85e1-fd3116bc998a\") " pod="openstack/glance-default-external-api-0" Jan 26 17:08:04 crc kubenswrapper[4754]: I0126 17:08:04.276168 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/428912eb-4aed-4926-8b00-f716a43a4399-config-data\") pod \"glance-default-internal-api-0\" (UID: \"428912eb-4aed-4926-8b00-f716a43a4399\") " pod="openstack/glance-default-internal-api-0" Jan 26 17:08:04 crc kubenswrapper[4754]: I0126 17:08:04.276197 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/428912eb-4aed-4926-8b00-f716a43a4399-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"428912eb-4aed-4926-8b00-f716a43a4399\") " pod="openstack/glance-default-internal-api-0" Jan 26 17:08:04 crc kubenswrapper[4754]: I0126 17:08:04.276227 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4a8fafa1-3d82-4a16-85e1-fd3116bc998a-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"4a8fafa1-3d82-4a16-85e1-fd3116bc998a\") " pod="openstack/glance-default-external-api-0" Jan 26 17:08:04 crc kubenswrapper[4754]: I0126 17:08:04.276260 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/428912eb-4aed-4926-8b00-f716a43a4399-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"428912eb-4aed-4926-8b00-f716a43a4399\") " pod="openstack/glance-default-internal-api-0" Jan 26 17:08:04 crc kubenswrapper[4754]: I0126 17:08:04.276294 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4a8fafa1-3d82-4a16-85e1-fd3116bc998a-config-data\") pod \"glance-default-external-api-0\" (UID: \"4a8fafa1-3d82-4a16-85e1-fd3116bc998a\") " pod="openstack/glance-default-external-api-0" Jan 26 17:08:04 crc kubenswrapper[4754]: I0126 17:08:04.276333 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"428912eb-4aed-4926-8b00-f716a43a4399\") " pod="openstack/glance-default-internal-api-0" Jan 26 17:08:04 crc kubenswrapper[4754]: I0126 17:08:04.276369 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r5q6t\" (UniqueName: \"kubernetes.io/projected/428912eb-4aed-4926-8b00-f716a43a4399-kube-api-access-r5q6t\") pod \"glance-default-internal-api-0\" (UID: \"428912eb-4aed-4926-8b00-f716a43a4399\") " pod="openstack/glance-default-internal-api-0" Jan 26 17:08:04 crc kubenswrapper[4754]: I0126 17:08:04.276416 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/428912eb-4aed-4926-8b00-f716a43a4399-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"428912eb-4aed-4926-8b00-f716a43a4399\") " pod="openstack/glance-default-internal-api-0" Jan 26 17:08:04 crc kubenswrapper[4754]: I0126 17:08:04.276439 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4a8fafa1-3d82-4a16-85e1-fd3116bc998a-logs\") pod \"glance-default-external-api-0\" (UID: \"4a8fafa1-3d82-4a16-85e1-fd3116bc998a\") " pod="openstack/glance-default-external-api-0" Jan 26 17:08:04 crc kubenswrapper[4754]: I0126 17:08:04.276483 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"4a8fafa1-3d82-4a16-85e1-fd3116bc998a\") " pod="openstack/glance-default-external-api-0" Jan 26 17:08:04 crc kubenswrapper[4754]: I0126 17:08:04.276515 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/428912eb-4aed-4926-8b00-f716a43a4399-scripts\") pod \"glance-default-internal-api-0\" (UID: \"428912eb-4aed-4926-8b00-f716a43a4399\") " pod="openstack/glance-default-internal-api-0" Jan 26 17:08:04 crc kubenswrapper[4754]: I0126 17:08:04.276553 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g62ms\" (UniqueName: \"kubernetes.io/projected/4a8fafa1-3d82-4a16-85e1-fd3116bc998a-kube-api-access-g62ms\") pod \"glance-default-external-api-0\" (UID: \"4a8fafa1-3d82-4a16-85e1-fd3116bc998a\") " pod="openstack/glance-default-external-api-0" Jan 26 17:08:04 crc kubenswrapper[4754]: I0126 17:08:04.276578 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a8fafa1-3d82-4a16-85e1-fd3116bc998a-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"4a8fafa1-3d82-4a16-85e1-fd3116bc998a\") " pod="openstack/glance-default-external-api-0" Jan 26 17:08:04 crc kubenswrapper[4754]: I0126 17:08:04.276609 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/428912eb-4aed-4926-8b00-f716a43a4399-logs\") pod \"glance-default-internal-api-0\" (UID: \"428912eb-4aed-4926-8b00-f716a43a4399\") " pod="openstack/glance-default-internal-api-0" Jan 26 17:08:04 crc kubenswrapper[4754]: I0126 17:08:04.278532 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4a8fafa1-3d82-4a16-85e1-fd3116bc998a-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"4a8fafa1-3d82-4a16-85e1-fd3116bc998a\") " pod="openstack/glance-default-external-api-0" Jan 26 17:08:04 crc kubenswrapper[4754]: I0126 17:08:04.278970 4754 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"4a8fafa1-3d82-4a16-85e1-fd3116bc998a\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/glance-default-external-api-0" Jan 26 17:08:04 crc kubenswrapper[4754]: I0126 17:08:04.279031 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4a8fafa1-3d82-4a16-85e1-fd3116bc998a-logs\") pod \"glance-default-external-api-0\" (UID: \"4a8fafa1-3d82-4a16-85e1-fd3116bc998a\") " pod="openstack/glance-default-external-api-0" Jan 26 17:08:04 crc kubenswrapper[4754]: I0126 17:08:04.283988 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4a8fafa1-3d82-4a16-85e1-fd3116bc998a-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"4a8fafa1-3d82-4a16-85e1-fd3116bc998a\") " pod="openstack/glance-default-external-api-0" Jan 26 17:08:04 crc kubenswrapper[4754]: I0126 17:08:04.302508 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4a8fafa1-3d82-4a16-85e1-fd3116bc998a-scripts\") pod \"glance-default-external-api-0\" (UID: \"4a8fafa1-3d82-4a16-85e1-fd3116bc998a\") " pod="openstack/glance-default-external-api-0" Jan 26 17:08:04 crc kubenswrapper[4754]: I0126 17:08:04.303693 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4a8fafa1-3d82-4a16-85e1-fd3116bc998a-config-data\") pod \"glance-default-external-api-0\" (UID: \"4a8fafa1-3d82-4a16-85e1-fd3116bc998a\") " pod="openstack/glance-default-external-api-0" Jan 26 17:08:04 crc kubenswrapper[4754]: I0126 17:08:04.307271 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a8fafa1-3d82-4a16-85e1-fd3116bc998a-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"4a8fafa1-3d82-4a16-85e1-fd3116bc998a\") " pod="openstack/glance-default-external-api-0" Jan 26 17:08:04 crc kubenswrapper[4754]: I0126 17:08:04.317884 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g62ms\" (UniqueName: \"kubernetes.io/projected/4a8fafa1-3d82-4a16-85e1-fd3116bc998a-kube-api-access-g62ms\") pod \"glance-default-external-api-0\" (UID: \"4a8fafa1-3d82-4a16-85e1-fd3116bc998a\") " pod="openstack/glance-default-external-api-0" Jan 26 17:08:04 crc kubenswrapper[4754]: I0126 17:08:04.346682 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"4a8fafa1-3d82-4a16-85e1-fd3116bc998a\") " pod="openstack/glance-default-external-api-0" Jan 26 17:08:04 crc kubenswrapper[4754]: I0126 17:08:04.377792 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/428912eb-4aed-4926-8b00-f716a43a4399-scripts\") pod \"glance-default-internal-api-0\" (UID: \"428912eb-4aed-4926-8b00-f716a43a4399\") " pod="openstack/glance-default-internal-api-0" Jan 26 17:08:04 crc kubenswrapper[4754]: I0126 17:08:04.378017 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/428912eb-4aed-4926-8b00-f716a43a4399-logs\") pod \"glance-default-internal-api-0\" (UID: \"428912eb-4aed-4926-8b00-f716a43a4399\") " pod="openstack/glance-default-internal-api-0" Jan 26 17:08:04 crc kubenswrapper[4754]: I0126 17:08:04.378075 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/428912eb-4aed-4926-8b00-f716a43a4399-config-data\") pod \"glance-default-internal-api-0\" (UID: \"428912eb-4aed-4926-8b00-f716a43a4399\") " pod="openstack/glance-default-internal-api-0" Jan 26 17:08:04 crc kubenswrapper[4754]: I0126 17:08:04.378095 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/428912eb-4aed-4926-8b00-f716a43a4399-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"428912eb-4aed-4926-8b00-f716a43a4399\") " pod="openstack/glance-default-internal-api-0" Jan 26 17:08:04 crc kubenswrapper[4754]: I0126 17:08:04.378125 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/428912eb-4aed-4926-8b00-f716a43a4399-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"428912eb-4aed-4926-8b00-f716a43a4399\") " pod="openstack/glance-default-internal-api-0" Jan 26 17:08:04 crc kubenswrapper[4754]: I0126 17:08:04.378164 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"428912eb-4aed-4926-8b00-f716a43a4399\") " pod="openstack/glance-default-internal-api-0" Jan 26 17:08:04 crc kubenswrapper[4754]: I0126 17:08:04.378200 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r5q6t\" (UniqueName: \"kubernetes.io/projected/428912eb-4aed-4926-8b00-f716a43a4399-kube-api-access-r5q6t\") pod \"glance-default-internal-api-0\" (UID: \"428912eb-4aed-4926-8b00-f716a43a4399\") " pod="openstack/glance-default-internal-api-0" Jan 26 17:08:04 crc kubenswrapper[4754]: I0126 17:08:04.378245 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/428912eb-4aed-4926-8b00-f716a43a4399-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"428912eb-4aed-4926-8b00-f716a43a4399\") " pod="openstack/glance-default-internal-api-0" Jan 26 17:08:04 crc kubenswrapper[4754]: I0126 17:08:04.378400 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/428912eb-4aed-4926-8b00-f716a43a4399-logs\") pod \"glance-default-internal-api-0\" (UID: \"428912eb-4aed-4926-8b00-f716a43a4399\") " pod="openstack/glance-default-internal-api-0" Jan 26 17:08:04 crc kubenswrapper[4754]: I0126 17:08:04.378653 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/428912eb-4aed-4926-8b00-f716a43a4399-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"428912eb-4aed-4926-8b00-f716a43a4399\") " pod="openstack/glance-default-internal-api-0" Jan 26 17:08:04 crc kubenswrapper[4754]: I0126 17:08:04.378991 4754 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"428912eb-4aed-4926-8b00-f716a43a4399\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/glance-default-internal-api-0" Jan 26 17:08:04 crc kubenswrapper[4754]: I0126 17:08:04.383353 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/428912eb-4aed-4926-8b00-f716a43a4399-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"428912eb-4aed-4926-8b00-f716a43a4399\") " pod="openstack/glance-default-internal-api-0" Jan 26 17:08:04 crc kubenswrapper[4754]: I0126 17:08:04.383923 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/428912eb-4aed-4926-8b00-f716a43a4399-scripts\") pod \"glance-default-internal-api-0\" (UID: \"428912eb-4aed-4926-8b00-f716a43a4399\") " pod="openstack/glance-default-internal-api-0" Jan 26 17:08:04 crc kubenswrapper[4754]: I0126 17:08:04.383923 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/428912eb-4aed-4926-8b00-f716a43a4399-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"428912eb-4aed-4926-8b00-f716a43a4399\") " pod="openstack/glance-default-internal-api-0" Jan 26 17:08:04 crc kubenswrapper[4754]: I0126 17:08:04.392923 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/428912eb-4aed-4926-8b00-f716a43a4399-config-data\") pod \"glance-default-internal-api-0\" (UID: \"428912eb-4aed-4926-8b00-f716a43a4399\") " pod="openstack/glance-default-internal-api-0" Jan 26 17:08:04 crc kubenswrapper[4754]: I0126 17:08:04.396980 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r5q6t\" (UniqueName: \"kubernetes.io/projected/428912eb-4aed-4926-8b00-f716a43a4399-kube-api-access-r5q6t\") pod \"glance-default-internal-api-0\" (UID: \"428912eb-4aed-4926-8b00-f716a43a4399\") " pod="openstack/glance-default-internal-api-0" Jan 26 17:08:04 crc kubenswrapper[4754]: I0126 17:08:04.414702 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"428912eb-4aed-4926-8b00-f716a43a4399\") " pod="openstack/glance-default-internal-api-0" Jan 26 17:08:04 crc kubenswrapper[4754]: I0126 17:08:04.458059 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Jan 26 17:08:04 crc kubenswrapper[4754]: I0126 17:08:04.548984 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Jan 26 17:08:04 crc kubenswrapper[4754]: I0126 17:08:04.936374 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-85f95cb57b-99z67" event={"ID":"ea911a5c-f353-4da9-9900-bfcbb234a377","Type":"ContainerStarted","Data":"a63002959b317ff98d2e99a38ab4b17c6e3ea0a8c5ee9526988780fb2de6d4a2"} Jan 26 17:08:04 crc kubenswrapper[4754]: I0126 17:08:04.936750 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-85f95cb57b-99z67" event={"ID":"ea911a5c-f353-4da9-9900-bfcbb234a377","Type":"ContainerStarted","Data":"e920e8dd93fe23ed63bbaaac5aa05943a59d0ede5d920fc5ddf3c89066cffba4"} Jan 26 17:08:04 crc kubenswrapper[4754]: I0126 17:08:04.946639 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-5xmfz" event={"ID":"e934228b-3d4e-4c15-baa6-5c03e0fceb23","Type":"ContainerStarted","Data":"3c17b0e81ba837680723ffaf2c7fb810b45ec801d7900bb2517aa72f884caa3d"} Jan 26 17:08:04 crc kubenswrapper[4754]: I0126 17:08:04.946715 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-5xmfz" event={"ID":"e934228b-3d4e-4c15-baa6-5c03e0fceb23","Type":"ContainerStarted","Data":"bb542a6a9a29f8c1e21fb86297b004fcdf04ae3fc10defb5b4300eb4c1211473"} Jan 26 17:08:04 crc kubenswrapper[4754]: I0126 17:08:04.960512 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-85f95cb57b-99z67" podStartSLOduration=28.960494407 podStartE2EDuration="28.960494407s" podCreationTimestamp="2026-01-26 17:07:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 17:08:04.959968033 +0000 UTC m=+1251.484148477" watchObservedRunningTime="2026-01-26 17:08:04.960494407 +0000 UTC m=+1251.484674841" Jan 26 17:08:04 crc kubenswrapper[4754]: I0126 17:08:04.965031 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6db5cbfc7c-nxhnr" event={"ID":"743bf5b2-5143-406b-b3d7-fa3a067974fd","Type":"ContainerStarted","Data":"7887ee09373110d081d243bf9a15a5d13219c87c84d3b2b6840818f2a57e985a"} Jan 26 17:08:04 crc kubenswrapper[4754]: I0126 17:08:04.965118 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-6db5cbfc7c-nxhnr" podUID="743bf5b2-5143-406b-b3d7-fa3a067974fd" containerName="horizon-log" containerID="cri-o://2634e3c695e0a29fa8d18c8ece6dbbd81e5f69b92a0b1c6ad95ae9182c0884c7" gracePeriod=30 Jan 26 17:08:04 crc kubenswrapper[4754]: I0126 17:08:04.965160 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-6db5cbfc7c-nxhnr" podUID="743bf5b2-5143-406b-b3d7-fa3a067974fd" containerName="horizon" containerID="cri-o://7887ee09373110d081d243bf9a15a5d13219c87c84d3b2b6840818f2a57e985a" gracePeriod=30 Jan 26 17:08:04 crc kubenswrapper[4754]: I0126 17:08:04.973037 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6dff6969b8-5bklb" event={"ID":"0313cdfe-e309-41aa-a5ab-83d7713628f3","Type":"ContainerStarted","Data":"1a5e07ac197d4d5b9cc756714ecaa61f8e3c076c4c9a48a6a1638afc1610d35e"} Jan 26 17:08:04 crc kubenswrapper[4754]: I0126 17:08:04.973076 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6dff6969b8-5bklb" event={"ID":"0313cdfe-e309-41aa-a5ab-83d7713628f3","Type":"ContainerStarted","Data":"4b0cc607a3b447ef731c1de736bbd1af5156aa1beb5462a0b64f5ebed2291561"} Jan 26 17:08:04 crc kubenswrapper[4754]: I0126 17:08:04.975729 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-585cbdbc67-qwfhd" event={"ID":"eabead53-af6d-4405-b76b-a20c5a84f469","Type":"ContainerStarted","Data":"8e092eb22ebc9cba28e6b243f38f358bf2cca40a97352fd16fe630e746bd0f10"} Jan 26 17:08:04 crc kubenswrapper[4754]: I0126 17:08:04.975754 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-585cbdbc67-qwfhd" event={"ID":"eabead53-af6d-4405-b76b-a20c5a84f469","Type":"ContainerStarted","Data":"04100d29f24966812f52c17d77db99b3275e2115f5be21feb58b8a03dbfbd039"} Jan 26 17:08:04 crc kubenswrapper[4754]: I0126 17:08:04.975859 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-585cbdbc67-qwfhd" podUID="eabead53-af6d-4405-b76b-a20c5a84f469" containerName="horizon-log" containerID="cri-o://8e092eb22ebc9cba28e6b243f38f358bf2cca40a97352fd16fe630e746bd0f10" gracePeriod=30 Jan 26 17:08:04 crc kubenswrapper[4754]: I0126 17:08:04.976085 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-585cbdbc67-qwfhd" podUID="eabead53-af6d-4405-b76b-a20c5a84f469" containerName="horizon" containerID="cri-o://04100d29f24966812f52c17d77db99b3275e2115f5be21feb58b8a03dbfbd039" gracePeriod=30 Jan 26 17:08:04 crc kubenswrapper[4754]: I0126 17:08:04.993528 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-5xmfz" podStartSLOduration=13.993509421 podStartE2EDuration="13.993509421s" podCreationTimestamp="2026-01-26 17:07:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 17:08:04.983041617 +0000 UTC m=+1251.507222041" watchObservedRunningTime="2026-01-26 17:08:04.993509421 +0000 UTC m=+1251.517689855" Jan 26 17:08:05 crc kubenswrapper[4754]: I0126 17:08:05.012454 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-585cbdbc67-qwfhd" podStartSLOduration=3.615983367 podStartE2EDuration="37.012430827s" podCreationTimestamp="2026-01-26 17:07:28 +0000 UTC" firstStartedPulling="2026-01-26 17:07:29.615983558 +0000 UTC m=+1216.140163992" lastFinishedPulling="2026-01-26 17:08:03.012431018 +0000 UTC m=+1249.536611452" observedRunningTime="2026-01-26 17:08:05.005901935 +0000 UTC m=+1251.530082369" watchObservedRunningTime="2026-01-26 17:08:05.012430827 +0000 UTC m=+1251.536611271" Jan 26 17:08:05 crc kubenswrapper[4754]: I0126 17:08:05.048361 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-6db5cbfc7c-nxhnr" podStartSLOduration=4.178907835 podStartE2EDuration="38.048340106s" podCreationTimestamp="2026-01-26 17:07:27 +0000 UTC" firstStartedPulling="2026-01-26 17:07:29.405801675 +0000 UTC m=+1215.929982109" lastFinishedPulling="2026-01-26 17:08:03.275233946 +0000 UTC m=+1249.799414380" observedRunningTime="2026-01-26 17:08:05.037967545 +0000 UTC m=+1251.562147989" watchObservedRunningTime="2026-01-26 17:08:05.048340106 +0000 UTC m=+1251.572520540" Jan 26 17:08:05 crc kubenswrapper[4754]: I0126 17:08:05.076072 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Jan 26 17:08:05 crc kubenswrapper[4754]: W0126 17:08:05.485685 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod428912eb_4aed_4926_8b00_f716a43a4399.slice/crio-e4ef61878ff787d988699a25e243588a87ca9d9e98ae9807c4af1e5a053dfa12 WatchSource:0}: Error finding container e4ef61878ff787d988699a25e243588a87ca9d9e98ae9807c4af1e5a053dfa12: Status 404 returned error can't find the container with id e4ef61878ff787d988699a25e243588a87ca9d9e98ae9807c4af1e5a053dfa12 Jan 26 17:08:05 crc kubenswrapper[4754]: I0126 17:08:05.778130 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="15b16489-4c64-4a41-bee1-c590e38a87e5" path="/var/lib/kubelet/pods/15b16489-4c64-4a41-bee1-c590e38a87e5/volumes" Jan 26 17:08:05 crc kubenswrapper[4754]: I0126 17:08:05.779493 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d93e77-fe09-47ec-acd8-eced4e88a416" path="/var/lib/kubelet/pods/31d93e77-fe09-47ec-acd8-eced4e88a416/volumes" Jan 26 17:08:05 crc kubenswrapper[4754]: I0126 17:08:05.780032 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="522901db-3c0d-4cf5-a534-690eac9ae4db" path="/var/lib/kubelet/pods/522901db-3c0d-4cf5-a534-690eac9ae4db/volumes" Jan 26 17:08:05 crc kubenswrapper[4754]: I0126 17:08:05.781440 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5d5e322d-aa8b-4f36-8257-09d1ccab02c6" path="/var/lib/kubelet/pods/5d5e322d-aa8b-4f36-8257-09d1ccab02c6/volumes" Jan 26 17:08:05 crc kubenswrapper[4754]: I0126 17:08:05.935846 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Jan 26 17:08:05 crc kubenswrapper[4754]: I0126 17:08:05.991169 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"428912eb-4aed-4926-8b00-f716a43a4399","Type":"ContainerStarted","Data":"e4ef61878ff787d988699a25e243588a87ca9d9e98ae9807c4af1e5a053dfa12"} Jan 26 17:08:05 crc kubenswrapper[4754]: I0126 17:08:05.994175 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6dff6969b8-5bklb" event={"ID":"0313cdfe-e309-41aa-a5ab-83d7713628f3","Type":"ContainerStarted","Data":"202d0a33717d86a8ef263b4eb4f9b301c52aee53b2a12762a1c621995a8e045a"} Jan 26 17:08:05 crc kubenswrapper[4754]: I0126 17:08:05.995713 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"4a8fafa1-3d82-4a16-85e1-fd3116bc998a","Type":"ContainerStarted","Data":"1816ede461917fc9aed1037d8a2a51d98ac1b2d9c969bae3690e1f12841c4111"} Jan 26 17:08:07 crc kubenswrapper[4754]: I0126 17:08:07.026702 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-6dff6969b8-5bklb" podStartSLOduration=31.026682337 podStartE2EDuration="31.026682337s" podCreationTimestamp="2026-01-26 17:07:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 17:08:07.017940578 +0000 UTC m=+1253.542121012" watchObservedRunningTime="2026-01-26 17:08:07.026682337 +0000 UTC m=+1253.550862771" Jan 26 17:08:07 crc kubenswrapper[4754]: I0126 17:08:07.078705 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-85f95cb57b-99z67" Jan 26 17:08:07 crc kubenswrapper[4754]: I0126 17:08:07.078766 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-85f95cb57b-99z67" Jan 26 17:08:07 crc kubenswrapper[4754]: I0126 17:08:07.216968 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-6dff6969b8-5bklb" Jan 26 17:08:07 crc kubenswrapper[4754]: I0126 17:08:07.218009 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-6dff6969b8-5bklb" Jan 26 17:08:08 crc kubenswrapper[4754]: I0126 17:08:08.010786 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"4a8fafa1-3d82-4a16-85e1-fd3116bc998a","Type":"ContainerStarted","Data":"68b9e43495925deaaeb41ff5279207dddd543bad6b2dd86201727f1e8bd1577d"} Jan 26 17:08:08 crc kubenswrapper[4754]: I0126 17:08:08.012607 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"428912eb-4aed-4926-8b00-f716a43a4399","Type":"ContainerStarted","Data":"17a24ca2bd71c5f9883d7d6763e457a9ad1791b53a390081e373f23d8c1cc0af"} Jan 26 17:08:08 crc kubenswrapper[4754]: I0126 17:08:08.219230 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-6db5cbfc7c-nxhnr" Jan 26 17:08:08 crc kubenswrapper[4754]: I0126 17:08:08.631085 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-585cbdbc67-qwfhd" Jan 26 17:08:17 crc kubenswrapper[4754]: I0126 17:08:17.082561 4754 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-85f95cb57b-99z67" podUID="ea911a5c-f353-4da9-9900-bfcbb234a377" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.154:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.154:8443: connect: connection refused" Jan 26 17:08:17 crc kubenswrapper[4754]: I0126 17:08:17.105224 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"4a8fafa1-3d82-4a16-85e1-fd3116bc998a","Type":"ContainerStarted","Data":"0c78645b52c9649d8e28a70685c6baf4ee822767bcf49eaf0c693d92ab3e3521"} Jan 26 17:08:17 crc kubenswrapper[4754]: I0126 17:08:17.107879 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"428912eb-4aed-4926-8b00-f716a43a4399","Type":"ContainerStarted","Data":"34521a84db431d82e60898db22e047282e63e4072af221baedb12427e8c15b10"} Jan 26 17:08:17 crc kubenswrapper[4754]: I0126 17:08:17.219356 4754 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-6dff6969b8-5bklb" podUID="0313cdfe-e309-41aa-a5ab-83d7713628f3" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.155:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.155:8443: connect: connection refused" Jan 26 17:08:20 crc kubenswrapper[4754]: I0126 17:08:20.155832 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=17.155814793 podStartE2EDuration="17.155814793s" podCreationTimestamp="2026-01-26 17:08:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 17:08:20.150942627 +0000 UTC m=+1266.675123081" watchObservedRunningTime="2026-01-26 17:08:20.155814793 +0000 UTC m=+1266.679995227" Jan 26 17:08:20 crc kubenswrapper[4754]: I0126 17:08:20.175519 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=17.175504329 podStartE2EDuration="17.175504329s" podCreationTimestamp="2026-01-26 17:08:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 17:08:20.17282771 +0000 UTC m=+1266.697008144" watchObservedRunningTime="2026-01-26 17:08:20.175504329 +0000 UTC m=+1266.699684763" Jan 26 17:08:24 crc kubenswrapper[4754]: I0126 17:08:24.460284 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Jan 26 17:08:24 crc kubenswrapper[4754]: I0126 17:08:24.460632 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Jan 26 17:08:24 crc kubenswrapper[4754]: I0126 17:08:24.503394 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Jan 26 17:08:24 crc kubenswrapper[4754]: I0126 17:08:24.531682 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Jan 26 17:08:24 crc kubenswrapper[4754]: I0126 17:08:24.549584 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Jan 26 17:08:24 crc kubenswrapper[4754]: I0126 17:08:24.549643 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Jan 26 17:08:24 crc kubenswrapper[4754]: I0126 17:08:24.611172 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Jan 26 17:08:24 crc kubenswrapper[4754]: I0126 17:08:24.611280 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Jan 26 17:08:25 crc kubenswrapper[4754]: I0126 17:08:25.172360 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Jan 26 17:08:25 crc kubenswrapper[4754]: I0126 17:08:25.172400 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Jan 26 17:08:25 crc kubenswrapper[4754]: I0126 17:08:25.172413 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Jan 26 17:08:25 crc kubenswrapper[4754]: I0126 17:08:25.172427 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Jan 26 17:08:26 crc kubenswrapper[4754]: I0126 17:08:26.181046 4754 generic.go:334] "Generic (PLEG): container finished" podID="e934228b-3d4e-4c15-baa6-5c03e0fceb23" containerID="3c17b0e81ba837680723ffaf2c7fb810b45ec801d7900bb2517aa72f884caa3d" exitCode=0 Jan 26 17:08:26 crc kubenswrapper[4754]: I0126 17:08:26.181169 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-5xmfz" event={"ID":"e934228b-3d4e-4c15-baa6-5c03e0fceb23","Type":"ContainerDied","Data":"3c17b0e81ba837680723ffaf2c7fb810b45ec801d7900bb2517aa72f884caa3d"} Jan 26 17:08:27 crc kubenswrapper[4754]: I0126 17:08:27.200425 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Jan 26 17:08:27 crc kubenswrapper[4754]: I0126 17:08:27.201597 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Jan 26 17:08:27 crc kubenswrapper[4754]: I0126 17:08:27.254208 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Jan 26 17:08:27 crc kubenswrapper[4754]: I0126 17:08:27.254611 4754 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Jan 26 17:08:27 crc kubenswrapper[4754]: I0126 17:08:27.278128 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Jan 26 17:08:29 crc kubenswrapper[4754]: I0126 17:08:29.056218 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-85f95cb57b-99z67" Jan 26 17:08:29 crc kubenswrapper[4754]: I0126 17:08:29.078465 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-6dff6969b8-5bklb" Jan 26 17:08:29 crc kubenswrapper[4754]: I0126 17:08:29.688712 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-5xmfz" Jan 26 17:08:29 crc kubenswrapper[4754]: I0126 17:08:29.726330 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e934228b-3d4e-4c15-baa6-5c03e0fceb23-scripts\") pod \"e934228b-3d4e-4c15-baa6-5c03e0fceb23\" (UID: \"e934228b-3d4e-4c15-baa6-5c03e0fceb23\") " Jan 26 17:08:29 crc kubenswrapper[4754]: I0126 17:08:29.726714 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e934228b-3d4e-4c15-baa6-5c03e0fceb23-combined-ca-bundle\") pod \"e934228b-3d4e-4c15-baa6-5c03e0fceb23\" (UID: \"e934228b-3d4e-4c15-baa6-5c03e0fceb23\") " Jan 26 17:08:29 crc kubenswrapper[4754]: I0126 17:08:29.726850 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fwqwc\" (UniqueName: \"kubernetes.io/projected/e934228b-3d4e-4c15-baa6-5c03e0fceb23-kube-api-access-fwqwc\") pod \"e934228b-3d4e-4c15-baa6-5c03e0fceb23\" (UID: \"e934228b-3d4e-4c15-baa6-5c03e0fceb23\") " Jan 26 17:08:29 crc kubenswrapper[4754]: I0126 17:08:29.726952 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/e934228b-3d4e-4c15-baa6-5c03e0fceb23-fernet-keys\") pod \"e934228b-3d4e-4c15-baa6-5c03e0fceb23\" (UID: \"e934228b-3d4e-4c15-baa6-5c03e0fceb23\") " Jan 26 17:08:29 crc kubenswrapper[4754]: I0126 17:08:29.727000 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e934228b-3d4e-4c15-baa6-5c03e0fceb23-config-data\") pod \"e934228b-3d4e-4c15-baa6-5c03e0fceb23\" (UID: \"e934228b-3d4e-4c15-baa6-5c03e0fceb23\") " Jan 26 17:08:29 crc kubenswrapper[4754]: I0126 17:08:29.727094 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/e934228b-3d4e-4c15-baa6-5c03e0fceb23-credential-keys\") pod \"e934228b-3d4e-4c15-baa6-5c03e0fceb23\" (UID: \"e934228b-3d4e-4c15-baa6-5c03e0fceb23\") " Jan 26 17:08:29 crc kubenswrapper[4754]: I0126 17:08:29.735398 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e934228b-3d4e-4c15-baa6-5c03e0fceb23-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "e934228b-3d4e-4c15-baa6-5c03e0fceb23" (UID: "e934228b-3d4e-4c15-baa6-5c03e0fceb23"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:08:29 crc kubenswrapper[4754]: I0126 17:08:29.737898 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e934228b-3d4e-4c15-baa6-5c03e0fceb23-kube-api-access-fwqwc" (OuterVolumeSpecName: "kube-api-access-fwqwc") pod "e934228b-3d4e-4c15-baa6-5c03e0fceb23" (UID: "e934228b-3d4e-4c15-baa6-5c03e0fceb23"). InnerVolumeSpecName "kube-api-access-fwqwc". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:08:29 crc kubenswrapper[4754]: I0126 17:08:29.744834 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e934228b-3d4e-4c15-baa6-5c03e0fceb23-scripts" (OuterVolumeSpecName: "scripts") pod "e934228b-3d4e-4c15-baa6-5c03e0fceb23" (UID: "e934228b-3d4e-4c15-baa6-5c03e0fceb23"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:08:29 crc kubenswrapper[4754]: I0126 17:08:29.749542 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e934228b-3d4e-4c15-baa6-5c03e0fceb23-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "e934228b-3d4e-4c15-baa6-5c03e0fceb23" (UID: "e934228b-3d4e-4c15-baa6-5c03e0fceb23"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:08:29 crc kubenswrapper[4754]: I0126 17:08:29.767930 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e934228b-3d4e-4c15-baa6-5c03e0fceb23-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e934228b-3d4e-4c15-baa6-5c03e0fceb23" (UID: "e934228b-3d4e-4c15-baa6-5c03e0fceb23"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:08:29 crc kubenswrapper[4754]: I0126 17:08:29.769335 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e934228b-3d4e-4c15-baa6-5c03e0fceb23-config-data" (OuterVolumeSpecName: "config-data") pod "e934228b-3d4e-4c15-baa6-5c03e0fceb23" (UID: "e934228b-3d4e-4c15-baa6-5c03e0fceb23"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:08:29 crc kubenswrapper[4754]: I0126 17:08:29.830982 4754 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e934228b-3d4e-4c15-baa6-5c03e0fceb23-scripts\") on node \"crc\" DevicePath \"\"" Jan 26 17:08:29 crc kubenswrapper[4754]: I0126 17:08:29.831207 4754 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e934228b-3d4e-4c15-baa6-5c03e0fceb23-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 26 17:08:29 crc kubenswrapper[4754]: I0126 17:08:29.831304 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fwqwc\" (UniqueName: \"kubernetes.io/projected/e934228b-3d4e-4c15-baa6-5c03e0fceb23-kube-api-access-fwqwc\") on node \"crc\" DevicePath \"\"" Jan 26 17:08:29 crc kubenswrapper[4754]: I0126 17:08:29.831380 4754 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/e934228b-3d4e-4c15-baa6-5c03e0fceb23-fernet-keys\") on node \"crc\" DevicePath \"\"" Jan 26 17:08:29 crc kubenswrapper[4754]: I0126 17:08:29.831439 4754 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e934228b-3d4e-4c15-baa6-5c03e0fceb23-config-data\") on node \"crc\" DevicePath \"\"" Jan 26 17:08:29 crc kubenswrapper[4754]: I0126 17:08:29.831496 4754 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/e934228b-3d4e-4c15-baa6-5c03e0fceb23-credential-keys\") on node \"crc\" DevicePath \"\"" Jan 26 17:08:30 crc kubenswrapper[4754]: I0126 17:08:30.238587 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-5xmfz" event={"ID":"e934228b-3d4e-4c15-baa6-5c03e0fceb23","Type":"ContainerDied","Data":"bb542a6a9a29f8c1e21fb86297b004fcdf04ae3fc10defb5b4300eb4c1211473"} Jan 26 17:08:30 crc kubenswrapper[4754]: I0126 17:08:30.238629 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bb542a6a9a29f8c1e21fb86297b004fcdf04ae3fc10defb5b4300eb4c1211473" Jan 26 17:08:30 crc kubenswrapper[4754]: I0126 17:08:30.238707 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-5xmfz" Jan 26 17:08:30 crc kubenswrapper[4754]: I0126 17:08:30.792195 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-6c774f65c8-h55rj"] Jan 26 17:08:30 crc kubenswrapper[4754]: E0126 17:08:30.793127 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e934228b-3d4e-4c15-baa6-5c03e0fceb23" containerName="keystone-bootstrap" Jan 26 17:08:30 crc kubenswrapper[4754]: I0126 17:08:30.793253 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="e934228b-3d4e-4c15-baa6-5c03e0fceb23" containerName="keystone-bootstrap" Jan 26 17:08:30 crc kubenswrapper[4754]: I0126 17:08:30.793503 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="e934228b-3d4e-4c15-baa6-5c03e0fceb23" containerName="keystone-bootstrap" Jan 26 17:08:30 crc kubenswrapper[4754]: I0126 17:08:30.794112 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-6c774f65c8-h55rj" Jan 26 17:08:30 crc kubenswrapper[4754]: I0126 17:08:30.795755 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Jan 26 17:08:30 crc kubenswrapper[4754]: I0126 17:08:30.796958 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Jan 26 17:08:30 crc kubenswrapper[4754]: I0126 17:08:30.797289 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Jan 26 17:08:30 crc kubenswrapper[4754]: I0126 17:08:30.809102 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Jan 26 17:08:30 crc kubenswrapper[4754]: I0126 17:08:30.809366 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-r42mh" Jan 26 17:08:30 crc kubenswrapper[4754]: I0126 17:08:30.814156 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Jan 26 17:08:30 crc kubenswrapper[4754]: I0126 17:08:30.828202 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-6c774f65c8-h55rj"] Jan 26 17:08:30 crc kubenswrapper[4754]: I0126 17:08:30.830486 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-85f95cb57b-99z67" Jan 26 17:08:30 crc kubenswrapper[4754]: I0126 17:08:30.847619 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/928f07ae-77be-48d4-ba56-daaa5ff400f0-combined-ca-bundle\") pod \"keystone-6c774f65c8-h55rj\" (UID: \"928f07ae-77be-48d4-ba56-daaa5ff400f0\") " pod="openstack/keystone-6c774f65c8-h55rj" Jan 26 17:08:30 crc kubenswrapper[4754]: I0126 17:08:30.847660 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/928f07ae-77be-48d4-ba56-daaa5ff400f0-config-data\") pod \"keystone-6c774f65c8-h55rj\" (UID: \"928f07ae-77be-48d4-ba56-daaa5ff400f0\") " pod="openstack/keystone-6c774f65c8-h55rj" Jan 26 17:08:30 crc kubenswrapper[4754]: I0126 17:08:30.847771 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-prfdw\" (UniqueName: \"kubernetes.io/projected/928f07ae-77be-48d4-ba56-daaa5ff400f0-kube-api-access-prfdw\") pod \"keystone-6c774f65c8-h55rj\" (UID: \"928f07ae-77be-48d4-ba56-daaa5ff400f0\") " pod="openstack/keystone-6c774f65c8-h55rj" Jan 26 17:08:30 crc kubenswrapper[4754]: I0126 17:08:30.847827 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/928f07ae-77be-48d4-ba56-daaa5ff400f0-fernet-keys\") pod \"keystone-6c774f65c8-h55rj\" (UID: \"928f07ae-77be-48d4-ba56-daaa5ff400f0\") " pod="openstack/keystone-6c774f65c8-h55rj" Jan 26 17:08:30 crc kubenswrapper[4754]: I0126 17:08:30.847865 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/928f07ae-77be-48d4-ba56-daaa5ff400f0-public-tls-certs\") pod \"keystone-6c774f65c8-h55rj\" (UID: \"928f07ae-77be-48d4-ba56-daaa5ff400f0\") " pod="openstack/keystone-6c774f65c8-h55rj" Jan 26 17:08:30 crc kubenswrapper[4754]: I0126 17:08:30.847911 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/928f07ae-77be-48d4-ba56-daaa5ff400f0-credential-keys\") pod \"keystone-6c774f65c8-h55rj\" (UID: \"928f07ae-77be-48d4-ba56-daaa5ff400f0\") " pod="openstack/keystone-6c774f65c8-h55rj" Jan 26 17:08:30 crc kubenswrapper[4754]: I0126 17:08:30.847989 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/928f07ae-77be-48d4-ba56-daaa5ff400f0-internal-tls-certs\") pod \"keystone-6c774f65c8-h55rj\" (UID: \"928f07ae-77be-48d4-ba56-daaa5ff400f0\") " pod="openstack/keystone-6c774f65c8-h55rj" Jan 26 17:08:30 crc kubenswrapper[4754]: I0126 17:08:30.848082 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/928f07ae-77be-48d4-ba56-daaa5ff400f0-scripts\") pod \"keystone-6c774f65c8-h55rj\" (UID: \"928f07ae-77be-48d4-ba56-daaa5ff400f0\") " pod="openstack/keystone-6c774f65c8-h55rj" Jan 26 17:08:30 crc kubenswrapper[4754]: I0126 17:08:30.949592 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/928f07ae-77be-48d4-ba56-daaa5ff400f0-fernet-keys\") pod \"keystone-6c774f65c8-h55rj\" (UID: \"928f07ae-77be-48d4-ba56-daaa5ff400f0\") " pod="openstack/keystone-6c774f65c8-h55rj" Jan 26 17:08:30 crc kubenswrapper[4754]: I0126 17:08:30.949656 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/928f07ae-77be-48d4-ba56-daaa5ff400f0-public-tls-certs\") pod \"keystone-6c774f65c8-h55rj\" (UID: \"928f07ae-77be-48d4-ba56-daaa5ff400f0\") " pod="openstack/keystone-6c774f65c8-h55rj" Jan 26 17:08:30 crc kubenswrapper[4754]: I0126 17:08:30.949719 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/928f07ae-77be-48d4-ba56-daaa5ff400f0-credential-keys\") pod \"keystone-6c774f65c8-h55rj\" (UID: \"928f07ae-77be-48d4-ba56-daaa5ff400f0\") " pod="openstack/keystone-6c774f65c8-h55rj" Jan 26 17:08:30 crc kubenswrapper[4754]: I0126 17:08:30.949774 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/928f07ae-77be-48d4-ba56-daaa5ff400f0-internal-tls-certs\") pod \"keystone-6c774f65c8-h55rj\" (UID: \"928f07ae-77be-48d4-ba56-daaa5ff400f0\") " pod="openstack/keystone-6c774f65c8-h55rj" Jan 26 17:08:30 crc kubenswrapper[4754]: I0126 17:08:30.949831 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/928f07ae-77be-48d4-ba56-daaa5ff400f0-scripts\") pod \"keystone-6c774f65c8-h55rj\" (UID: \"928f07ae-77be-48d4-ba56-daaa5ff400f0\") " pod="openstack/keystone-6c774f65c8-h55rj" Jan 26 17:08:30 crc kubenswrapper[4754]: I0126 17:08:30.949877 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/928f07ae-77be-48d4-ba56-daaa5ff400f0-combined-ca-bundle\") pod \"keystone-6c774f65c8-h55rj\" (UID: \"928f07ae-77be-48d4-ba56-daaa5ff400f0\") " pod="openstack/keystone-6c774f65c8-h55rj" Jan 26 17:08:30 crc kubenswrapper[4754]: I0126 17:08:30.949902 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/928f07ae-77be-48d4-ba56-daaa5ff400f0-config-data\") pod \"keystone-6c774f65c8-h55rj\" (UID: \"928f07ae-77be-48d4-ba56-daaa5ff400f0\") " pod="openstack/keystone-6c774f65c8-h55rj" Jan 26 17:08:30 crc kubenswrapper[4754]: I0126 17:08:30.949944 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-prfdw\" (UniqueName: \"kubernetes.io/projected/928f07ae-77be-48d4-ba56-daaa5ff400f0-kube-api-access-prfdw\") pod \"keystone-6c774f65c8-h55rj\" (UID: \"928f07ae-77be-48d4-ba56-daaa5ff400f0\") " pod="openstack/keystone-6c774f65c8-h55rj" Jan 26 17:08:30 crc kubenswrapper[4754]: I0126 17:08:30.956565 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/928f07ae-77be-48d4-ba56-daaa5ff400f0-credential-keys\") pod \"keystone-6c774f65c8-h55rj\" (UID: \"928f07ae-77be-48d4-ba56-daaa5ff400f0\") " pod="openstack/keystone-6c774f65c8-h55rj" Jan 26 17:08:30 crc kubenswrapper[4754]: I0126 17:08:30.960360 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/928f07ae-77be-48d4-ba56-daaa5ff400f0-combined-ca-bundle\") pod \"keystone-6c774f65c8-h55rj\" (UID: \"928f07ae-77be-48d4-ba56-daaa5ff400f0\") " pod="openstack/keystone-6c774f65c8-h55rj" Jan 26 17:08:30 crc kubenswrapper[4754]: I0126 17:08:30.960992 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/928f07ae-77be-48d4-ba56-daaa5ff400f0-scripts\") pod \"keystone-6c774f65c8-h55rj\" (UID: \"928f07ae-77be-48d4-ba56-daaa5ff400f0\") " pod="openstack/keystone-6c774f65c8-h55rj" Jan 26 17:08:30 crc kubenswrapper[4754]: I0126 17:08:30.961394 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/928f07ae-77be-48d4-ba56-daaa5ff400f0-public-tls-certs\") pod \"keystone-6c774f65c8-h55rj\" (UID: \"928f07ae-77be-48d4-ba56-daaa5ff400f0\") " pod="openstack/keystone-6c774f65c8-h55rj" Jan 26 17:08:30 crc kubenswrapper[4754]: I0126 17:08:30.961842 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/928f07ae-77be-48d4-ba56-daaa5ff400f0-internal-tls-certs\") pod \"keystone-6c774f65c8-h55rj\" (UID: \"928f07ae-77be-48d4-ba56-daaa5ff400f0\") " pod="openstack/keystone-6c774f65c8-h55rj" Jan 26 17:08:30 crc kubenswrapper[4754]: I0126 17:08:30.963152 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/928f07ae-77be-48d4-ba56-daaa5ff400f0-fernet-keys\") pod \"keystone-6c774f65c8-h55rj\" (UID: \"928f07ae-77be-48d4-ba56-daaa5ff400f0\") " pod="openstack/keystone-6c774f65c8-h55rj" Jan 26 17:08:30 crc kubenswrapper[4754]: I0126 17:08:30.966476 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/928f07ae-77be-48d4-ba56-daaa5ff400f0-config-data\") pod \"keystone-6c774f65c8-h55rj\" (UID: \"928f07ae-77be-48d4-ba56-daaa5ff400f0\") " pod="openstack/keystone-6c774f65c8-h55rj" Jan 26 17:08:30 crc kubenswrapper[4754]: I0126 17:08:30.967140 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-6dff6969b8-5bklb" Jan 26 17:08:30 crc kubenswrapper[4754]: I0126 17:08:30.978058 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-prfdw\" (UniqueName: \"kubernetes.io/projected/928f07ae-77be-48d4-ba56-daaa5ff400f0-kube-api-access-prfdw\") pod \"keystone-6c774f65c8-h55rj\" (UID: \"928f07ae-77be-48d4-ba56-daaa5ff400f0\") " pod="openstack/keystone-6c774f65c8-h55rj" Jan 26 17:08:31 crc kubenswrapper[4754]: I0126 17:08:31.059467 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-85f95cb57b-99z67"] Jan 26 17:08:31 crc kubenswrapper[4754]: I0126 17:08:31.114091 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-6c774f65c8-h55rj" Jan 26 17:08:31 crc kubenswrapper[4754]: I0126 17:08:31.263844 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-9spps" event={"ID":"ea094395-bb3c-4464-829f-313ab6b9e14c","Type":"ContainerStarted","Data":"a0aac0c2b2fa1e7140caf64c6f53502aa2cdbab58168b149446176ac7c4866bb"} Jan 26 17:08:31 crc kubenswrapper[4754]: I0126 17:08:31.266954 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-7hskr" event={"ID":"ab3f008d-78ab-4b73-962d-4447f312c9fa","Type":"ContainerStarted","Data":"eda7e173767be2f94247fe84877d1dc7a95556f64d1cc275cf4ccf7dd0a1ae64"} Jan 26 17:08:31 crc kubenswrapper[4754]: I0126 17:08:31.274292 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d84d5918-51f5-4131-9caf-9380004a0ca5","Type":"ContainerStarted","Data":"2b6a9e46662327f03d044bfee7f9926d7e4287b3bd44ae06415f5f1bf5a68abc"} Jan 26 17:08:31 crc kubenswrapper[4754]: I0126 17:08:31.277027 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-whnhn" event={"ID":"01246f69-d6bc-4f70-a50f-4d7c8e4a7620","Type":"ContainerStarted","Data":"fa8641d204e41cbb652b48277e0e1d4c609dfb5b9fb94837cbf80deac687370c"} Jan 26 17:08:31 crc kubenswrapper[4754]: I0126 17:08:31.277178 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-85f95cb57b-99z67" podUID="ea911a5c-f353-4da9-9900-bfcbb234a377" containerName="horizon-log" containerID="cri-o://e920e8dd93fe23ed63bbaaac5aa05943a59d0ede5d920fc5ddf3c89066cffba4" gracePeriod=30 Jan 26 17:08:31 crc kubenswrapper[4754]: I0126 17:08:31.277261 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-85f95cb57b-99z67" podUID="ea911a5c-f353-4da9-9900-bfcbb234a377" containerName="horizon" containerID="cri-o://a63002959b317ff98d2e99a38ab4b17c6e3ea0a8c5ee9526988780fb2de6d4a2" gracePeriod=30 Jan 26 17:08:31 crc kubenswrapper[4754]: I0126 17:08:31.340460 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-9spps" podStartSLOduration=2.78560709 podStartE2EDuration="1m3.340438725s" podCreationTimestamp="2026-01-26 17:07:28 +0000 UTC" firstStartedPulling="2026-01-26 17:07:29.635829272 +0000 UTC m=+1216.160009706" lastFinishedPulling="2026-01-26 17:08:30.190660907 +0000 UTC m=+1276.714841341" observedRunningTime="2026-01-26 17:08:31.295890209 +0000 UTC m=+1277.820070643" watchObservedRunningTime="2026-01-26 17:08:31.340438725 +0000 UTC m=+1277.864619169" Jan 26 17:08:31 crc kubenswrapper[4754]: I0126 17:08:31.343088 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-whnhn" podStartSLOduration=3.575311961 podStartE2EDuration="1m4.343052933s" podCreationTimestamp="2026-01-26 17:07:27 +0000 UTC" firstStartedPulling="2026-01-26 17:07:29.423158921 +0000 UTC m=+1215.947339345" lastFinishedPulling="2026-01-26 17:08:30.190899883 +0000 UTC m=+1276.715080317" observedRunningTime="2026-01-26 17:08:31.314144027 +0000 UTC m=+1277.838324461" watchObservedRunningTime="2026-01-26 17:08:31.343052933 +0000 UTC m=+1277.867233367" Jan 26 17:08:31 crc kubenswrapper[4754]: I0126 17:08:31.352056 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-7hskr" podStartSLOduration=3.496307057 podStartE2EDuration="1m4.352039408s" podCreationTimestamp="2026-01-26 17:07:27 +0000 UTC" firstStartedPulling="2026-01-26 17:07:29.334895265 +0000 UTC m=+1215.859075699" lastFinishedPulling="2026-01-26 17:08:30.190627616 +0000 UTC m=+1276.714808050" observedRunningTime="2026-01-26 17:08:31.349771419 +0000 UTC m=+1277.873951853" watchObservedRunningTime="2026-01-26 17:08:31.352039408 +0000 UTC m=+1277.876219842" Jan 26 17:08:31 crc kubenswrapper[4754]: I0126 17:08:31.632953 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-6c774f65c8-h55rj"] Jan 26 17:08:32 crc kubenswrapper[4754]: I0126 17:08:32.296468 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-6c774f65c8-h55rj" event={"ID":"928f07ae-77be-48d4-ba56-daaa5ff400f0","Type":"ContainerStarted","Data":"3e64a6cabbd73653330237bd6752ed0ffa6bc7cb92cebda2fded2fc36c6bcec8"} Jan 26 17:08:32 crc kubenswrapper[4754]: I0126 17:08:32.296872 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-6c774f65c8-h55rj" event={"ID":"928f07ae-77be-48d4-ba56-daaa5ff400f0","Type":"ContainerStarted","Data":"1719a7a1dbe0a563ec7eeef8be642c29831052849f1c8c45f5b5485bc44cac1a"} Jan 26 17:08:32 crc kubenswrapper[4754]: I0126 17:08:32.298324 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-6c774f65c8-h55rj" Jan 26 17:08:35 crc kubenswrapper[4754]: E0126 17:08:35.307683 4754 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podeabead53_af6d_4405_b76b_a20c5a84f469.slice/crio-conmon-04100d29f24966812f52c17d77db99b3275e2115f5be21feb58b8a03dbfbd039.scope\": RecentStats: unable to find data in memory cache]" Jan 26 17:08:35 crc kubenswrapper[4754]: I0126 17:08:35.342402 4754 generic.go:334] "Generic (PLEG): container finished" podID="eabead53-af6d-4405-b76b-a20c5a84f469" containerID="04100d29f24966812f52c17d77db99b3275e2115f5be21feb58b8a03dbfbd039" exitCode=137 Jan 26 17:08:35 crc kubenswrapper[4754]: I0126 17:08:35.342434 4754 generic.go:334] "Generic (PLEG): container finished" podID="eabead53-af6d-4405-b76b-a20c5a84f469" containerID="8e092eb22ebc9cba28e6b243f38f358bf2cca40a97352fd16fe630e746bd0f10" exitCode=137 Jan 26 17:08:35 crc kubenswrapper[4754]: I0126 17:08:35.342480 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-585cbdbc67-qwfhd" event={"ID":"eabead53-af6d-4405-b76b-a20c5a84f469","Type":"ContainerDied","Data":"04100d29f24966812f52c17d77db99b3275e2115f5be21feb58b8a03dbfbd039"} Jan 26 17:08:35 crc kubenswrapper[4754]: I0126 17:08:35.342505 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-585cbdbc67-qwfhd" event={"ID":"eabead53-af6d-4405-b76b-a20c5a84f469","Type":"ContainerDied","Data":"8e092eb22ebc9cba28e6b243f38f358bf2cca40a97352fd16fe630e746bd0f10"} Jan 26 17:08:35 crc kubenswrapper[4754]: I0126 17:08:35.352209 4754 generic.go:334] "Generic (PLEG): container finished" podID="ea911a5c-f353-4da9-9900-bfcbb234a377" containerID="a63002959b317ff98d2e99a38ab4b17c6e3ea0a8c5ee9526988780fb2de6d4a2" exitCode=0 Jan 26 17:08:35 crc kubenswrapper[4754]: I0126 17:08:35.352249 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-85f95cb57b-99z67" event={"ID":"ea911a5c-f353-4da9-9900-bfcbb234a377","Type":"ContainerDied","Data":"a63002959b317ff98d2e99a38ab4b17c6e3ea0a8c5ee9526988780fb2de6d4a2"} Jan 26 17:08:35 crc kubenswrapper[4754]: I0126 17:08:35.354295 4754 generic.go:334] "Generic (PLEG): container finished" podID="743bf5b2-5143-406b-b3d7-fa3a067974fd" containerID="7887ee09373110d081d243bf9a15a5d13219c87c84d3b2b6840818f2a57e985a" exitCode=137 Jan 26 17:08:35 crc kubenswrapper[4754]: I0126 17:08:35.354318 4754 generic.go:334] "Generic (PLEG): container finished" podID="743bf5b2-5143-406b-b3d7-fa3a067974fd" containerID="2634e3c695e0a29fa8d18c8ece6dbbd81e5f69b92a0b1c6ad95ae9182c0884c7" exitCode=137 Jan 26 17:08:35 crc kubenswrapper[4754]: I0126 17:08:35.354338 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6db5cbfc7c-nxhnr" event={"ID":"743bf5b2-5143-406b-b3d7-fa3a067974fd","Type":"ContainerDied","Data":"7887ee09373110d081d243bf9a15a5d13219c87c84d3b2b6840818f2a57e985a"} Jan 26 17:08:35 crc kubenswrapper[4754]: I0126 17:08:35.354363 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6db5cbfc7c-nxhnr" event={"ID":"743bf5b2-5143-406b-b3d7-fa3a067974fd","Type":"ContainerDied","Data":"2634e3c695e0a29fa8d18c8ece6dbbd81e5f69b92a0b1c6ad95ae9182c0884c7"} Jan 26 17:08:35 crc kubenswrapper[4754]: I0126 17:08:35.437279 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6db5cbfc7c-nxhnr" Jan 26 17:08:35 crc kubenswrapper[4754]: I0126 17:08:35.451769 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-585cbdbc67-qwfhd" Jan 26 17:08:35 crc kubenswrapper[4754]: I0126 17:08:35.470821 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-6c774f65c8-h55rj" podStartSLOduration=5.470800752 podStartE2EDuration="5.470800752s" podCreationTimestamp="2026-01-26 17:08:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 17:08:32.322144515 +0000 UTC m=+1278.846324979" watchObservedRunningTime="2026-01-26 17:08:35.470800752 +0000 UTC m=+1281.994981186" Jan 26 17:08:35 crc kubenswrapper[4754]: I0126 17:08:35.553973 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/743bf5b2-5143-406b-b3d7-fa3a067974fd-horizon-secret-key\") pod \"743bf5b2-5143-406b-b3d7-fa3a067974fd\" (UID: \"743bf5b2-5143-406b-b3d7-fa3a067974fd\") " Jan 26 17:08:35 crc kubenswrapper[4754]: I0126 17:08:35.554111 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/743bf5b2-5143-406b-b3d7-fa3a067974fd-scripts\") pod \"743bf5b2-5143-406b-b3d7-fa3a067974fd\" (UID: \"743bf5b2-5143-406b-b3d7-fa3a067974fd\") " Jan 26 17:08:35 crc kubenswrapper[4754]: I0126 17:08:35.554148 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/743bf5b2-5143-406b-b3d7-fa3a067974fd-logs\") pod \"743bf5b2-5143-406b-b3d7-fa3a067974fd\" (UID: \"743bf5b2-5143-406b-b3d7-fa3a067974fd\") " Jan 26 17:08:35 crc kubenswrapper[4754]: I0126 17:08:35.554169 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/743bf5b2-5143-406b-b3d7-fa3a067974fd-config-data\") pod \"743bf5b2-5143-406b-b3d7-fa3a067974fd\" (UID: \"743bf5b2-5143-406b-b3d7-fa3a067974fd\") " Jan 26 17:08:35 crc kubenswrapper[4754]: I0126 17:08:35.554196 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2tn9b\" (UniqueName: \"kubernetes.io/projected/743bf5b2-5143-406b-b3d7-fa3a067974fd-kube-api-access-2tn9b\") pod \"743bf5b2-5143-406b-b3d7-fa3a067974fd\" (UID: \"743bf5b2-5143-406b-b3d7-fa3a067974fd\") " Jan 26 17:08:35 crc kubenswrapper[4754]: I0126 17:08:35.554215 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c6spp\" (UniqueName: \"kubernetes.io/projected/eabead53-af6d-4405-b76b-a20c5a84f469-kube-api-access-c6spp\") pod \"eabead53-af6d-4405-b76b-a20c5a84f469\" (UID: \"eabead53-af6d-4405-b76b-a20c5a84f469\") " Jan 26 17:08:35 crc kubenswrapper[4754]: I0126 17:08:35.554261 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/eabead53-af6d-4405-b76b-a20c5a84f469-scripts\") pod \"eabead53-af6d-4405-b76b-a20c5a84f469\" (UID: \"eabead53-af6d-4405-b76b-a20c5a84f469\") " Jan 26 17:08:35 crc kubenswrapper[4754]: I0126 17:08:35.554307 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/eabead53-af6d-4405-b76b-a20c5a84f469-horizon-secret-key\") pod \"eabead53-af6d-4405-b76b-a20c5a84f469\" (UID: \"eabead53-af6d-4405-b76b-a20c5a84f469\") " Jan 26 17:08:35 crc kubenswrapper[4754]: I0126 17:08:35.554325 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/eabead53-af6d-4405-b76b-a20c5a84f469-config-data\") pod \"eabead53-af6d-4405-b76b-a20c5a84f469\" (UID: \"eabead53-af6d-4405-b76b-a20c5a84f469\") " Jan 26 17:08:35 crc kubenswrapper[4754]: I0126 17:08:35.554340 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/eabead53-af6d-4405-b76b-a20c5a84f469-logs\") pod \"eabead53-af6d-4405-b76b-a20c5a84f469\" (UID: \"eabead53-af6d-4405-b76b-a20c5a84f469\") " Jan 26 17:08:35 crc kubenswrapper[4754]: I0126 17:08:35.555120 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eabead53-af6d-4405-b76b-a20c5a84f469-logs" (OuterVolumeSpecName: "logs") pod "eabead53-af6d-4405-b76b-a20c5a84f469" (UID: "eabead53-af6d-4405-b76b-a20c5a84f469"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:08:35 crc kubenswrapper[4754]: I0126 17:08:35.557376 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/743bf5b2-5143-406b-b3d7-fa3a067974fd-logs" (OuterVolumeSpecName: "logs") pod "743bf5b2-5143-406b-b3d7-fa3a067974fd" (UID: "743bf5b2-5143-406b-b3d7-fa3a067974fd"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:08:35 crc kubenswrapper[4754]: I0126 17:08:35.563825 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/743bf5b2-5143-406b-b3d7-fa3a067974fd-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "743bf5b2-5143-406b-b3d7-fa3a067974fd" (UID: "743bf5b2-5143-406b-b3d7-fa3a067974fd"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:08:35 crc kubenswrapper[4754]: I0126 17:08:35.563825 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eabead53-af6d-4405-b76b-a20c5a84f469-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "eabead53-af6d-4405-b76b-a20c5a84f469" (UID: "eabead53-af6d-4405-b76b-a20c5a84f469"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:08:35 crc kubenswrapper[4754]: I0126 17:08:35.563922 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/743bf5b2-5143-406b-b3d7-fa3a067974fd-kube-api-access-2tn9b" (OuterVolumeSpecName: "kube-api-access-2tn9b") pod "743bf5b2-5143-406b-b3d7-fa3a067974fd" (UID: "743bf5b2-5143-406b-b3d7-fa3a067974fd"). InnerVolumeSpecName "kube-api-access-2tn9b". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:08:35 crc kubenswrapper[4754]: I0126 17:08:35.564924 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eabead53-af6d-4405-b76b-a20c5a84f469-kube-api-access-c6spp" (OuterVolumeSpecName: "kube-api-access-c6spp") pod "eabead53-af6d-4405-b76b-a20c5a84f469" (UID: "eabead53-af6d-4405-b76b-a20c5a84f469"). InnerVolumeSpecName "kube-api-access-c6spp". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:08:35 crc kubenswrapper[4754]: I0126 17:08:35.580406 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/743bf5b2-5143-406b-b3d7-fa3a067974fd-scripts" (OuterVolumeSpecName: "scripts") pod "743bf5b2-5143-406b-b3d7-fa3a067974fd" (UID: "743bf5b2-5143-406b-b3d7-fa3a067974fd"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:08:35 crc kubenswrapper[4754]: I0126 17:08:35.580526 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/743bf5b2-5143-406b-b3d7-fa3a067974fd-config-data" (OuterVolumeSpecName: "config-data") pod "743bf5b2-5143-406b-b3d7-fa3a067974fd" (UID: "743bf5b2-5143-406b-b3d7-fa3a067974fd"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:08:35 crc kubenswrapper[4754]: I0126 17:08:35.582305 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eabead53-af6d-4405-b76b-a20c5a84f469-config-data" (OuterVolumeSpecName: "config-data") pod "eabead53-af6d-4405-b76b-a20c5a84f469" (UID: "eabead53-af6d-4405-b76b-a20c5a84f469"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:08:35 crc kubenswrapper[4754]: I0126 17:08:35.583447 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eabead53-af6d-4405-b76b-a20c5a84f469-scripts" (OuterVolumeSpecName: "scripts") pod "eabead53-af6d-4405-b76b-a20c5a84f469" (UID: "eabead53-af6d-4405-b76b-a20c5a84f469"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:08:35 crc kubenswrapper[4754]: I0126 17:08:35.656586 4754 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/743bf5b2-5143-406b-b3d7-fa3a067974fd-logs\") on node \"crc\" DevicePath \"\"" Jan 26 17:08:35 crc kubenswrapper[4754]: I0126 17:08:35.656622 4754 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/743bf5b2-5143-406b-b3d7-fa3a067974fd-config-data\") on node \"crc\" DevicePath \"\"" Jan 26 17:08:35 crc kubenswrapper[4754]: I0126 17:08:35.656636 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2tn9b\" (UniqueName: \"kubernetes.io/projected/743bf5b2-5143-406b-b3d7-fa3a067974fd-kube-api-access-2tn9b\") on node \"crc\" DevicePath \"\"" Jan 26 17:08:35 crc kubenswrapper[4754]: I0126 17:08:35.656649 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c6spp\" (UniqueName: \"kubernetes.io/projected/eabead53-af6d-4405-b76b-a20c5a84f469-kube-api-access-c6spp\") on node \"crc\" DevicePath \"\"" Jan 26 17:08:35 crc kubenswrapper[4754]: I0126 17:08:35.656704 4754 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/eabead53-af6d-4405-b76b-a20c5a84f469-scripts\") on node \"crc\" DevicePath \"\"" Jan 26 17:08:35 crc kubenswrapper[4754]: I0126 17:08:35.656716 4754 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/eabead53-af6d-4405-b76b-a20c5a84f469-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Jan 26 17:08:35 crc kubenswrapper[4754]: I0126 17:08:35.656727 4754 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/eabead53-af6d-4405-b76b-a20c5a84f469-config-data\") on node \"crc\" DevicePath \"\"" Jan 26 17:08:35 crc kubenswrapper[4754]: I0126 17:08:35.656737 4754 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/eabead53-af6d-4405-b76b-a20c5a84f469-logs\") on node \"crc\" DevicePath \"\"" Jan 26 17:08:35 crc kubenswrapper[4754]: I0126 17:08:35.656746 4754 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/743bf5b2-5143-406b-b3d7-fa3a067974fd-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Jan 26 17:08:35 crc kubenswrapper[4754]: I0126 17:08:35.656758 4754 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/743bf5b2-5143-406b-b3d7-fa3a067974fd-scripts\") on node \"crc\" DevicePath \"\"" Jan 26 17:08:36 crc kubenswrapper[4754]: I0126 17:08:36.366444 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d84d5918-51f5-4131-9caf-9380004a0ca5","Type":"ContainerStarted","Data":"114035b4c3d582ed66db03f17a7b8f1e20a340ca4a189ef2996ccd2fcb60ce3b"} Jan 26 17:08:36 crc kubenswrapper[4754]: I0126 17:08:36.369488 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6db5cbfc7c-nxhnr" event={"ID":"743bf5b2-5143-406b-b3d7-fa3a067974fd","Type":"ContainerDied","Data":"19399a2650468f91411b2156cba66da58aa77b72b18d53d56e937132b8c7dec0"} Jan 26 17:08:36 crc kubenswrapper[4754]: I0126 17:08:36.369541 4754 scope.go:117] "RemoveContainer" containerID="7887ee09373110d081d243bf9a15a5d13219c87c84d3b2b6840818f2a57e985a" Jan 26 17:08:36 crc kubenswrapper[4754]: I0126 17:08:36.369548 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6db5cbfc7c-nxhnr" Jan 26 17:08:36 crc kubenswrapper[4754]: I0126 17:08:36.371808 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-585cbdbc67-qwfhd" event={"ID":"eabead53-af6d-4405-b76b-a20c5a84f469","Type":"ContainerDied","Data":"732ce722519d0100e33e34175dbac50759bc15630f8b052b15e80f9ca50aa5ad"} Jan 26 17:08:36 crc kubenswrapper[4754]: I0126 17:08:36.371918 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-585cbdbc67-qwfhd" Jan 26 17:08:36 crc kubenswrapper[4754]: I0126 17:08:36.395549 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-585cbdbc67-qwfhd"] Jan 26 17:08:36 crc kubenswrapper[4754]: I0126 17:08:36.403026 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-585cbdbc67-qwfhd"] Jan 26 17:08:36 crc kubenswrapper[4754]: I0126 17:08:36.422533 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-6db5cbfc7c-nxhnr"] Jan 26 17:08:36 crc kubenswrapper[4754]: I0126 17:08:36.433144 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-6db5cbfc7c-nxhnr"] Jan 26 17:08:36 crc kubenswrapper[4754]: I0126 17:08:36.551741 4754 scope.go:117] "RemoveContainer" containerID="2634e3c695e0a29fa8d18c8ece6dbbd81e5f69b92a0b1c6ad95ae9182c0884c7" Jan 26 17:08:36 crc kubenswrapper[4754]: I0126 17:08:36.572522 4754 scope.go:117] "RemoveContainer" containerID="04100d29f24966812f52c17d77db99b3275e2115f5be21feb58b8a03dbfbd039" Jan 26 17:08:36 crc kubenswrapper[4754]: I0126 17:08:36.746597 4754 scope.go:117] "RemoveContainer" containerID="8e092eb22ebc9cba28e6b243f38f358bf2cca40a97352fd16fe630e746bd0f10" Jan 26 17:08:37 crc kubenswrapper[4754]: I0126 17:08:37.079774 4754 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-85f95cb57b-99z67" podUID="ea911a5c-f353-4da9-9900-bfcbb234a377" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.154:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.154:8443: connect: connection refused" Jan 26 17:08:37 crc kubenswrapper[4754]: I0126 17:08:37.385117 4754 generic.go:334] "Generic (PLEG): container finished" podID="01246f69-d6bc-4f70-a50f-4d7c8e4a7620" containerID="fa8641d204e41cbb652b48277e0e1d4c609dfb5b9fb94837cbf80deac687370c" exitCode=0 Jan 26 17:08:37 crc kubenswrapper[4754]: I0126 17:08:37.385191 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-whnhn" event={"ID":"01246f69-d6bc-4f70-a50f-4d7c8e4a7620","Type":"ContainerDied","Data":"fa8641d204e41cbb652b48277e0e1d4c609dfb5b9fb94837cbf80deac687370c"} Jan 26 17:08:37 crc kubenswrapper[4754]: I0126 17:08:37.777894 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="743bf5b2-5143-406b-b3d7-fa3a067974fd" path="/var/lib/kubelet/pods/743bf5b2-5143-406b-b3d7-fa3a067974fd/volumes" Jan 26 17:08:37 crc kubenswrapper[4754]: I0126 17:08:37.778919 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eabead53-af6d-4405-b76b-a20c5a84f469" path="/var/lib/kubelet/pods/eabead53-af6d-4405-b76b-a20c5a84f469/volumes" Jan 26 17:08:38 crc kubenswrapper[4754]: I0126 17:08:38.725109 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-whnhn" Jan 26 17:08:38 crc kubenswrapper[4754]: I0126 17:08:38.807476 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01246f69-d6bc-4f70-a50f-4d7c8e4a7620-combined-ca-bundle\") pod \"01246f69-d6bc-4f70-a50f-4d7c8e4a7620\" (UID: \"01246f69-d6bc-4f70-a50f-4d7c8e4a7620\") " Jan 26 17:08:38 crc kubenswrapper[4754]: I0126 17:08:38.807534 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/01246f69-d6bc-4f70-a50f-4d7c8e4a7620-logs\") pod \"01246f69-d6bc-4f70-a50f-4d7c8e4a7620\" (UID: \"01246f69-d6bc-4f70-a50f-4d7c8e4a7620\") " Jan 26 17:08:38 crc kubenswrapper[4754]: I0126 17:08:38.807683 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/01246f69-d6bc-4f70-a50f-4d7c8e4a7620-config-data\") pod \"01246f69-d6bc-4f70-a50f-4d7c8e4a7620\" (UID: \"01246f69-d6bc-4f70-a50f-4d7c8e4a7620\") " Jan 26 17:08:38 crc kubenswrapper[4754]: I0126 17:08:38.807775 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/01246f69-d6bc-4f70-a50f-4d7c8e4a7620-scripts\") pod \"01246f69-d6bc-4f70-a50f-4d7c8e4a7620\" (UID: \"01246f69-d6bc-4f70-a50f-4d7c8e4a7620\") " Jan 26 17:08:38 crc kubenswrapper[4754]: I0126 17:08:38.807823 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rr5pm\" (UniqueName: \"kubernetes.io/projected/01246f69-d6bc-4f70-a50f-4d7c8e4a7620-kube-api-access-rr5pm\") pod \"01246f69-d6bc-4f70-a50f-4d7c8e4a7620\" (UID: \"01246f69-d6bc-4f70-a50f-4d7c8e4a7620\") " Jan 26 17:08:38 crc kubenswrapper[4754]: I0126 17:08:38.810271 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/01246f69-d6bc-4f70-a50f-4d7c8e4a7620-logs" (OuterVolumeSpecName: "logs") pod "01246f69-d6bc-4f70-a50f-4d7c8e4a7620" (UID: "01246f69-d6bc-4f70-a50f-4d7c8e4a7620"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:08:38 crc kubenswrapper[4754]: I0126 17:08:38.828704 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01246f69-d6bc-4f70-a50f-4d7c8e4a7620-scripts" (OuterVolumeSpecName: "scripts") pod "01246f69-d6bc-4f70-a50f-4d7c8e4a7620" (UID: "01246f69-d6bc-4f70-a50f-4d7c8e4a7620"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:08:38 crc kubenswrapper[4754]: I0126 17:08:38.829218 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01246f69-d6bc-4f70-a50f-4d7c8e4a7620-kube-api-access-rr5pm" (OuterVolumeSpecName: "kube-api-access-rr5pm") pod "01246f69-d6bc-4f70-a50f-4d7c8e4a7620" (UID: "01246f69-d6bc-4f70-a50f-4d7c8e4a7620"). InnerVolumeSpecName "kube-api-access-rr5pm". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:08:38 crc kubenswrapper[4754]: I0126 17:08:38.837895 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01246f69-d6bc-4f70-a50f-4d7c8e4a7620-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "01246f69-d6bc-4f70-a50f-4d7c8e4a7620" (UID: "01246f69-d6bc-4f70-a50f-4d7c8e4a7620"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:08:38 crc kubenswrapper[4754]: I0126 17:08:38.842226 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01246f69-d6bc-4f70-a50f-4d7c8e4a7620-config-data" (OuterVolumeSpecName: "config-data") pod "01246f69-d6bc-4f70-a50f-4d7c8e4a7620" (UID: "01246f69-d6bc-4f70-a50f-4d7c8e4a7620"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:08:38 crc kubenswrapper[4754]: I0126 17:08:38.909828 4754 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/01246f69-d6bc-4f70-a50f-4d7c8e4a7620-config-data\") on node \"crc\" DevicePath \"\"" Jan 26 17:08:38 crc kubenswrapper[4754]: I0126 17:08:38.909871 4754 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/01246f69-d6bc-4f70-a50f-4d7c8e4a7620-scripts\") on node \"crc\" DevicePath \"\"" Jan 26 17:08:38 crc kubenswrapper[4754]: I0126 17:08:38.909884 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rr5pm\" (UniqueName: \"kubernetes.io/projected/01246f69-d6bc-4f70-a50f-4d7c8e4a7620-kube-api-access-rr5pm\") on node \"crc\" DevicePath \"\"" Jan 26 17:08:38 crc kubenswrapper[4754]: I0126 17:08:38.909897 4754 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01246f69-d6bc-4f70-a50f-4d7c8e4a7620-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 26 17:08:38 crc kubenswrapper[4754]: I0126 17:08:38.909910 4754 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/01246f69-d6bc-4f70-a50f-4d7c8e4a7620-logs\") on node \"crc\" DevicePath \"\"" Jan 26 17:08:39 crc kubenswrapper[4754]: I0126 17:08:39.405694 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-whnhn" event={"ID":"01246f69-d6bc-4f70-a50f-4d7c8e4a7620","Type":"ContainerDied","Data":"f3ffb4c4669fc103b7394b3bffeca790d3f2c6ffc4578b08a6b116b2fb7efe2c"} Jan 26 17:08:39 crc kubenswrapper[4754]: I0126 17:08:39.406038 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f3ffb4c4669fc103b7394b3bffeca790d3f2c6ffc4578b08a6b116b2fb7efe2c" Jan 26 17:08:39 crc kubenswrapper[4754]: I0126 17:08:39.405723 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-whnhn" Jan 26 17:08:39 crc kubenswrapper[4754]: I0126 17:08:39.499440 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-577b5f64f6-qwc6n"] Jan 26 17:08:39 crc kubenswrapper[4754]: E0126 17:08:39.499876 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eabead53-af6d-4405-b76b-a20c5a84f469" containerName="horizon-log" Jan 26 17:08:39 crc kubenswrapper[4754]: I0126 17:08:39.499892 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="eabead53-af6d-4405-b76b-a20c5a84f469" containerName="horizon-log" Jan 26 17:08:39 crc kubenswrapper[4754]: E0126 17:08:39.499913 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="743bf5b2-5143-406b-b3d7-fa3a067974fd" containerName="horizon-log" Jan 26 17:08:39 crc kubenswrapper[4754]: I0126 17:08:39.499920 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="743bf5b2-5143-406b-b3d7-fa3a067974fd" containerName="horizon-log" Jan 26 17:08:39 crc kubenswrapper[4754]: E0126 17:08:39.499934 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eabead53-af6d-4405-b76b-a20c5a84f469" containerName="horizon" Jan 26 17:08:39 crc kubenswrapper[4754]: I0126 17:08:39.499941 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="eabead53-af6d-4405-b76b-a20c5a84f469" containerName="horizon" Jan 26 17:08:39 crc kubenswrapper[4754]: E0126 17:08:39.499962 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="01246f69-d6bc-4f70-a50f-4d7c8e4a7620" containerName="placement-db-sync" Jan 26 17:08:39 crc kubenswrapper[4754]: I0126 17:08:39.499969 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="01246f69-d6bc-4f70-a50f-4d7c8e4a7620" containerName="placement-db-sync" Jan 26 17:08:39 crc kubenswrapper[4754]: E0126 17:08:39.499980 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="743bf5b2-5143-406b-b3d7-fa3a067974fd" containerName="horizon" Jan 26 17:08:39 crc kubenswrapper[4754]: I0126 17:08:39.499986 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="743bf5b2-5143-406b-b3d7-fa3a067974fd" containerName="horizon" Jan 26 17:08:39 crc kubenswrapper[4754]: I0126 17:08:39.500209 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="eabead53-af6d-4405-b76b-a20c5a84f469" containerName="horizon" Jan 26 17:08:39 crc kubenswrapper[4754]: I0126 17:08:39.500231 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="743bf5b2-5143-406b-b3d7-fa3a067974fd" containerName="horizon-log" Jan 26 17:08:39 crc kubenswrapper[4754]: I0126 17:08:39.500253 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="743bf5b2-5143-406b-b3d7-fa3a067974fd" containerName="horizon" Jan 26 17:08:39 crc kubenswrapper[4754]: I0126 17:08:39.500276 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="eabead53-af6d-4405-b76b-a20c5a84f469" containerName="horizon-log" Jan 26 17:08:39 crc kubenswrapper[4754]: I0126 17:08:39.500292 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="01246f69-d6bc-4f70-a50f-4d7c8e4a7620" containerName="placement-db-sync" Jan 26 17:08:39 crc kubenswrapper[4754]: I0126 17:08:39.502253 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-577b5f64f6-qwc6n" Jan 26 17:08:39 crc kubenswrapper[4754]: I0126 17:08:39.505149 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Jan 26 17:08:39 crc kubenswrapper[4754]: I0126 17:08:39.505216 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Jan 26 17:08:39 crc kubenswrapper[4754]: I0126 17:08:39.506244 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-92svj" Jan 26 17:08:39 crc kubenswrapper[4754]: I0126 17:08:39.509741 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Jan 26 17:08:39 crc kubenswrapper[4754]: I0126 17:08:39.509978 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Jan 26 17:08:39 crc kubenswrapper[4754]: I0126 17:08:39.531047 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-577b5f64f6-qwc6n"] Jan 26 17:08:39 crc kubenswrapper[4754]: I0126 17:08:39.626524 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/854736cc-7859-4621-865c-69cab5f7dbeb-internal-tls-certs\") pod \"placement-577b5f64f6-qwc6n\" (UID: \"854736cc-7859-4621-865c-69cab5f7dbeb\") " pod="openstack/placement-577b5f64f6-qwc6n" Jan 26 17:08:39 crc kubenswrapper[4754]: I0126 17:08:39.626792 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/854736cc-7859-4621-865c-69cab5f7dbeb-public-tls-certs\") pod \"placement-577b5f64f6-qwc6n\" (UID: \"854736cc-7859-4621-865c-69cab5f7dbeb\") " pod="openstack/placement-577b5f64f6-qwc6n" Jan 26 17:08:39 crc kubenswrapper[4754]: I0126 17:08:39.626879 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/854736cc-7859-4621-865c-69cab5f7dbeb-config-data\") pod \"placement-577b5f64f6-qwc6n\" (UID: \"854736cc-7859-4621-865c-69cab5f7dbeb\") " pod="openstack/placement-577b5f64f6-qwc6n" Jan 26 17:08:39 crc kubenswrapper[4754]: I0126 17:08:39.626996 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/854736cc-7859-4621-865c-69cab5f7dbeb-scripts\") pod \"placement-577b5f64f6-qwc6n\" (UID: \"854736cc-7859-4621-865c-69cab5f7dbeb\") " pod="openstack/placement-577b5f64f6-qwc6n" Jan 26 17:08:39 crc kubenswrapper[4754]: I0126 17:08:39.627079 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/854736cc-7859-4621-865c-69cab5f7dbeb-combined-ca-bundle\") pod \"placement-577b5f64f6-qwc6n\" (UID: \"854736cc-7859-4621-865c-69cab5f7dbeb\") " pod="openstack/placement-577b5f64f6-qwc6n" Jan 26 17:08:39 crc kubenswrapper[4754]: I0126 17:08:39.627334 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/854736cc-7859-4621-865c-69cab5f7dbeb-logs\") pod \"placement-577b5f64f6-qwc6n\" (UID: \"854736cc-7859-4621-865c-69cab5f7dbeb\") " pod="openstack/placement-577b5f64f6-qwc6n" Jan 26 17:08:39 crc kubenswrapper[4754]: I0126 17:08:39.627449 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7bbjs\" (UniqueName: \"kubernetes.io/projected/854736cc-7859-4621-865c-69cab5f7dbeb-kube-api-access-7bbjs\") pod \"placement-577b5f64f6-qwc6n\" (UID: \"854736cc-7859-4621-865c-69cab5f7dbeb\") " pod="openstack/placement-577b5f64f6-qwc6n" Jan 26 17:08:39 crc kubenswrapper[4754]: I0126 17:08:39.729694 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/854736cc-7859-4621-865c-69cab5f7dbeb-scripts\") pod \"placement-577b5f64f6-qwc6n\" (UID: \"854736cc-7859-4621-865c-69cab5f7dbeb\") " pod="openstack/placement-577b5f64f6-qwc6n" Jan 26 17:08:39 crc kubenswrapper[4754]: I0126 17:08:39.729792 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/854736cc-7859-4621-865c-69cab5f7dbeb-combined-ca-bundle\") pod \"placement-577b5f64f6-qwc6n\" (UID: \"854736cc-7859-4621-865c-69cab5f7dbeb\") " pod="openstack/placement-577b5f64f6-qwc6n" Jan 26 17:08:39 crc kubenswrapper[4754]: I0126 17:08:39.729820 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/854736cc-7859-4621-865c-69cab5f7dbeb-logs\") pod \"placement-577b5f64f6-qwc6n\" (UID: \"854736cc-7859-4621-865c-69cab5f7dbeb\") " pod="openstack/placement-577b5f64f6-qwc6n" Jan 26 17:08:39 crc kubenswrapper[4754]: I0126 17:08:39.730372 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/854736cc-7859-4621-865c-69cab5f7dbeb-logs\") pod \"placement-577b5f64f6-qwc6n\" (UID: \"854736cc-7859-4621-865c-69cab5f7dbeb\") " pod="openstack/placement-577b5f64f6-qwc6n" Jan 26 17:08:39 crc kubenswrapper[4754]: I0126 17:08:39.730558 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7bbjs\" (UniqueName: \"kubernetes.io/projected/854736cc-7859-4621-865c-69cab5f7dbeb-kube-api-access-7bbjs\") pod \"placement-577b5f64f6-qwc6n\" (UID: \"854736cc-7859-4621-865c-69cab5f7dbeb\") " pod="openstack/placement-577b5f64f6-qwc6n" Jan 26 17:08:39 crc kubenswrapper[4754]: I0126 17:08:39.730698 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/854736cc-7859-4621-865c-69cab5f7dbeb-internal-tls-certs\") pod \"placement-577b5f64f6-qwc6n\" (UID: \"854736cc-7859-4621-865c-69cab5f7dbeb\") " pod="openstack/placement-577b5f64f6-qwc6n" Jan 26 17:08:39 crc kubenswrapper[4754]: I0126 17:08:39.730721 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/854736cc-7859-4621-865c-69cab5f7dbeb-public-tls-certs\") pod \"placement-577b5f64f6-qwc6n\" (UID: \"854736cc-7859-4621-865c-69cab5f7dbeb\") " pod="openstack/placement-577b5f64f6-qwc6n" Jan 26 17:08:39 crc kubenswrapper[4754]: I0126 17:08:39.730750 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/854736cc-7859-4621-865c-69cab5f7dbeb-config-data\") pod \"placement-577b5f64f6-qwc6n\" (UID: \"854736cc-7859-4621-865c-69cab5f7dbeb\") " pod="openstack/placement-577b5f64f6-qwc6n" Jan 26 17:08:39 crc kubenswrapper[4754]: I0126 17:08:39.734387 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/854736cc-7859-4621-865c-69cab5f7dbeb-scripts\") pod \"placement-577b5f64f6-qwc6n\" (UID: \"854736cc-7859-4621-865c-69cab5f7dbeb\") " pod="openstack/placement-577b5f64f6-qwc6n" Jan 26 17:08:39 crc kubenswrapper[4754]: I0126 17:08:39.734409 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/854736cc-7859-4621-865c-69cab5f7dbeb-internal-tls-certs\") pod \"placement-577b5f64f6-qwc6n\" (UID: \"854736cc-7859-4621-865c-69cab5f7dbeb\") " pod="openstack/placement-577b5f64f6-qwc6n" Jan 26 17:08:39 crc kubenswrapper[4754]: I0126 17:08:39.735307 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/854736cc-7859-4621-865c-69cab5f7dbeb-public-tls-certs\") pod \"placement-577b5f64f6-qwc6n\" (UID: \"854736cc-7859-4621-865c-69cab5f7dbeb\") " pod="openstack/placement-577b5f64f6-qwc6n" Jan 26 17:08:39 crc kubenswrapper[4754]: I0126 17:08:39.735439 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/854736cc-7859-4621-865c-69cab5f7dbeb-config-data\") pod \"placement-577b5f64f6-qwc6n\" (UID: \"854736cc-7859-4621-865c-69cab5f7dbeb\") " pod="openstack/placement-577b5f64f6-qwc6n" Jan 26 17:08:39 crc kubenswrapper[4754]: I0126 17:08:39.737691 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/854736cc-7859-4621-865c-69cab5f7dbeb-combined-ca-bundle\") pod \"placement-577b5f64f6-qwc6n\" (UID: \"854736cc-7859-4621-865c-69cab5f7dbeb\") " pod="openstack/placement-577b5f64f6-qwc6n" Jan 26 17:08:39 crc kubenswrapper[4754]: I0126 17:08:39.748072 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7bbjs\" (UniqueName: \"kubernetes.io/projected/854736cc-7859-4621-865c-69cab5f7dbeb-kube-api-access-7bbjs\") pod \"placement-577b5f64f6-qwc6n\" (UID: \"854736cc-7859-4621-865c-69cab5f7dbeb\") " pod="openstack/placement-577b5f64f6-qwc6n" Jan 26 17:08:39 crc kubenswrapper[4754]: I0126 17:08:39.825346 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-577b5f64f6-qwc6n" Jan 26 17:08:40 crc kubenswrapper[4754]: I0126 17:08:40.335981 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-577b5f64f6-qwc6n"] Jan 26 17:08:40 crc kubenswrapper[4754]: I0126 17:08:40.423950 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-577b5f64f6-qwc6n" event={"ID":"854736cc-7859-4621-865c-69cab5f7dbeb","Type":"ContainerStarted","Data":"d282f28381c10fc0d60592142a61d34462633351958541225f0b1213d5c203fb"} Jan 26 17:08:41 crc kubenswrapper[4754]: I0126 17:08:41.435112 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-577b5f64f6-qwc6n" event={"ID":"854736cc-7859-4621-865c-69cab5f7dbeb","Type":"ContainerStarted","Data":"d171fcf5bb3fc684f0ae798d93818d58c41b6fdc199c265563571c4736f984e7"} Jan 26 17:08:41 crc kubenswrapper[4754]: I0126 17:08:41.435422 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-577b5f64f6-qwc6n" event={"ID":"854736cc-7859-4621-865c-69cab5f7dbeb","Type":"ContainerStarted","Data":"e71deadee17691f4c69b962159d0f926a5a600e535aca9c59a0b28f5fe3011f6"} Jan 26 17:08:41 crc kubenswrapper[4754]: I0126 17:08:41.435437 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-577b5f64f6-qwc6n" Jan 26 17:08:41 crc kubenswrapper[4754]: I0126 17:08:41.435460 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-577b5f64f6-qwc6n" Jan 26 17:08:41 crc kubenswrapper[4754]: I0126 17:08:41.456822 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-577b5f64f6-qwc6n" podStartSLOduration=2.45679903 podStartE2EDuration="2.45679903s" podCreationTimestamp="2026-01-26 17:08:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 17:08:41.451981264 +0000 UTC m=+1287.976161728" watchObservedRunningTime="2026-01-26 17:08:41.45679903 +0000 UTC m=+1287.980979464" Jan 26 17:08:47 crc kubenswrapper[4754]: I0126 17:08:47.079465 4754 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-85f95cb57b-99z67" podUID="ea911a5c-f353-4da9-9900-bfcbb234a377" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.154:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.154:8443: connect: connection refused" Jan 26 17:08:51 crc kubenswrapper[4754]: I0126 17:08:51.512809 4754 generic.go:334] "Generic (PLEG): container finished" podID="ea094395-bb3c-4464-829f-313ab6b9e14c" containerID="a0aac0c2b2fa1e7140caf64c6f53502aa2cdbab58168b149446176ac7c4866bb" exitCode=0 Jan 26 17:08:51 crc kubenswrapper[4754]: I0126 17:08:51.512892 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-9spps" event={"ID":"ea094395-bb3c-4464-829f-313ab6b9e14c","Type":"ContainerDied","Data":"a0aac0c2b2fa1e7140caf64c6f53502aa2cdbab58168b149446176ac7c4866bb"} Jan 26 17:08:54 crc kubenswrapper[4754]: I0126 17:08:54.050986 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-9spps" Jan 26 17:08:54 crc kubenswrapper[4754]: I0126 17:08:54.253455 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ea094395-bb3c-4464-829f-313ab6b9e14c-db-sync-config-data\") pod \"ea094395-bb3c-4464-829f-313ab6b9e14c\" (UID: \"ea094395-bb3c-4464-829f-313ab6b9e14c\") " Jan 26 17:08:54 crc kubenswrapper[4754]: I0126 17:08:54.253663 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wk6gz\" (UniqueName: \"kubernetes.io/projected/ea094395-bb3c-4464-829f-313ab6b9e14c-kube-api-access-wk6gz\") pod \"ea094395-bb3c-4464-829f-313ab6b9e14c\" (UID: \"ea094395-bb3c-4464-829f-313ab6b9e14c\") " Jan 26 17:08:54 crc kubenswrapper[4754]: I0126 17:08:54.253735 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ea094395-bb3c-4464-829f-313ab6b9e14c-combined-ca-bundle\") pod \"ea094395-bb3c-4464-829f-313ab6b9e14c\" (UID: \"ea094395-bb3c-4464-829f-313ab6b9e14c\") " Jan 26 17:08:54 crc kubenswrapper[4754]: I0126 17:08:54.259164 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ea094395-bb3c-4464-829f-313ab6b9e14c-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "ea094395-bb3c-4464-829f-313ab6b9e14c" (UID: "ea094395-bb3c-4464-829f-313ab6b9e14c"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:08:54 crc kubenswrapper[4754]: I0126 17:08:54.259233 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ea094395-bb3c-4464-829f-313ab6b9e14c-kube-api-access-wk6gz" (OuterVolumeSpecName: "kube-api-access-wk6gz") pod "ea094395-bb3c-4464-829f-313ab6b9e14c" (UID: "ea094395-bb3c-4464-829f-313ab6b9e14c"). InnerVolumeSpecName "kube-api-access-wk6gz". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:08:54 crc kubenswrapper[4754]: I0126 17:08:54.278784 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ea094395-bb3c-4464-829f-313ab6b9e14c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ea094395-bb3c-4464-829f-313ab6b9e14c" (UID: "ea094395-bb3c-4464-829f-313ab6b9e14c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:08:54 crc kubenswrapper[4754]: I0126 17:08:54.355492 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wk6gz\" (UniqueName: \"kubernetes.io/projected/ea094395-bb3c-4464-829f-313ab6b9e14c-kube-api-access-wk6gz\") on node \"crc\" DevicePath \"\"" Jan 26 17:08:54 crc kubenswrapper[4754]: I0126 17:08:54.355523 4754 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ea094395-bb3c-4464-829f-313ab6b9e14c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 26 17:08:54 crc kubenswrapper[4754]: I0126 17:08:54.355532 4754 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ea094395-bb3c-4464-829f-313ab6b9e14c-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Jan 26 17:08:54 crc kubenswrapper[4754]: I0126 17:08:54.545942 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-9spps" event={"ID":"ea094395-bb3c-4464-829f-313ab6b9e14c","Type":"ContainerDied","Data":"ede0b1d99871904079a1c996a8496a9574c987d4380c1fe223865c6fc4de3cad"} Jan 26 17:08:54 crc kubenswrapper[4754]: I0126 17:08:54.545987 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ede0b1d99871904079a1c996a8496a9574c987d4380c1fe223865c6fc4de3cad" Jan 26 17:08:54 crc kubenswrapper[4754]: I0126 17:08:54.546046 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-9spps" Jan 26 17:08:55 crc kubenswrapper[4754]: I0126 17:08:55.448366 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-5677764859-tswcr"] Jan 26 17:08:55 crc kubenswrapper[4754]: E0126 17:08:55.448800 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea094395-bb3c-4464-829f-313ab6b9e14c" containerName="barbican-db-sync" Jan 26 17:08:55 crc kubenswrapper[4754]: I0126 17:08:55.448816 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea094395-bb3c-4464-829f-313ab6b9e14c" containerName="barbican-db-sync" Jan 26 17:08:55 crc kubenswrapper[4754]: I0126 17:08:55.448999 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="ea094395-bb3c-4464-829f-313ab6b9e14c" containerName="barbican-db-sync" Jan 26 17:08:55 crc kubenswrapper[4754]: I0126 17:08:55.449972 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-5677764859-tswcr" Jan 26 17:08:55 crc kubenswrapper[4754]: I0126 17:08:55.454389 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Jan 26 17:08:55 crc kubenswrapper[4754]: I0126 17:08:55.454627 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Jan 26 17:08:55 crc kubenswrapper[4754]: I0126 17:08:55.454966 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-sfxbx" Jan 26 17:08:55 crc kubenswrapper[4754]: I0126 17:08:55.476058 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d64625ff-6586-4c80-a720-b2febd49a966-config-data-custom\") pod \"barbican-worker-5677764859-tswcr\" (UID: \"d64625ff-6586-4c80-a720-b2febd49a966\") " pod="openstack/barbican-worker-5677764859-tswcr" Jan 26 17:08:55 crc kubenswrapper[4754]: I0126 17:08:55.476178 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d64625ff-6586-4c80-a720-b2febd49a966-logs\") pod \"barbican-worker-5677764859-tswcr\" (UID: \"d64625ff-6586-4c80-a720-b2febd49a966\") " pod="openstack/barbican-worker-5677764859-tswcr" Jan 26 17:08:55 crc kubenswrapper[4754]: I0126 17:08:55.476326 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d64625ff-6586-4c80-a720-b2febd49a966-combined-ca-bundle\") pod \"barbican-worker-5677764859-tswcr\" (UID: \"d64625ff-6586-4c80-a720-b2febd49a966\") " pod="openstack/barbican-worker-5677764859-tswcr" Jan 26 17:08:55 crc kubenswrapper[4754]: I0126 17:08:55.476389 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d64625ff-6586-4c80-a720-b2febd49a966-config-data\") pod \"barbican-worker-5677764859-tswcr\" (UID: \"d64625ff-6586-4c80-a720-b2febd49a966\") " pod="openstack/barbican-worker-5677764859-tswcr" Jan 26 17:08:55 crc kubenswrapper[4754]: I0126 17:08:55.476532 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gktjq\" (UniqueName: \"kubernetes.io/projected/d64625ff-6586-4c80-a720-b2febd49a966-kube-api-access-gktjq\") pod \"barbican-worker-5677764859-tswcr\" (UID: \"d64625ff-6586-4c80-a720-b2febd49a966\") " pod="openstack/barbican-worker-5677764859-tswcr" Jan 26 17:08:55 crc kubenswrapper[4754]: I0126 17:08:55.484748 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-5677764859-tswcr"] Jan 26 17:08:55 crc kubenswrapper[4754]: I0126 17:08:55.528737 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-586bdc5f9-lqn4x"] Jan 26 17:08:55 crc kubenswrapper[4754]: I0126 17:08:55.530783 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-586bdc5f9-lqn4x" Jan 26 17:08:55 crc kubenswrapper[4754]: I0126 17:08:55.536762 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-99649c64-sldzz"] Jan 26 17:08:55 crc kubenswrapper[4754]: I0126 17:08:55.538442 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-99649c64-sldzz" Jan 26 17:08:55 crc kubenswrapper[4754]: I0126 17:08:55.542283 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Jan 26 17:08:55 crc kubenswrapper[4754]: I0126 17:08:55.598328 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d64625ff-6586-4c80-a720-b2febd49a966-combined-ca-bundle\") pod \"barbican-worker-5677764859-tswcr\" (UID: \"d64625ff-6586-4c80-a720-b2febd49a966\") " pod="openstack/barbican-worker-5677764859-tswcr" Jan 26 17:08:55 crc kubenswrapper[4754]: I0126 17:08:55.598416 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d64625ff-6586-4c80-a720-b2febd49a966-config-data\") pod \"barbican-worker-5677764859-tswcr\" (UID: \"d64625ff-6586-4c80-a720-b2febd49a966\") " pod="openstack/barbican-worker-5677764859-tswcr" Jan 26 17:08:55 crc kubenswrapper[4754]: I0126 17:08:55.598491 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gktjq\" (UniqueName: \"kubernetes.io/projected/d64625ff-6586-4c80-a720-b2febd49a966-kube-api-access-gktjq\") pod \"barbican-worker-5677764859-tswcr\" (UID: \"d64625ff-6586-4c80-a720-b2febd49a966\") " pod="openstack/barbican-worker-5677764859-tswcr" Jan 26 17:08:55 crc kubenswrapper[4754]: I0126 17:08:55.598574 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d64625ff-6586-4c80-a720-b2febd49a966-config-data-custom\") pod \"barbican-worker-5677764859-tswcr\" (UID: \"d64625ff-6586-4c80-a720-b2febd49a966\") " pod="openstack/barbican-worker-5677764859-tswcr" Jan 26 17:08:55 crc kubenswrapper[4754]: I0126 17:08:55.598670 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d64625ff-6586-4c80-a720-b2febd49a966-logs\") pod \"barbican-worker-5677764859-tswcr\" (UID: \"d64625ff-6586-4c80-a720-b2febd49a966\") " pod="openstack/barbican-worker-5677764859-tswcr" Jan 26 17:08:55 crc kubenswrapper[4754]: I0126 17:08:55.600161 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d64625ff-6586-4c80-a720-b2febd49a966-logs\") pod \"barbican-worker-5677764859-tswcr\" (UID: \"d64625ff-6586-4c80-a720-b2febd49a966\") " pod="openstack/barbican-worker-5677764859-tswcr" Jan 26 17:08:55 crc kubenswrapper[4754]: I0126 17:08:55.610793 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-99649c64-sldzz"] Jan 26 17:08:55 crc kubenswrapper[4754]: I0126 17:08:55.641828 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d64625ff-6586-4c80-a720-b2febd49a966-config-data\") pod \"barbican-worker-5677764859-tswcr\" (UID: \"d64625ff-6586-4c80-a720-b2febd49a966\") " pod="openstack/barbican-worker-5677764859-tswcr" Jan 26 17:08:55 crc kubenswrapper[4754]: I0126 17:08:55.648439 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d64625ff-6586-4c80-a720-b2febd49a966-combined-ca-bundle\") pod \"barbican-worker-5677764859-tswcr\" (UID: \"d64625ff-6586-4c80-a720-b2febd49a966\") " pod="openstack/barbican-worker-5677764859-tswcr" Jan 26 17:08:55 crc kubenswrapper[4754]: I0126 17:08:55.688193 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gktjq\" (UniqueName: \"kubernetes.io/projected/d64625ff-6586-4c80-a720-b2febd49a966-kube-api-access-gktjq\") pod \"barbican-worker-5677764859-tswcr\" (UID: \"d64625ff-6586-4c80-a720-b2febd49a966\") " pod="openstack/barbican-worker-5677764859-tswcr" Jan 26 17:08:55 crc kubenswrapper[4754]: I0126 17:08:55.689747 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d64625ff-6586-4c80-a720-b2febd49a966-config-data-custom\") pod \"barbican-worker-5677764859-tswcr\" (UID: \"d64625ff-6586-4c80-a720-b2febd49a966\") " pod="openstack/barbican-worker-5677764859-tswcr" Jan 26 17:08:55 crc kubenswrapper[4754]: I0126 17:08:55.706996 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/057f7e48-aac2-407d-9fd4-555b5f8c6dad-ovsdbserver-sb\") pod \"dnsmasq-dns-586bdc5f9-lqn4x\" (UID: \"057f7e48-aac2-407d-9fd4-555b5f8c6dad\") " pod="openstack/dnsmasq-dns-586bdc5f9-lqn4x" Jan 26 17:08:55 crc kubenswrapper[4754]: I0126 17:08:55.707047 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/057f7e48-aac2-407d-9fd4-555b5f8c6dad-config\") pod \"dnsmasq-dns-586bdc5f9-lqn4x\" (UID: \"057f7e48-aac2-407d-9fd4-555b5f8c6dad\") " pod="openstack/dnsmasq-dns-586bdc5f9-lqn4x" Jan 26 17:08:55 crc kubenswrapper[4754]: I0126 17:08:55.707073 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jh76b\" (UniqueName: \"kubernetes.io/projected/057f7e48-aac2-407d-9fd4-555b5f8c6dad-kube-api-access-jh76b\") pod \"dnsmasq-dns-586bdc5f9-lqn4x\" (UID: \"057f7e48-aac2-407d-9fd4-555b5f8c6dad\") " pod="openstack/dnsmasq-dns-586bdc5f9-lqn4x" Jan 26 17:08:55 crc kubenswrapper[4754]: I0126 17:08:55.707121 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0b004ba0-01f6-49f9-8d8d-56033dcd3533-config-data-custom\") pod \"barbican-keystone-listener-99649c64-sldzz\" (UID: \"0b004ba0-01f6-49f9-8d8d-56033dcd3533\") " pod="openstack/barbican-keystone-listener-99649c64-sldzz" Jan 26 17:08:55 crc kubenswrapper[4754]: I0126 17:08:55.707139 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/057f7e48-aac2-407d-9fd4-555b5f8c6dad-dns-swift-storage-0\") pod \"dnsmasq-dns-586bdc5f9-lqn4x\" (UID: \"057f7e48-aac2-407d-9fd4-555b5f8c6dad\") " pod="openstack/dnsmasq-dns-586bdc5f9-lqn4x" Jan 26 17:08:55 crc kubenswrapper[4754]: I0126 17:08:55.707154 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b004ba0-01f6-49f9-8d8d-56033dcd3533-combined-ca-bundle\") pod \"barbican-keystone-listener-99649c64-sldzz\" (UID: \"0b004ba0-01f6-49f9-8d8d-56033dcd3533\") " pod="openstack/barbican-keystone-listener-99649c64-sldzz" Jan 26 17:08:55 crc kubenswrapper[4754]: I0126 17:08:55.707167 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/057f7e48-aac2-407d-9fd4-555b5f8c6dad-ovsdbserver-nb\") pod \"dnsmasq-dns-586bdc5f9-lqn4x\" (UID: \"057f7e48-aac2-407d-9fd4-555b5f8c6dad\") " pod="openstack/dnsmasq-dns-586bdc5f9-lqn4x" Jan 26 17:08:55 crc kubenswrapper[4754]: I0126 17:08:55.707182 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0b004ba0-01f6-49f9-8d8d-56033dcd3533-config-data\") pod \"barbican-keystone-listener-99649c64-sldzz\" (UID: \"0b004ba0-01f6-49f9-8d8d-56033dcd3533\") " pod="openstack/barbican-keystone-listener-99649c64-sldzz" Jan 26 17:08:55 crc kubenswrapper[4754]: I0126 17:08:55.707241 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/057f7e48-aac2-407d-9fd4-555b5f8c6dad-dns-svc\") pod \"dnsmasq-dns-586bdc5f9-lqn4x\" (UID: \"057f7e48-aac2-407d-9fd4-555b5f8c6dad\") " pod="openstack/dnsmasq-dns-586bdc5f9-lqn4x" Jan 26 17:08:55 crc kubenswrapper[4754]: I0126 17:08:55.707257 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0b004ba0-01f6-49f9-8d8d-56033dcd3533-logs\") pod \"barbican-keystone-listener-99649c64-sldzz\" (UID: \"0b004ba0-01f6-49f9-8d8d-56033dcd3533\") " pod="openstack/barbican-keystone-listener-99649c64-sldzz" Jan 26 17:08:55 crc kubenswrapper[4754]: I0126 17:08:55.707274 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fcpsj\" (UniqueName: \"kubernetes.io/projected/0b004ba0-01f6-49f9-8d8d-56033dcd3533-kube-api-access-fcpsj\") pod \"barbican-keystone-listener-99649c64-sldzz\" (UID: \"0b004ba0-01f6-49f9-8d8d-56033dcd3533\") " pod="openstack/barbican-keystone-listener-99649c64-sldzz" Jan 26 17:08:55 crc kubenswrapper[4754]: I0126 17:08:55.736648 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-586bdc5f9-lqn4x"] Jan 26 17:08:55 crc kubenswrapper[4754]: I0126 17:08:55.770857 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-5677764859-tswcr" Jan 26 17:08:55 crc kubenswrapper[4754]: I0126 17:08:55.820906 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/057f7e48-aac2-407d-9fd4-555b5f8c6dad-dns-svc\") pod \"dnsmasq-dns-586bdc5f9-lqn4x\" (UID: \"057f7e48-aac2-407d-9fd4-555b5f8c6dad\") " pod="openstack/dnsmasq-dns-586bdc5f9-lqn4x" Jan 26 17:08:55 crc kubenswrapper[4754]: I0126 17:08:55.821293 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0b004ba0-01f6-49f9-8d8d-56033dcd3533-logs\") pod \"barbican-keystone-listener-99649c64-sldzz\" (UID: \"0b004ba0-01f6-49f9-8d8d-56033dcd3533\") " pod="openstack/barbican-keystone-listener-99649c64-sldzz" Jan 26 17:08:55 crc kubenswrapper[4754]: I0126 17:08:55.821329 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fcpsj\" (UniqueName: \"kubernetes.io/projected/0b004ba0-01f6-49f9-8d8d-56033dcd3533-kube-api-access-fcpsj\") pod \"barbican-keystone-listener-99649c64-sldzz\" (UID: \"0b004ba0-01f6-49f9-8d8d-56033dcd3533\") " pod="openstack/barbican-keystone-listener-99649c64-sldzz" Jan 26 17:08:55 crc kubenswrapper[4754]: I0126 17:08:55.821434 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/057f7e48-aac2-407d-9fd4-555b5f8c6dad-ovsdbserver-sb\") pod \"dnsmasq-dns-586bdc5f9-lqn4x\" (UID: \"057f7e48-aac2-407d-9fd4-555b5f8c6dad\") " pod="openstack/dnsmasq-dns-586bdc5f9-lqn4x" Jan 26 17:08:55 crc kubenswrapper[4754]: I0126 17:08:55.821489 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/057f7e48-aac2-407d-9fd4-555b5f8c6dad-config\") pod \"dnsmasq-dns-586bdc5f9-lqn4x\" (UID: \"057f7e48-aac2-407d-9fd4-555b5f8c6dad\") " pod="openstack/dnsmasq-dns-586bdc5f9-lqn4x" Jan 26 17:08:55 crc kubenswrapper[4754]: I0126 17:08:55.821537 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jh76b\" (UniqueName: \"kubernetes.io/projected/057f7e48-aac2-407d-9fd4-555b5f8c6dad-kube-api-access-jh76b\") pod \"dnsmasq-dns-586bdc5f9-lqn4x\" (UID: \"057f7e48-aac2-407d-9fd4-555b5f8c6dad\") " pod="openstack/dnsmasq-dns-586bdc5f9-lqn4x" Jan 26 17:08:55 crc kubenswrapper[4754]: I0126 17:08:55.821685 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0b004ba0-01f6-49f9-8d8d-56033dcd3533-config-data-custom\") pod \"barbican-keystone-listener-99649c64-sldzz\" (UID: \"0b004ba0-01f6-49f9-8d8d-56033dcd3533\") " pod="openstack/barbican-keystone-listener-99649c64-sldzz" Jan 26 17:08:55 crc kubenswrapper[4754]: I0126 17:08:55.821759 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/057f7e48-aac2-407d-9fd4-555b5f8c6dad-dns-swift-storage-0\") pod \"dnsmasq-dns-586bdc5f9-lqn4x\" (UID: \"057f7e48-aac2-407d-9fd4-555b5f8c6dad\") " pod="openstack/dnsmasq-dns-586bdc5f9-lqn4x" Jan 26 17:08:55 crc kubenswrapper[4754]: I0126 17:08:55.821787 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b004ba0-01f6-49f9-8d8d-56033dcd3533-combined-ca-bundle\") pod \"barbican-keystone-listener-99649c64-sldzz\" (UID: \"0b004ba0-01f6-49f9-8d8d-56033dcd3533\") " pod="openstack/barbican-keystone-listener-99649c64-sldzz" Jan 26 17:08:55 crc kubenswrapper[4754]: I0126 17:08:55.821811 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/057f7e48-aac2-407d-9fd4-555b5f8c6dad-ovsdbserver-nb\") pod \"dnsmasq-dns-586bdc5f9-lqn4x\" (UID: \"057f7e48-aac2-407d-9fd4-555b5f8c6dad\") " pod="openstack/dnsmasq-dns-586bdc5f9-lqn4x" Jan 26 17:08:55 crc kubenswrapper[4754]: I0126 17:08:55.821837 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0b004ba0-01f6-49f9-8d8d-56033dcd3533-config-data\") pod \"barbican-keystone-listener-99649c64-sldzz\" (UID: \"0b004ba0-01f6-49f9-8d8d-56033dcd3533\") " pod="openstack/barbican-keystone-listener-99649c64-sldzz" Jan 26 17:08:55 crc kubenswrapper[4754]: I0126 17:08:55.822120 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/057f7e48-aac2-407d-9fd4-555b5f8c6dad-dns-svc\") pod \"dnsmasq-dns-586bdc5f9-lqn4x\" (UID: \"057f7e48-aac2-407d-9fd4-555b5f8c6dad\") " pod="openstack/dnsmasq-dns-586bdc5f9-lqn4x" Jan 26 17:08:55 crc kubenswrapper[4754]: I0126 17:08:55.822739 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0b004ba0-01f6-49f9-8d8d-56033dcd3533-logs\") pod \"barbican-keystone-listener-99649c64-sldzz\" (UID: \"0b004ba0-01f6-49f9-8d8d-56033dcd3533\") " pod="openstack/barbican-keystone-listener-99649c64-sldzz" Jan 26 17:08:55 crc kubenswrapper[4754]: I0126 17:08:55.823460 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/057f7e48-aac2-407d-9fd4-555b5f8c6dad-ovsdbserver-sb\") pod \"dnsmasq-dns-586bdc5f9-lqn4x\" (UID: \"057f7e48-aac2-407d-9fd4-555b5f8c6dad\") " pod="openstack/dnsmasq-dns-586bdc5f9-lqn4x" Jan 26 17:08:55 crc kubenswrapper[4754]: I0126 17:08:55.829831 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0b004ba0-01f6-49f9-8d8d-56033dcd3533-config-data-custom\") pod \"barbican-keystone-listener-99649c64-sldzz\" (UID: \"0b004ba0-01f6-49f9-8d8d-56033dcd3533\") " pod="openstack/barbican-keystone-listener-99649c64-sldzz" Jan 26 17:08:55 crc kubenswrapper[4754]: I0126 17:08:55.831465 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-5d7f8b89fd-jwf5r"] Jan 26 17:08:55 crc kubenswrapper[4754]: I0126 17:08:55.832800 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0b004ba0-01f6-49f9-8d8d-56033dcd3533-config-data\") pod \"barbican-keystone-listener-99649c64-sldzz\" (UID: \"0b004ba0-01f6-49f9-8d8d-56033dcd3533\") " pod="openstack/barbican-keystone-listener-99649c64-sldzz" Jan 26 17:08:55 crc kubenswrapper[4754]: I0126 17:08:55.835675 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b004ba0-01f6-49f9-8d8d-56033dcd3533-combined-ca-bundle\") pod \"barbican-keystone-listener-99649c64-sldzz\" (UID: \"0b004ba0-01f6-49f9-8d8d-56033dcd3533\") " pod="openstack/barbican-keystone-listener-99649c64-sldzz" Jan 26 17:08:55 crc kubenswrapper[4754]: I0126 17:08:55.837533 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/057f7e48-aac2-407d-9fd4-555b5f8c6dad-config\") pod \"dnsmasq-dns-586bdc5f9-lqn4x\" (UID: \"057f7e48-aac2-407d-9fd4-555b5f8c6dad\") " pod="openstack/dnsmasq-dns-586bdc5f9-lqn4x" Jan 26 17:08:55 crc kubenswrapper[4754]: I0126 17:08:55.840339 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5d7f8b89fd-jwf5r" Jan 26 17:08:55 crc kubenswrapper[4754]: I0126 17:08:55.842699 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/057f7e48-aac2-407d-9fd4-555b5f8c6dad-dns-swift-storage-0\") pod \"dnsmasq-dns-586bdc5f9-lqn4x\" (UID: \"057f7e48-aac2-407d-9fd4-555b5f8c6dad\") " pod="openstack/dnsmasq-dns-586bdc5f9-lqn4x" Jan 26 17:08:55 crc kubenswrapper[4754]: I0126 17:08:55.843831 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/057f7e48-aac2-407d-9fd4-555b5f8c6dad-ovsdbserver-nb\") pod \"dnsmasq-dns-586bdc5f9-lqn4x\" (UID: \"057f7e48-aac2-407d-9fd4-555b5f8c6dad\") " pod="openstack/dnsmasq-dns-586bdc5f9-lqn4x" Jan 26 17:08:55 crc kubenswrapper[4754]: I0126 17:08:55.848018 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jh76b\" (UniqueName: \"kubernetes.io/projected/057f7e48-aac2-407d-9fd4-555b5f8c6dad-kube-api-access-jh76b\") pod \"dnsmasq-dns-586bdc5f9-lqn4x\" (UID: \"057f7e48-aac2-407d-9fd4-555b5f8c6dad\") " pod="openstack/dnsmasq-dns-586bdc5f9-lqn4x" Jan 26 17:08:55 crc kubenswrapper[4754]: I0126 17:08:55.848329 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Jan 26 17:08:55 crc kubenswrapper[4754]: I0126 17:08:55.851237 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fcpsj\" (UniqueName: \"kubernetes.io/projected/0b004ba0-01f6-49f9-8d8d-56033dcd3533-kube-api-access-fcpsj\") pod \"barbican-keystone-listener-99649c64-sldzz\" (UID: \"0b004ba0-01f6-49f9-8d8d-56033dcd3533\") " pod="openstack/barbican-keystone-listener-99649c64-sldzz" Jan 26 17:08:55 crc kubenswrapper[4754]: I0126 17:08:55.857201 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-5d7f8b89fd-jwf5r"] Jan 26 17:08:55 crc kubenswrapper[4754]: I0126 17:08:55.873827 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-586bdc5f9-lqn4x" Jan 26 17:08:55 crc kubenswrapper[4754]: I0126 17:08:55.897893 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-99649c64-sldzz" Jan 26 17:08:56 crc kubenswrapper[4754]: I0126 17:08:56.029329 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4sjpt\" (UniqueName: \"kubernetes.io/projected/0dee3ce6-9657-4b8a-9034-7c744db2d209-kube-api-access-4sjpt\") pod \"barbican-api-5d7f8b89fd-jwf5r\" (UID: \"0dee3ce6-9657-4b8a-9034-7c744db2d209\") " pod="openstack/barbican-api-5d7f8b89fd-jwf5r" Jan 26 17:08:56 crc kubenswrapper[4754]: I0126 17:08:56.029967 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0dee3ce6-9657-4b8a-9034-7c744db2d209-logs\") pod \"barbican-api-5d7f8b89fd-jwf5r\" (UID: \"0dee3ce6-9657-4b8a-9034-7c744db2d209\") " pod="openstack/barbican-api-5d7f8b89fd-jwf5r" Jan 26 17:08:56 crc kubenswrapper[4754]: I0126 17:08:56.030033 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0dee3ce6-9657-4b8a-9034-7c744db2d209-config-data-custom\") pod \"barbican-api-5d7f8b89fd-jwf5r\" (UID: \"0dee3ce6-9657-4b8a-9034-7c744db2d209\") " pod="openstack/barbican-api-5d7f8b89fd-jwf5r" Jan 26 17:08:56 crc kubenswrapper[4754]: I0126 17:08:56.030086 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0dee3ce6-9657-4b8a-9034-7c744db2d209-config-data\") pod \"barbican-api-5d7f8b89fd-jwf5r\" (UID: \"0dee3ce6-9657-4b8a-9034-7c744db2d209\") " pod="openstack/barbican-api-5d7f8b89fd-jwf5r" Jan 26 17:08:56 crc kubenswrapper[4754]: I0126 17:08:56.030323 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0dee3ce6-9657-4b8a-9034-7c744db2d209-combined-ca-bundle\") pod \"barbican-api-5d7f8b89fd-jwf5r\" (UID: \"0dee3ce6-9657-4b8a-9034-7c744db2d209\") " pod="openstack/barbican-api-5d7f8b89fd-jwf5r" Jan 26 17:08:56 crc kubenswrapper[4754]: I0126 17:08:56.131773 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0dee3ce6-9657-4b8a-9034-7c744db2d209-combined-ca-bundle\") pod \"barbican-api-5d7f8b89fd-jwf5r\" (UID: \"0dee3ce6-9657-4b8a-9034-7c744db2d209\") " pod="openstack/barbican-api-5d7f8b89fd-jwf5r" Jan 26 17:08:56 crc kubenswrapper[4754]: I0126 17:08:56.131867 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4sjpt\" (UniqueName: \"kubernetes.io/projected/0dee3ce6-9657-4b8a-9034-7c744db2d209-kube-api-access-4sjpt\") pod \"barbican-api-5d7f8b89fd-jwf5r\" (UID: \"0dee3ce6-9657-4b8a-9034-7c744db2d209\") " pod="openstack/barbican-api-5d7f8b89fd-jwf5r" Jan 26 17:08:56 crc kubenswrapper[4754]: I0126 17:08:56.131894 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0dee3ce6-9657-4b8a-9034-7c744db2d209-logs\") pod \"barbican-api-5d7f8b89fd-jwf5r\" (UID: \"0dee3ce6-9657-4b8a-9034-7c744db2d209\") " pod="openstack/barbican-api-5d7f8b89fd-jwf5r" Jan 26 17:08:56 crc kubenswrapper[4754]: I0126 17:08:56.131928 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0dee3ce6-9657-4b8a-9034-7c744db2d209-config-data-custom\") pod \"barbican-api-5d7f8b89fd-jwf5r\" (UID: \"0dee3ce6-9657-4b8a-9034-7c744db2d209\") " pod="openstack/barbican-api-5d7f8b89fd-jwf5r" Jan 26 17:08:56 crc kubenswrapper[4754]: I0126 17:08:56.132275 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0dee3ce6-9657-4b8a-9034-7c744db2d209-config-data\") pod \"barbican-api-5d7f8b89fd-jwf5r\" (UID: \"0dee3ce6-9657-4b8a-9034-7c744db2d209\") " pod="openstack/barbican-api-5d7f8b89fd-jwf5r" Jan 26 17:08:56 crc kubenswrapper[4754]: I0126 17:08:56.132690 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0dee3ce6-9657-4b8a-9034-7c744db2d209-logs\") pod \"barbican-api-5d7f8b89fd-jwf5r\" (UID: \"0dee3ce6-9657-4b8a-9034-7c744db2d209\") " pod="openstack/barbican-api-5d7f8b89fd-jwf5r" Jan 26 17:08:56 crc kubenswrapper[4754]: I0126 17:08:56.136496 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0dee3ce6-9657-4b8a-9034-7c744db2d209-config-data-custom\") pod \"barbican-api-5d7f8b89fd-jwf5r\" (UID: \"0dee3ce6-9657-4b8a-9034-7c744db2d209\") " pod="openstack/barbican-api-5d7f8b89fd-jwf5r" Jan 26 17:08:56 crc kubenswrapper[4754]: I0126 17:08:56.138621 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0dee3ce6-9657-4b8a-9034-7c744db2d209-combined-ca-bundle\") pod \"barbican-api-5d7f8b89fd-jwf5r\" (UID: \"0dee3ce6-9657-4b8a-9034-7c744db2d209\") " pod="openstack/barbican-api-5d7f8b89fd-jwf5r" Jan 26 17:08:56 crc kubenswrapper[4754]: I0126 17:08:56.139601 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0dee3ce6-9657-4b8a-9034-7c744db2d209-config-data\") pod \"barbican-api-5d7f8b89fd-jwf5r\" (UID: \"0dee3ce6-9657-4b8a-9034-7c744db2d209\") " pod="openstack/barbican-api-5d7f8b89fd-jwf5r" Jan 26 17:08:56 crc kubenswrapper[4754]: I0126 17:08:56.153602 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4sjpt\" (UniqueName: \"kubernetes.io/projected/0dee3ce6-9657-4b8a-9034-7c744db2d209-kube-api-access-4sjpt\") pod \"barbican-api-5d7f8b89fd-jwf5r\" (UID: \"0dee3ce6-9657-4b8a-9034-7c744db2d209\") " pod="openstack/barbican-api-5d7f8b89fd-jwf5r" Jan 26 17:08:56 crc kubenswrapper[4754]: I0126 17:08:56.198606 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5d7f8b89fd-jwf5r" Jan 26 17:08:56 crc kubenswrapper[4754]: I0126 17:08:56.307314 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-5677764859-tswcr"] Jan 26 17:08:56 crc kubenswrapper[4754]: I0126 17:08:56.420972 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-586bdc5f9-lqn4x"] Jan 26 17:08:56 crc kubenswrapper[4754]: I0126 17:08:56.433482 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-99649c64-sldzz"] Jan 26 17:08:56 crc kubenswrapper[4754]: W0126 17:08:56.437570 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0b004ba0_01f6_49f9_8d8d_56033dcd3533.slice/crio-78b2f30f9cc7f260c6b32f2c9c147cb251c84644132a6642c7b082af3fcd7a17 WatchSource:0}: Error finding container 78b2f30f9cc7f260c6b32f2c9c147cb251c84644132a6642c7b082af3fcd7a17: Status 404 returned error can't find the container with id 78b2f30f9cc7f260c6b32f2c9c147cb251c84644132a6642c7b082af3fcd7a17 Jan 26 17:08:56 crc kubenswrapper[4754]: I0126 17:08:56.575541 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-99649c64-sldzz" event={"ID":"0b004ba0-01f6-49f9-8d8d-56033dcd3533","Type":"ContainerStarted","Data":"78b2f30f9cc7f260c6b32f2c9c147cb251c84644132a6642c7b082af3fcd7a17"} Jan 26 17:08:56 crc kubenswrapper[4754]: I0126 17:08:56.576870 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-5677764859-tswcr" event={"ID":"d64625ff-6586-4c80-a720-b2febd49a966","Type":"ContainerStarted","Data":"524bc32ae3b8384c9c1971cee91cf21ddd48c2a502f64950051a38343c46f81c"} Jan 26 17:08:56 crc kubenswrapper[4754]: I0126 17:08:56.578121 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-586bdc5f9-lqn4x" event={"ID":"057f7e48-aac2-407d-9fd4-555b5f8c6dad","Type":"ContainerStarted","Data":"1d797d02e0a18f2e8fd1c14ae47ae138d6e9f382b84be194f3b2a811d63f5c2e"} Jan 26 17:08:56 crc kubenswrapper[4754]: W0126 17:08:56.615906 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0dee3ce6_9657_4b8a_9034_7c744db2d209.slice/crio-17f7bc11c09ec7d6307edb1ea875db377ca62b684188c883a065b30028bc2138 WatchSource:0}: Error finding container 17f7bc11c09ec7d6307edb1ea875db377ca62b684188c883a065b30028bc2138: Status 404 returned error can't find the container with id 17f7bc11c09ec7d6307edb1ea875db377ca62b684188c883a065b30028bc2138 Jan 26 17:08:56 crc kubenswrapper[4754]: I0126 17:08:56.621476 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-5d7f8b89fd-jwf5r"] Jan 26 17:08:56 crc kubenswrapper[4754]: E0126 17:08:56.643882 4754 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/ubi9/httpd-24:latest" Jan 26 17:08:56 crc kubenswrapper[4754]: E0126 17:08:56.644152 4754 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:proxy-httpd,Image:registry.redhat.io/ubi9/httpd-24:latest,Command:[/usr/sbin/httpd],Args:[-DFOREGROUND],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:proxy-httpd,HostPort:0,ContainerPort:3000,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/httpd/conf/httpd.conf,SubPath:httpd.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/httpd/conf.d/ssl.conf,SubPath:ssl.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:run-httpd,ReadOnly:false,MountPath:/run/httpd,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:log-httpd,ReadOnly:false,MountPath:/var/log/httpd,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-2bpjj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/,Port:{0 3000 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:300,TimeoutSeconds:30,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/,Port:{0 3000 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:10,TimeoutSeconds:30,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(d84d5918-51f5-4131-9caf-9380004a0ca5): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Jan 26 17:08:56 crc kubenswrapper[4754]: E0126 17:08:56.646092 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"ceilometer-central-agent\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"proxy-httpd\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"]" pod="openstack/ceilometer-0" podUID="d84d5918-51f5-4131-9caf-9380004a0ca5" Jan 26 17:08:57 crc kubenswrapper[4754]: I0126 17:08:57.079649 4754 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-85f95cb57b-99z67" podUID="ea911a5c-f353-4da9-9900-bfcbb234a377" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.154:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.154:8443: connect: connection refused" Jan 26 17:08:57 crc kubenswrapper[4754]: I0126 17:08:57.080023 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-85f95cb57b-99z67" Jan 26 17:08:57 crc kubenswrapper[4754]: I0126 17:08:57.590032 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5d7f8b89fd-jwf5r" event={"ID":"0dee3ce6-9657-4b8a-9034-7c744db2d209","Type":"ContainerStarted","Data":"17f7bc11c09ec7d6307edb1ea875db377ca62b684188c883a065b30028bc2138"} Jan 26 17:08:57 crc kubenswrapper[4754]: I0126 17:08:57.591035 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d84d5918-51f5-4131-9caf-9380004a0ca5" containerName="ceilometer-notification-agent" containerID="cri-o://2b6a9e46662327f03d044bfee7f9926d7e4287b3bd44ae06415f5f1bf5a68abc" gracePeriod=30 Jan 26 17:08:57 crc kubenswrapper[4754]: I0126 17:08:57.591136 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d84d5918-51f5-4131-9caf-9380004a0ca5" containerName="sg-core" containerID="cri-o://114035b4c3d582ed66db03f17a7b8f1e20a340ca4a189ef2996ccd2fcb60ce3b" gracePeriod=30 Jan 26 17:08:58 crc kubenswrapper[4754]: I0126 17:08:58.314946 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-5d4bd978-rj87r"] Jan 26 17:08:58 crc kubenswrapper[4754]: I0126 17:08:58.316628 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5d4bd978-rj87r" Jan 26 17:08:58 crc kubenswrapper[4754]: I0126 17:08:58.321319 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Jan 26 17:08:58 crc kubenswrapper[4754]: I0126 17:08:58.324576 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Jan 26 17:08:58 crc kubenswrapper[4754]: I0126 17:08:58.331173 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-5d4bd978-rj87r"] Jan 26 17:08:58 crc kubenswrapper[4754]: I0126 17:08:58.474702 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8c7859b2-be6e-49c3-8392-47ed649a1f68-internal-tls-certs\") pod \"barbican-api-5d4bd978-rj87r\" (UID: \"8c7859b2-be6e-49c3-8392-47ed649a1f68\") " pod="openstack/barbican-api-5d4bd978-rj87r" Jan 26 17:08:58 crc kubenswrapper[4754]: I0126 17:08:58.474811 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8c7859b2-be6e-49c3-8392-47ed649a1f68-config-data-custom\") pod \"barbican-api-5d4bd978-rj87r\" (UID: \"8c7859b2-be6e-49c3-8392-47ed649a1f68\") " pod="openstack/barbican-api-5d4bd978-rj87r" Jan 26 17:08:58 crc kubenswrapper[4754]: I0126 17:08:58.474959 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8c7859b2-be6e-49c3-8392-47ed649a1f68-logs\") pod \"barbican-api-5d4bd978-rj87r\" (UID: \"8c7859b2-be6e-49c3-8392-47ed649a1f68\") " pod="openstack/barbican-api-5d4bd978-rj87r" Jan 26 17:08:58 crc kubenswrapper[4754]: I0126 17:08:58.475088 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8c7859b2-be6e-49c3-8392-47ed649a1f68-public-tls-certs\") pod \"barbican-api-5d4bd978-rj87r\" (UID: \"8c7859b2-be6e-49c3-8392-47ed649a1f68\") " pod="openstack/barbican-api-5d4bd978-rj87r" Jan 26 17:08:58 crc kubenswrapper[4754]: I0126 17:08:58.475153 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c7859b2-be6e-49c3-8392-47ed649a1f68-config-data\") pod \"barbican-api-5d4bd978-rj87r\" (UID: \"8c7859b2-be6e-49c3-8392-47ed649a1f68\") " pod="openstack/barbican-api-5d4bd978-rj87r" Jan 26 17:08:58 crc kubenswrapper[4754]: I0126 17:08:58.475256 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-98frb\" (UniqueName: \"kubernetes.io/projected/8c7859b2-be6e-49c3-8392-47ed649a1f68-kube-api-access-98frb\") pod \"barbican-api-5d4bd978-rj87r\" (UID: \"8c7859b2-be6e-49c3-8392-47ed649a1f68\") " pod="openstack/barbican-api-5d4bd978-rj87r" Jan 26 17:08:58 crc kubenswrapper[4754]: I0126 17:08:58.475324 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c7859b2-be6e-49c3-8392-47ed649a1f68-combined-ca-bundle\") pod \"barbican-api-5d4bd978-rj87r\" (UID: \"8c7859b2-be6e-49c3-8392-47ed649a1f68\") " pod="openstack/barbican-api-5d4bd978-rj87r" Jan 26 17:08:58 crc kubenswrapper[4754]: I0126 17:08:58.576861 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8c7859b2-be6e-49c3-8392-47ed649a1f68-internal-tls-certs\") pod \"barbican-api-5d4bd978-rj87r\" (UID: \"8c7859b2-be6e-49c3-8392-47ed649a1f68\") " pod="openstack/barbican-api-5d4bd978-rj87r" Jan 26 17:08:58 crc kubenswrapper[4754]: I0126 17:08:58.577188 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8c7859b2-be6e-49c3-8392-47ed649a1f68-config-data-custom\") pod \"barbican-api-5d4bd978-rj87r\" (UID: \"8c7859b2-be6e-49c3-8392-47ed649a1f68\") " pod="openstack/barbican-api-5d4bd978-rj87r" Jan 26 17:08:58 crc kubenswrapper[4754]: I0126 17:08:58.577329 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8c7859b2-be6e-49c3-8392-47ed649a1f68-logs\") pod \"barbican-api-5d4bd978-rj87r\" (UID: \"8c7859b2-be6e-49c3-8392-47ed649a1f68\") " pod="openstack/barbican-api-5d4bd978-rj87r" Jan 26 17:08:58 crc kubenswrapper[4754]: I0126 17:08:58.577467 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8c7859b2-be6e-49c3-8392-47ed649a1f68-public-tls-certs\") pod \"barbican-api-5d4bd978-rj87r\" (UID: \"8c7859b2-be6e-49c3-8392-47ed649a1f68\") " pod="openstack/barbican-api-5d4bd978-rj87r" Jan 26 17:08:58 crc kubenswrapper[4754]: I0126 17:08:58.577568 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c7859b2-be6e-49c3-8392-47ed649a1f68-config-data\") pod \"barbican-api-5d4bd978-rj87r\" (UID: \"8c7859b2-be6e-49c3-8392-47ed649a1f68\") " pod="openstack/barbican-api-5d4bd978-rj87r" Jan 26 17:08:58 crc kubenswrapper[4754]: I0126 17:08:58.577722 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-98frb\" (UniqueName: \"kubernetes.io/projected/8c7859b2-be6e-49c3-8392-47ed649a1f68-kube-api-access-98frb\") pod \"barbican-api-5d4bd978-rj87r\" (UID: \"8c7859b2-be6e-49c3-8392-47ed649a1f68\") " pod="openstack/barbican-api-5d4bd978-rj87r" Jan 26 17:08:58 crc kubenswrapper[4754]: I0126 17:08:58.577836 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c7859b2-be6e-49c3-8392-47ed649a1f68-combined-ca-bundle\") pod \"barbican-api-5d4bd978-rj87r\" (UID: \"8c7859b2-be6e-49c3-8392-47ed649a1f68\") " pod="openstack/barbican-api-5d4bd978-rj87r" Jan 26 17:08:58 crc kubenswrapper[4754]: I0126 17:08:58.578207 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8c7859b2-be6e-49c3-8392-47ed649a1f68-logs\") pod \"barbican-api-5d4bd978-rj87r\" (UID: \"8c7859b2-be6e-49c3-8392-47ed649a1f68\") " pod="openstack/barbican-api-5d4bd978-rj87r" Jan 26 17:08:58 crc kubenswrapper[4754]: I0126 17:08:58.583903 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c7859b2-be6e-49c3-8392-47ed649a1f68-combined-ca-bundle\") pod \"barbican-api-5d4bd978-rj87r\" (UID: \"8c7859b2-be6e-49c3-8392-47ed649a1f68\") " pod="openstack/barbican-api-5d4bd978-rj87r" Jan 26 17:08:58 crc kubenswrapper[4754]: I0126 17:08:58.584700 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c7859b2-be6e-49c3-8392-47ed649a1f68-config-data\") pod \"barbican-api-5d4bd978-rj87r\" (UID: \"8c7859b2-be6e-49c3-8392-47ed649a1f68\") " pod="openstack/barbican-api-5d4bd978-rj87r" Jan 26 17:08:58 crc kubenswrapper[4754]: I0126 17:08:58.584828 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8c7859b2-be6e-49c3-8392-47ed649a1f68-config-data-custom\") pod \"barbican-api-5d4bd978-rj87r\" (UID: \"8c7859b2-be6e-49c3-8392-47ed649a1f68\") " pod="openstack/barbican-api-5d4bd978-rj87r" Jan 26 17:08:58 crc kubenswrapper[4754]: I0126 17:08:58.584928 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8c7859b2-be6e-49c3-8392-47ed649a1f68-internal-tls-certs\") pod \"barbican-api-5d4bd978-rj87r\" (UID: \"8c7859b2-be6e-49c3-8392-47ed649a1f68\") " pod="openstack/barbican-api-5d4bd978-rj87r" Jan 26 17:08:58 crc kubenswrapper[4754]: I0126 17:08:58.585264 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8c7859b2-be6e-49c3-8392-47ed649a1f68-public-tls-certs\") pod \"barbican-api-5d4bd978-rj87r\" (UID: \"8c7859b2-be6e-49c3-8392-47ed649a1f68\") " pod="openstack/barbican-api-5d4bd978-rj87r" Jan 26 17:08:58 crc kubenswrapper[4754]: I0126 17:08:58.599378 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-98frb\" (UniqueName: \"kubernetes.io/projected/8c7859b2-be6e-49c3-8392-47ed649a1f68-kube-api-access-98frb\") pod \"barbican-api-5d4bd978-rj87r\" (UID: \"8c7859b2-be6e-49c3-8392-47ed649a1f68\") " pod="openstack/barbican-api-5d4bd978-rj87r" Jan 26 17:08:58 crc kubenswrapper[4754]: I0126 17:08:58.638143 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5d4bd978-rj87r" Jan 26 17:08:59 crc kubenswrapper[4754]: I0126 17:08:59.058132 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-5d4bd978-rj87r"] Jan 26 17:08:59 crc kubenswrapper[4754]: I0126 17:08:59.606097 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5d4bd978-rj87r" event={"ID":"8c7859b2-be6e-49c3-8392-47ed649a1f68","Type":"ContainerStarted","Data":"b92a60cec0c097e4e2066332499ec60419579fabe0fe43cf9b73a35157ed84e4"} Jan 26 17:09:00 crc kubenswrapper[4754]: I0126 17:09:00.623043 4754 generic.go:334] "Generic (PLEG): container finished" podID="d84d5918-51f5-4131-9caf-9380004a0ca5" containerID="114035b4c3d582ed66db03f17a7b8f1e20a340ca4a189ef2996ccd2fcb60ce3b" exitCode=2 Jan 26 17:09:00 crc kubenswrapper[4754]: I0126 17:09:00.623337 4754 generic.go:334] "Generic (PLEG): container finished" podID="d84d5918-51f5-4131-9caf-9380004a0ca5" containerID="2b6a9e46662327f03d044bfee7f9926d7e4287b3bd44ae06415f5f1bf5a68abc" exitCode=0 Jan 26 17:09:00 crc kubenswrapper[4754]: I0126 17:09:00.623127 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d84d5918-51f5-4131-9caf-9380004a0ca5","Type":"ContainerDied","Data":"114035b4c3d582ed66db03f17a7b8f1e20a340ca4a189ef2996ccd2fcb60ce3b"} Jan 26 17:09:00 crc kubenswrapper[4754]: I0126 17:09:00.623401 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d84d5918-51f5-4131-9caf-9380004a0ca5","Type":"ContainerDied","Data":"2b6a9e46662327f03d044bfee7f9926d7e4287b3bd44ae06415f5f1bf5a68abc"} Jan 26 17:09:00 crc kubenswrapper[4754]: I0126 17:09:00.625055 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5d7f8b89fd-jwf5r" event={"ID":"0dee3ce6-9657-4b8a-9034-7c744db2d209","Type":"ContainerStarted","Data":"a5dcc43e1e1644072f8d0a7424cbbc5cfda3573dc5d056f0c4781edb61fa886e"} Jan 26 17:09:00 crc kubenswrapper[4754]: I0126 17:09:00.626521 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5d4bd978-rj87r" event={"ID":"8c7859b2-be6e-49c3-8392-47ed649a1f68","Type":"ContainerStarted","Data":"36a08bd76370961f776f1050a2f3b2d775a55d4bc8a4a3f3c07a92f542423ce1"} Jan 26 17:09:00 crc kubenswrapper[4754]: I0126 17:09:00.635603 4754 generic.go:334] "Generic (PLEG): container finished" podID="057f7e48-aac2-407d-9fd4-555b5f8c6dad" containerID="cf88012f2cc61bc7a32fa7efc83cfc617efba3a5eaf348518664d834f71460ab" exitCode=0 Jan 26 17:09:00 crc kubenswrapper[4754]: I0126 17:09:00.635648 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-586bdc5f9-lqn4x" event={"ID":"057f7e48-aac2-407d-9fd4-555b5f8c6dad","Type":"ContainerDied","Data":"cf88012f2cc61bc7a32fa7efc83cfc617efba3a5eaf348518664d834f71460ab"} Jan 26 17:09:00 crc kubenswrapper[4754]: I0126 17:09:00.821799 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 26 17:09:00 crc kubenswrapper[4754]: I0126 17:09:00.918306 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d84d5918-51f5-4131-9caf-9380004a0ca5-run-httpd\") pod \"d84d5918-51f5-4131-9caf-9380004a0ca5\" (UID: \"d84d5918-51f5-4131-9caf-9380004a0ca5\") " Jan 26 17:09:00 crc kubenswrapper[4754]: I0126 17:09:00.918358 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2bpjj\" (UniqueName: \"kubernetes.io/projected/d84d5918-51f5-4131-9caf-9380004a0ca5-kube-api-access-2bpjj\") pod \"d84d5918-51f5-4131-9caf-9380004a0ca5\" (UID: \"d84d5918-51f5-4131-9caf-9380004a0ca5\") " Jan 26 17:09:00 crc kubenswrapper[4754]: I0126 17:09:00.918391 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d84d5918-51f5-4131-9caf-9380004a0ca5-sg-core-conf-yaml\") pod \"d84d5918-51f5-4131-9caf-9380004a0ca5\" (UID: \"d84d5918-51f5-4131-9caf-9380004a0ca5\") " Jan 26 17:09:00 crc kubenswrapper[4754]: I0126 17:09:00.918474 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d84d5918-51f5-4131-9caf-9380004a0ca5-config-data\") pod \"d84d5918-51f5-4131-9caf-9380004a0ca5\" (UID: \"d84d5918-51f5-4131-9caf-9380004a0ca5\") " Jan 26 17:09:00 crc kubenswrapper[4754]: I0126 17:09:00.918506 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d84d5918-51f5-4131-9caf-9380004a0ca5-combined-ca-bundle\") pod \"d84d5918-51f5-4131-9caf-9380004a0ca5\" (UID: \"d84d5918-51f5-4131-9caf-9380004a0ca5\") " Jan 26 17:09:00 crc kubenswrapper[4754]: I0126 17:09:00.918547 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d84d5918-51f5-4131-9caf-9380004a0ca5-log-httpd\") pod \"d84d5918-51f5-4131-9caf-9380004a0ca5\" (UID: \"d84d5918-51f5-4131-9caf-9380004a0ca5\") " Jan 26 17:09:00 crc kubenswrapper[4754]: I0126 17:09:00.918578 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d84d5918-51f5-4131-9caf-9380004a0ca5-scripts\") pod \"d84d5918-51f5-4131-9caf-9380004a0ca5\" (UID: \"d84d5918-51f5-4131-9caf-9380004a0ca5\") " Jan 26 17:09:00 crc kubenswrapper[4754]: I0126 17:09:00.918985 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d84d5918-51f5-4131-9caf-9380004a0ca5-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "d84d5918-51f5-4131-9caf-9380004a0ca5" (UID: "d84d5918-51f5-4131-9caf-9380004a0ca5"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:09:00 crc kubenswrapper[4754]: I0126 17:09:00.919373 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d84d5918-51f5-4131-9caf-9380004a0ca5-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "d84d5918-51f5-4131-9caf-9380004a0ca5" (UID: "d84d5918-51f5-4131-9caf-9380004a0ca5"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:09:00 crc kubenswrapper[4754]: I0126 17:09:00.919943 4754 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d84d5918-51f5-4131-9caf-9380004a0ca5-run-httpd\") on node \"crc\" DevicePath \"\"" Jan 26 17:09:00 crc kubenswrapper[4754]: I0126 17:09:00.919963 4754 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d84d5918-51f5-4131-9caf-9380004a0ca5-log-httpd\") on node \"crc\" DevicePath \"\"" Jan 26 17:09:00 crc kubenswrapper[4754]: I0126 17:09:00.923526 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d84d5918-51f5-4131-9caf-9380004a0ca5-scripts" (OuterVolumeSpecName: "scripts") pod "d84d5918-51f5-4131-9caf-9380004a0ca5" (UID: "d84d5918-51f5-4131-9caf-9380004a0ca5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:09:00 crc kubenswrapper[4754]: I0126 17:09:00.924425 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d84d5918-51f5-4131-9caf-9380004a0ca5-kube-api-access-2bpjj" (OuterVolumeSpecName: "kube-api-access-2bpjj") pod "d84d5918-51f5-4131-9caf-9380004a0ca5" (UID: "d84d5918-51f5-4131-9caf-9380004a0ca5"). InnerVolumeSpecName "kube-api-access-2bpjj". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:09:00 crc kubenswrapper[4754]: I0126 17:09:00.945855 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d84d5918-51f5-4131-9caf-9380004a0ca5-config-data" (OuterVolumeSpecName: "config-data") pod "d84d5918-51f5-4131-9caf-9380004a0ca5" (UID: "d84d5918-51f5-4131-9caf-9380004a0ca5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:09:00 crc kubenswrapper[4754]: I0126 17:09:00.955446 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d84d5918-51f5-4131-9caf-9380004a0ca5-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "d84d5918-51f5-4131-9caf-9380004a0ca5" (UID: "d84d5918-51f5-4131-9caf-9380004a0ca5"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:09:00 crc kubenswrapper[4754]: I0126 17:09:00.958606 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d84d5918-51f5-4131-9caf-9380004a0ca5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d84d5918-51f5-4131-9caf-9380004a0ca5" (UID: "d84d5918-51f5-4131-9caf-9380004a0ca5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:09:01 crc kubenswrapper[4754]: I0126 17:09:01.021302 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2bpjj\" (UniqueName: \"kubernetes.io/projected/d84d5918-51f5-4131-9caf-9380004a0ca5-kube-api-access-2bpjj\") on node \"crc\" DevicePath \"\"" Jan 26 17:09:01 crc kubenswrapper[4754]: I0126 17:09:01.021338 4754 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d84d5918-51f5-4131-9caf-9380004a0ca5-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Jan 26 17:09:01 crc kubenswrapper[4754]: I0126 17:09:01.021347 4754 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d84d5918-51f5-4131-9caf-9380004a0ca5-config-data\") on node \"crc\" DevicePath \"\"" Jan 26 17:09:01 crc kubenswrapper[4754]: I0126 17:09:01.021355 4754 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d84d5918-51f5-4131-9caf-9380004a0ca5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 26 17:09:01 crc kubenswrapper[4754]: I0126 17:09:01.021363 4754 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d84d5918-51f5-4131-9caf-9380004a0ca5-scripts\") on node \"crc\" DevicePath \"\"" Jan 26 17:09:01 crc kubenswrapper[4754]: I0126 17:09:01.662629 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5d4bd978-rj87r" event={"ID":"8c7859b2-be6e-49c3-8392-47ed649a1f68","Type":"ContainerStarted","Data":"32df5ed79972eacc1b95aa8e9517356a22268613a5b10c77511da6ec0178fe40"} Jan 26 17:09:01 crc kubenswrapper[4754]: I0126 17:09:01.663000 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-5d4bd978-rj87r" Jan 26 17:09:01 crc kubenswrapper[4754]: I0126 17:09:01.663021 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-5d4bd978-rj87r" Jan 26 17:09:01 crc kubenswrapper[4754]: I0126 17:09:01.668282 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-586bdc5f9-lqn4x" event={"ID":"057f7e48-aac2-407d-9fd4-555b5f8c6dad","Type":"ContainerStarted","Data":"2e5ddf125f0ef1060a21de205d16388b079b15d704493d09a9b6fa0a3bc82c2d"} Jan 26 17:09:01 crc kubenswrapper[4754]: I0126 17:09:01.669206 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-586bdc5f9-lqn4x" Jan 26 17:09:01 crc kubenswrapper[4754]: I0126 17:09:01.672158 4754 generic.go:334] "Generic (PLEG): container finished" podID="ea911a5c-f353-4da9-9900-bfcbb234a377" containerID="e920e8dd93fe23ed63bbaaac5aa05943a59d0ede5d920fc5ddf3c89066cffba4" exitCode=137 Jan 26 17:09:01 crc kubenswrapper[4754]: I0126 17:09:01.672223 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-85f95cb57b-99z67" event={"ID":"ea911a5c-f353-4da9-9900-bfcbb234a377","Type":"ContainerDied","Data":"e920e8dd93fe23ed63bbaaac5aa05943a59d0ede5d920fc5ddf3c89066cffba4"} Jan 26 17:09:01 crc kubenswrapper[4754]: I0126 17:09:01.676206 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d84d5918-51f5-4131-9caf-9380004a0ca5","Type":"ContainerDied","Data":"d699729cd44d0e51f597fa6591ec44c18203dc73bc45ce3e89d2c8230052c8c9"} Jan 26 17:09:01 crc kubenswrapper[4754]: I0126 17:09:01.676245 4754 scope.go:117] "RemoveContainer" containerID="114035b4c3d582ed66db03f17a7b8f1e20a340ca4a189ef2996ccd2fcb60ce3b" Jan 26 17:09:01 crc kubenswrapper[4754]: I0126 17:09:01.676414 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 26 17:09:01 crc kubenswrapper[4754]: I0126 17:09:01.686539 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-5d4bd978-rj87r" podStartSLOduration=3.686523375 podStartE2EDuration="3.686523375s" podCreationTimestamp="2026-01-26 17:08:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 17:09:01.684143672 +0000 UTC m=+1308.208324116" watchObservedRunningTime="2026-01-26 17:09:01.686523375 +0000 UTC m=+1308.210703809" Jan 26 17:09:01 crc kubenswrapper[4754]: I0126 17:09:01.689810 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5d7f8b89fd-jwf5r" event={"ID":"0dee3ce6-9657-4b8a-9034-7c744db2d209","Type":"ContainerStarted","Data":"6401776521d0c72d1ea925c21389233c039edc6ed1e4450e98139bea1afe8ec3"} Jan 26 17:09:01 crc kubenswrapper[4754]: I0126 17:09:01.690556 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-5d7f8b89fd-jwf5r" Jan 26 17:09:01 crc kubenswrapper[4754]: I0126 17:09:01.690578 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-5d7f8b89fd-jwf5r" Jan 26 17:09:01 crc kubenswrapper[4754]: I0126 17:09:01.717833 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-586bdc5f9-lqn4x" podStartSLOduration=6.7178127629999995 podStartE2EDuration="6.717812763s" podCreationTimestamp="2026-01-26 17:08:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 17:09:01.709831904 +0000 UTC m=+1308.234012348" watchObservedRunningTime="2026-01-26 17:09:01.717812763 +0000 UTC m=+1308.241993197" Jan 26 17:09:01 crc kubenswrapper[4754]: I0126 17:09:01.741622 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-5d7f8b89fd-jwf5r" podStartSLOduration=6.741600295 podStartE2EDuration="6.741600295s" podCreationTimestamp="2026-01-26 17:08:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 17:09:01.728823491 +0000 UTC m=+1308.253003925" watchObservedRunningTime="2026-01-26 17:09:01.741600295 +0000 UTC m=+1308.265780729" Jan 26 17:09:01 crc kubenswrapper[4754]: I0126 17:09:01.795264 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Jan 26 17:09:01 crc kubenswrapper[4754]: I0126 17:09:01.800541 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Jan 26 17:09:01 crc kubenswrapper[4754]: I0126 17:09:01.815872 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Jan 26 17:09:01 crc kubenswrapper[4754]: E0126 17:09:01.820492 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d84d5918-51f5-4131-9caf-9380004a0ca5" containerName="ceilometer-notification-agent" Jan 26 17:09:01 crc kubenswrapper[4754]: I0126 17:09:01.820516 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="d84d5918-51f5-4131-9caf-9380004a0ca5" containerName="ceilometer-notification-agent" Jan 26 17:09:01 crc kubenswrapper[4754]: E0126 17:09:01.820532 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d84d5918-51f5-4131-9caf-9380004a0ca5" containerName="sg-core" Jan 26 17:09:01 crc kubenswrapper[4754]: I0126 17:09:01.820540 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="d84d5918-51f5-4131-9caf-9380004a0ca5" containerName="sg-core" Jan 26 17:09:01 crc kubenswrapper[4754]: I0126 17:09:01.820771 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="d84d5918-51f5-4131-9caf-9380004a0ca5" containerName="ceilometer-notification-agent" Jan 26 17:09:01 crc kubenswrapper[4754]: I0126 17:09:01.820794 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="d84d5918-51f5-4131-9caf-9380004a0ca5" containerName="sg-core" Jan 26 17:09:01 crc kubenswrapper[4754]: I0126 17:09:01.827516 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 26 17:09:01 crc kubenswrapper[4754]: I0126 17:09:01.838292 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Jan 26 17:09:01 crc kubenswrapper[4754]: I0126 17:09:01.838562 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Jan 26 17:09:01 crc kubenswrapper[4754]: I0126 17:09:01.839936 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Jan 26 17:09:01 crc kubenswrapper[4754]: I0126 17:09:01.936470 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4c65192f-a05a-45e2-9d56-33ddeb265374-log-httpd\") pod \"ceilometer-0\" (UID: \"4c65192f-a05a-45e2-9d56-33ddeb265374\") " pod="openstack/ceilometer-0" Jan 26 17:09:01 crc kubenswrapper[4754]: I0126 17:09:01.936836 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c65192f-a05a-45e2-9d56-33ddeb265374-config-data\") pod \"ceilometer-0\" (UID: \"4c65192f-a05a-45e2-9d56-33ddeb265374\") " pod="openstack/ceilometer-0" Jan 26 17:09:01 crc kubenswrapper[4754]: I0126 17:09:01.936906 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4c65192f-a05a-45e2-9d56-33ddeb265374-scripts\") pod \"ceilometer-0\" (UID: \"4c65192f-a05a-45e2-9d56-33ddeb265374\") " pod="openstack/ceilometer-0" Jan 26 17:09:01 crc kubenswrapper[4754]: I0126 17:09:01.936991 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4vqll\" (UniqueName: \"kubernetes.io/projected/4c65192f-a05a-45e2-9d56-33ddeb265374-kube-api-access-4vqll\") pod \"ceilometer-0\" (UID: \"4c65192f-a05a-45e2-9d56-33ddeb265374\") " pod="openstack/ceilometer-0" Jan 26 17:09:01 crc kubenswrapper[4754]: I0126 17:09:01.937026 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4c65192f-a05a-45e2-9d56-33ddeb265374-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4c65192f-a05a-45e2-9d56-33ddeb265374\") " pod="openstack/ceilometer-0" Jan 26 17:09:01 crc kubenswrapper[4754]: I0126 17:09:01.937056 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c65192f-a05a-45e2-9d56-33ddeb265374-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4c65192f-a05a-45e2-9d56-33ddeb265374\") " pod="openstack/ceilometer-0" Jan 26 17:09:01 crc kubenswrapper[4754]: I0126 17:09:01.937113 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4c65192f-a05a-45e2-9d56-33ddeb265374-run-httpd\") pod \"ceilometer-0\" (UID: \"4c65192f-a05a-45e2-9d56-33ddeb265374\") " pod="openstack/ceilometer-0" Jan 26 17:09:02 crc kubenswrapper[4754]: I0126 17:09:02.039162 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c65192f-a05a-45e2-9d56-33ddeb265374-config-data\") pod \"ceilometer-0\" (UID: \"4c65192f-a05a-45e2-9d56-33ddeb265374\") " pod="openstack/ceilometer-0" Jan 26 17:09:02 crc kubenswrapper[4754]: I0126 17:09:02.039266 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4c65192f-a05a-45e2-9d56-33ddeb265374-scripts\") pod \"ceilometer-0\" (UID: \"4c65192f-a05a-45e2-9d56-33ddeb265374\") " pod="openstack/ceilometer-0" Jan 26 17:09:02 crc kubenswrapper[4754]: I0126 17:09:02.039310 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4vqll\" (UniqueName: \"kubernetes.io/projected/4c65192f-a05a-45e2-9d56-33ddeb265374-kube-api-access-4vqll\") pod \"ceilometer-0\" (UID: \"4c65192f-a05a-45e2-9d56-33ddeb265374\") " pod="openstack/ceilometer-0" Jan 26 17:09:02 crc kubenswrapper[4754]: I0126 17:09:02.039327 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4c65192f-a05a-45e2-9d56-33ddeb265374-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4c65192f-a05a-45e2-9d56-33ddeb265374\") " pod="openstack/ceilometer-0" Jan 26 17:09:02 crc kubenswrapper[4754]: I0126 17:09:02.039349 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c65192f-a05a-45e2-9d56-33ddeb265374-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4c65192f-a05a-45e2-9d56-33ddeb265374\") " pod="openstack/ceilometer-0" Jan 26 17:09:02 crc kubenswrapper[4754]: I0126 17:09:02.039374 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4c65192f-a05a-45e2-9d56-33ddeb265374-run-httpd\") pod \"ceilometer-0\" (UID: \"4c65192f-a05a-45e2-9d56-33ddeb265374\") " pod="openstack/ceilometer-0" Jan 26 17:09:02 crc kubenswrapper[4754]: I0126 17:09:02.039401 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4c65192f-a05a-45e2-9d56-33ddeb265374-log-httpd\") pod \"ceilometer-0\" (UID: \"4c65192f-a05a-45e2-9d56-33ddeb265374\") " pod="openstack/ceilometer-0" Jan 26 17:09:02 crc kubenswrapper[4754]: I0126 17:09:02.039797 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4c65192f-a05a-45e2-9d56-33ddeb265374-log-httpd\") pod \"ceilometer-0\" (UID: \"4c65192f-a05a-45e2-9d56-33ddeb265374\") " pod="openstack/ceilometer-0" Jan 26 17:09:02 crc kubenswrapper[4754]: I0126 17:09:02.040146 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4c65192f-a05a-45e2-9d56-33ddeb265374-run-httpd\") pod \"ceilometer-0\" (UID: \"4c65192f-a05a-45e2-9d56-33ddeb265374\") " pod="openstack/ceilometer-0" Jan 26 17:09:02 crc kubenswrapper[4754]: I0126 17:09:02.044920 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c65192f-a05a-45e2-9d56-33ddeb265374-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4c65192f-a05a-45e2-9d56-33ddeb265374\") " pod="openstack/ceilometer-0" Jan 26 17:09:02 crc kubenswrapper[4754]: I0126 17:09:02.045562 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c65192f-a05a-45e2-9d56-33ddeb265374-config-data\") pod \"ceilometer-0\" (UID: \"4c65192f-a05a-45e2-9d56-33ddeb265374\") " pod="openstack/ceilometer-0" Jan 26 17:09:02 crc kubenswrapper[4754]: I0126 17:09:02.046175 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4c65192f-a05a-45e2-9d56-33ddeb265374-scripts\") pod \"ceilometer-0\" (UID: \"4c65192f-a05a-45e2-9d56-33ddeb265374\") " pod="openstack/ceilometer-0" Jan 26 17:09:02 crc kubenswrapper[4754]: I0126 17:09:02.062580 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4c65192f-a05a-45e2-9d56-33ddeb265374-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4c65192f-a05a-45e2-9d56-33ddeb265374\") " pod="openstack/ceilometer-0" Jan 26 17:09:02 crc kubenswrapper[4754]: I0126 17:09:02.069744 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4vqll\" (UniqueName: \"kubernetes.io/projected/4c65192f-a05a-45e2-9d56-33ddeb265374-kube-api-access-4vqll\") pod \"ceilometer-0\" (UID: \"4c65192f-a05a-45e2-9d56-33ddeb265374\") " pod="openstack/ceilometer-0" Jan 26 17:09:02 crc kubenswrapper[4754]: I0126 17:09:02.159355 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 26 17:09:03 crc kubenswrapper[4754]: I0126 17:09:03.776652 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d84d5918-51f5-4131-9caf-9380004a0ca5" path="/var/lib/kubelet/pods/d84d5918-51f5-4131-9caf-9380004a0ca5/volumes" Jan 26 17:09:04 crc kubenswrapper[4754]: I0126 17:09:04.481583 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-6c774f65c8-h55rj" Jan 26 17:09:05 crc kubenswrapper[4754]: I0126 17:09:05.874872 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-586bdc5f9-lqn4x" Jan 26 17:09:05 crc kubenswrapper[4754]: I0126 17:09:05.946202 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-rjqkz"] Jan 26 17:09:05 crc kubenswrapper[4754]: I0126 17:09:05.946460 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-785d8bcb8c-rjqkz" podUID="bcc405ec-8acd-4fd7-bb9e-f3ab471d03ec" containerName="dnsmasq-dns" containerID="cri-o://a08ed9215a80b3818664f82516ebea1b8ce063efca461fceaef02f735167f849" gracePeriod=10 Jan 26 17:09:07 crc kubenswrapper[4754]: I0126 17:09:07.129415 4754 patch_prober.go:28] interesting pod/machine-config-daemon-x65wv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 26 17:09:07 crc kubenswrapper[4754]: I0126 17:09:07.129768 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 26 17:09:07 crc kubenswrapper[4754]: I0126 17:09:07.138252 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Jan 26 17:09:07 crc kubenswrapper[4754]: I0126 17:09:07.139398 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Jan 26 17:09:07 crc kubenswrapper[4754]: I0126 17:09:07.140894 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Jan 26 17:09:07 crc kubenswrapper[4754]: I0126 17:09:07.141807 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Jan 26 17:09:07 crc kubenswrapper[4754]: I0126 17:09:07.142231 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-qkzpl" Jan 26 17:09:07 crc kubenswrapper[4754]: I0126 17:09:07.154526 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Jan 26 17:09:07 crc kubenswrapper[4754]: I0126 17:09:07.231111 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4mntt\" (UniqueName: \"kubernetes.io/projected/19dee3c5-51dc-46be-8454-9c10c76b3655-kube-api-access-4mntt\") pod \"openstackclient\" (UID: \"19dee3c5-51dc-46be-8454-9c10c76b3655\") " pod="openstack/openstackclient" Jan 26 17:09:07 crc kubenswrapper[4754]: I0126 17:09:07.231149 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19dee3c5-51dc-46be-8454-9c10c76b3655-combined-ca-bundle\") pod \"openstackclient\" (UID: \"19dee3c5-51dc-46be-8454-9c10c76b3655\") " pod="openstack/openstackclient" Jan 26 17:09:07 crc kubenswrapper[4754]: I0126 17:09:07.231241 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/19dee3c5-51dc-46be-8454-9c10c76b3655-openstack-config-secret\") pod \"openstackclient\" (UID: \"19dee3c5-51dc-46be-8454-9c10c76b3655\") " pod="openstack/openstackclient" Jan 26 17:09:07 crc kubenswrapper[4754]: I0126 17:09:07.231265 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/19dee3c5-51dc-46be-8454-9c10c76b3655-openstack-config\") pod \"openstackclient\" (UID: \"19dee3c5-51dc-46be-8454-9c10c76b3655\") " pod="openstack/openstackclient" Jan 26 17:09:07 crc kubenswrapper[4754]: I0126 17:09:07.332775 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/19dee3c5-51dc-46be-8454-9c10c76b3655-openstack-config-secret\") pod \"openstackclient\" (UID: \"19dee3c5-51dc-46be-8454-9c10c76b3655\") " pod="openstack/openstackclient" Jan 26 17:09:07 crc kubenswrapper[4754]: I0126 17:09:07.332841 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/19dee3c5-51dc-46be-8454-9c10c76b3655-openstack-config\") pod \"openstackclient\" (UID: \"19dee3c5-51dc-46be-8454-9c10c76b3655\") " pod="openstack/openstackclient" Jan 26 17:09:07 crc kubenswrapper[4754]: I0126 17:09:07.333120 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4mntt\" (UniqueName: \"kubernetes.io/projected/19dee3c5-51dc-46be-8454-9c10c76b3655-kube-api-access-4mntt\") pod \"openstackclient\" (UID: \"19dee3c5-51dc-46be-8454-9c10c76b3655\") " pod="openstack/openstackclient" Jan 26 17:09:07 crc kubenswrapper[4754]: I0126 17:09:07.333160 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19dee3c5-51dc-46be-8454-9c10c76b3655-combined-ca-bundle\") pod \"openstackclient\" (UID: \"19dee3c5-51dc-46be-8454-9c10c76b3655\") " pod="openstack/openstackclient" Jan 26 17:09:07 crc kubenswrapper[4754]: I0126 17:09:07.334392 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/19dee3c5-51dc-46be-8454-9c10c76b3655-openstack-config\") pod \"openstackclient\" (UID: \"19dee3c5-51dc-46be-8454-9c10c76b3655\") " pod="openstack/openstackclient" Jan 26 17:09:07 crc kubenswrapper[4754]: I0126 17:09:07.338814 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19dee3c5-51dc-46be-8454-9c10c76b3655-combined-ca-bundle\") pod \"openstackclient\" (UID: \"19dee3c5-51dc-46be-8454-9c10c76b3655\") " pod="openstack/openstackclient" Jan 26 17:09:07 crc kubenswrapper[4754]: I0126 17:09:07.339117 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/19dee3c5-51dc-46be-8454-9c10c76b3655-openstack-config-secret\") pod \"openstackclient\" (UID: \"19dee3c5-51dc-46be-8454-9c10c76b3655\") " pod="openstack/openstackclient" Jan 26 17:09:07 crc kubenswrapper[4754]: I0126 17:09:07.354882 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4mntt\" (UniqueName: \"kubernetes.io/projected/19dee3c5-51dc-46be-8454-9c10c76b3655-kube-api-access-4mntt\") pod \"openstackclient\" (UID: \"19dee3c5-51dc-46be-8454-9c10c76b3655\") " pod="openstack/openstackclient" Jan 26 17:09:07 crc kubenswrapper[4754]: I0126 17:09:07.456371 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Jan 26 17:09:08 crc kubenswrapper[4754]: I0126 17:09:08.723335 4754 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-785d8bcb8c-rjqkz" podUID="bcc405ec-8acd-4fd7-bb9e-f3ab471d03ec" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.149:5353: connect: connection refused" Jan 26 17:09:09 crc kubenswrapper[4754]: I0126 17:09:09.617374 4754 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-5d7f8b89fd-jwf5r" podUID="0dee3ce6-9657-4b8a-9034-7c744db2d209" containerName="barbican-api" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 26 17:09:09 crc kubenswrapper[4754]: I0126 17:09:09.619583 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-5d7f8b89fd-jwf5r" podUID="0dee3ce6-9657-4b8a-9034-7c744db2d209" containerName="barbican-api-log" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 26 17:09:09 crc kubenswrapper[4754]: I0126 17:09:09.620322 4754 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-5d7f8b89fd-jwf5r" podUID="0dee3ce6-9657-4b8a-9034-7c744db2d209" containerName="barbican-api-log" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 26 17:09:09 crc kubenswrapper[4754]: I0126 17:09:09.634518 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-5d4bd978-rj87r" podUID="8c7859b2-be6e-49c3-8392-47ed649a1f68" containerName="barbican-api" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 26 17:09:09 crc kubenswrapper[4754]: I0126 17:09:09.634570 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-5d4bd978-rj87r" podUID="8c7859b2-be6e-49c3-8392-47ed649a1f68" containerName="barbican-api-log" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 26 17:09:09 crc kubenswrapper[4754]: I0126 17:09:09.681985 4754 scope.go:117] "RemoveContainer" containerID="2b6a9e46662327f03d044bfee7f9926d7e4287b3bd44ae06415f5f1bf5a68abc" Jan 26 17:09:09 crc kubenswrapper[4754]: I0126 17:09:09.764049 4754 generic.go:334] "Generic (PLEG): container finished" podID="bcc405ec-8acd-4fd7-bb9e-f3ab471d03ec" containerID="a08ed9215a80b3818664f82516ebea1b8ce063efca461fceaef02f735167f849" exitCode=0 Jan 26 17:09:09 crc kubenswrapper[4754]: I0126 17:09:09.764130 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-rjqkz" event={"ID":"bcc405ec-8acd-4fd7-bb9e-f3ab471d03ec","Type":"ContainerDied","Data":"a08ed9215a80b3818664f82516ebea1b8ce063efca461fceaef02f735167f849"} Jan 26 17:09:09 crc kubenswrapper[4754]: I0126 17:09:09.778534 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-85f95cb57b-99z67" event={"ID":"ea911a5c-f353-4da9-9900-bfcbb234a377","Type":"ContainerDied","Data":"f0b91df38c4a01b5f49a3da68f73e9163c1c9e925d047455d811efa5231a8507"} Jan 26 17:09:09 crc kubenswrapper[4754]: I0126 17:09:09.778576 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f0b91df38c4a01b5f49a3da68f73e9163c1c9e925d047455d811efa5231a8507" Jan 26 17:09:09 crc kubenswrapper[4754]: I0126 17:09:09.847863 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-85f95cb57b-99z67" Jan 26 17:09:09 crc kubenswrapper[4754]: I0126 17:09:09.984186 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ea911a5c-f353-4da9-9900-bfcbb234a377-scripts\") pod \"ea911a5c-f353-4da9-9900-bfcbb234a377\" (UID: \"ea911a5c-f353-4da9-9900-bfcbb234a377\") " Jan 26 17:09:09 crc kubenswrapper[4754]: I0126 17:09:09.984242 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tl5tm\" (UniqueName: \"kubernetes.io/projected/ea911a5c-f353-4da9-9900-bfcbb234a377-kube-api-access-tl5tm\") pod \"ea911a5c-f353-4da9-9900-bfcbb234a377\" (UID: \"ea911a5c-f353-4da9-9900-bfcbb234a377\") " Jan 26 17:09:09 crc kubenswrapper[4754]: I0126 17:09:09.984258 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ea911a5c-f353-4da9-9900-bfcbb234a377-config-data\") pod \"ea911a5c-f353-4da9-9900-bfcbb234a377\" (UID: \"ea911a5c-f353-4da9-9900-bfcbb234a377\") " Jan 26 17:09:09 crc kubenswrapper[4754]: I0126 17:09:09.984282 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/ea911a5c-f353-4da9-9900-bfcbb234a377-horizon-tls-certs\") pod \"ea911a5c-f353-4da9-9900-bfcbb234a377\" (UID: \"ea911a5c-f353-4da9-9900-bfcbb234a377\") " Jan 26 17:09:09 crc kubenswrapper[4754]: I0126 17:09:09.984371 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ea911a5c-f353-4da9-9900-bfcbb234a377-logs\") pod \"ea911a5c-f353-4da9-9900-bfcbb234a377\" (UID: \"ea911a5c-f353-4da9-9900-bfcbb234a377\") " Jan 26 17:09:09 crc kubenswrapper[4754]: I0126 17:09:09.984444 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/ea911a5c-f353-4da9-9900-bfcbb234a377-horizon-secret-key\") pod \"ea911a5c-f353-4da9-9900-bfcbb234a377\" (UID: \"ea911a5c-f353-4da9-9900-bfcbb234a377\") " Jan 26 17:09:09 crc kubenswrapper[4754]: I0126 17:09:09.984523 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ea911a5c-f353-4da9-9900-bfcbb234a377-combined-ca-bundle\") pod \"ea911a5c-f353-4da9-9900-bfcbb234a377\" (UID: \"ea911a5c-f353-4da9-9900-bfcbb234a377\") " Jan 26 17:09:09 crc kubenswrapper[4754]: I0126 17:09:09.986580 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ea911a5c-f353-4da9-9900-bfcbb234a377-logs" (OuterVolumeSpecName: "logs") pod "ea911a5c-f353-4da9-9900-bfcbb234a377" (UID: "ea911a5c-f353-4da9-9900-bfcbb234a377"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:09:09 crc kubenswrapper[4754]: I0126 17:09:09.992946 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ea911a5c-f353-4da9-9900-bfcbb234a377-kube-api-access-tl5tm" (OuterVolumeSpecName: "kube-api-access-tl5tm") pod "ea911a5c-f353-4da9-9900-bfcbb234a377" (UID: "ea911a5c-f353-4da9-9900-bfcbb234a377"). InnerVolumeSpecName "kube-api-access-tl5tm". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:09:09 crc kubenswrapper[4754]: I0126 17:09:09.993282 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ea911a5c-f353-4da9-9900-bfcbb234a377-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "ea911a5c-f353-4da9-9900-bfcbb234a377" (UID: "ea911a5c-f353-4da9-9900-bfcbb234a377"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:09:10 crc kubenswrapper[4754]: I0126 17:09:10.087076 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-rjqkz" Jan 26 17:09:10 crc kubenswrapper[4754]: I0126 17:09:10.090139 4754 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ea911a5c-f353-4da9-9900-bfcbb234a377-logs\") on node \"crc\" DevicePath \"\"" Jan 26 17:09:10 crc kubenswrapper[4754]: I0126 17:09:10.090201 4754 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/ea911a5c-f353-4da9-9900-bfcbb234a377-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Jan 26 17:09:10 crc kubenswrapper[4754]: I0126 17:09:10.090218 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tl5tm\" (UniqueName: \"kubernetes.io/projected/ea911a5c-f353-4da9-9900-bfcbb234a377-kube-api-access-tl5tm\") on node \"crc\" DevicePath \"\"" Jan 26 17:09:10 crc kubenswrapper[4754]: I0126 17:09:10.165965 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ea911a5c-f353-4da9-9900-bfcbb234a377-scripts" (OuterVolumeSpecName: "scripts") pod "ea911a5c-f353-4da9-9900-bfcbb234a377" (UID: "ea911a5c-f353-4da9-9900-bfcbb234a377"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:09:10 crc kubenswrapper[4754]: I0126 17:09:10.171037 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ea911a5c-f353-4da9-9900-bfcbb234a377-config-data" (OuterVolumeSpecName: "config-data") pod "ea911a5c-f353-4da9-9900-bfcbb234a377" (UID: "ea911a5c-f353-4da9-9900-bfcbb234a377"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:09:10 crc kubenswrapper[4754]: I0126 17:09:10.180354 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ea911a5c-f353-4da9-9900-bfcbb234a377-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ea911a5c-f353-4da9-9900-bfcbb234a377" (UID: "ea911a5c-f353-4da9-9900-bfcbb234a377"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:09:10 crc kubenswrapper[4754]: I0126 17:09:10.190872 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bcc405ec-8acd-4fd7-bb9e-f3ab471d03ec-ovsdbserver-sb\") pod \"bcc405ec-8acd-4fd7-bb9e-f3ab471d03ec\" (UID: \"bcc405ec-8acd-4fd7-bb9e-f3ab471d03ec\") " Jan 26 17:09:10 crc kubenswrapper[4754]: I0126 17:09:10.191109 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nd7x6\" (UniqueName: \"kubernetes.io/projected/bcc405ec-8acd-4fd7-bb9e-f3ab471d03ec-kube-api-access-nd7x6\") pod \"bcc405ec-8acd-4fd7-bb9e-f3ab471d03ec\" (UID: \"bcc405ec-8acd-4fd7-bb9e-f3ab471d03ec\") " Jan 26 17:09:10 crc kubenswrapper[4754]: I0126 17:09:10.191180 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bcc405ec-8acd-4fd7-bb9e-f3ab471d03ec-dns-swift-storage-0\") pod \"bcc405ec-8acd-4fd7-bb9e-f3ab471d03ec\" (UID: \"bcc405ec-8acd-4fd7-bb9e-f3ab471d03ec\") " Jan 26 17:09:10 crc kubenswrapper[4754]: I0126 17:09:10.191414 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bcc405ec-8acd-4fd7-bb9e-f3ab471d03ec-dns-svc\") pod \"bcc405ec-8acd-4fd7-bb9e-f3ab471d03ec\" (UID: \"bcc405ec-8acd-4fd7-bb9e-f3ab471d03ec\") " Jan 26 17:09:10 crc kubenswrapper[4754]: I0126 17:09:10.191446 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bcc405ec-8acd-4fd7-bb9e-f3ab471d03ec-ovsdbserver-nb\") pod \"bcc405ec-8acd-4fd7-bb9e-f3ab471d03ec\" (UID: \"bcc405ec-8acd-4fd7-bb9e-f3ab471d03ec\") " Jan 26 17:09:10 crc kubenswrapper[4754]: I0126 17:09:10.191519 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bcc405ec-8acd-4fd7-bb9e-f3ab471d03ec-config\") pod \"bcc405ec-8acd-4fd7-bb9e-f3ab471d03ec\" (UID: \"bcc405ec-8acd-4fd7-bb9e-f3ab471d03ec\") " Jan 26 17:09:10 crc kubenswrapper[4754]: I0126 17:09:10.192054 4754 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ea911a5c-f353-4da9-9900-bfcbb234a377-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 26 17:09:10 crc kubenswrapper[4754]: I0126 17:09:10.192077 4754 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ea911a5c-f353-4da9-9900-bfcbb234a377-scripts\") on node \"crc\" DevicePath \"\"" Jan 26 17:09:10 crc kubenswrapper[4754]: I0126 17:09:10.192091 4754 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ea911a5c-f353-4da9-9900-bfcbb234a377-config-data\") on node \"crc\" DevicePath \"\"" Jan 26 17:09:10 crc kubenswrapper[4754]: I0126 17:09:10.214099 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bcc405ec-8acd-4fd7-bb9e-f3ab471d03ec-kube-api-access-nd7x6" (OuterVolumeSpecName: "kube-api-access-nd7x6") pod "bcc405ec-8acd-4fd7-bb9e-f3ab471d03ec" (UID: "bcc405ec-8acd-4fd7-bb9e-f3ab471d03ec"). InnerVolumeSpecName "kube-api-access-nd7x6". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:09:10 crc kubenswrapper[4754]: I0126 17:09:10.279634 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Jan 26 17:09:10 crc kubenswrapper[4754]: I0126 17:09:10.281120 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ea911a5c-f353-4da9-9900-bfcbb234a377-horizon-tls-certs" (OuterVolumeSpecName: "horizon-tls-certs") pod "ea911a5c-f353-4da9-9900-bfcbb234a377" (UID: "ea911a5c-f353-4da9-9900-bfcbb234a377"). InnerVolumeSpecName "horizon-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:09:10 crc kubenswrapper[4754]: I0126 17:09:10.300870 4754 reconciler_common.go:293] "Volume detached for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/ea911a5c-f353-4da9-9900-bfcbb234a377-horizon-tls-certs\") on node \"crc\" DevicePath \"\"" Jan 26 17:09:10 crc kubenswrapper[4754]: I0126 17:09:10.300902 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nd7x6\" (UniqueName: \"kubernetes.io/projected/bcc405ec-8acd-4fd7-bb9e-f3ab471d03ec-kube-api-access-nd7x6\") on node \"crc\" DevicePath \"\"" Jan 26 17:09:10 crc kubenswrapper[4754]: I0126 17:09:10.315664 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bcc405ec-8acd-4fd7-bb9e-f3ab471d03ec-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "bcc405ec-8acd-4fd7-bb9e-f3ab471d03ec" (UID: "bcc405ec-8acd-4fd7-bb9e-f3ab471d03ec"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:09:10 crc kubenswrapper[4754]: I0126 17:09:10.362116 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bcc405ec-8acd-4fd7-bb9e-f3ab471d03ec-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "bcc405ec-8acd-4fd7-bb9e-f3ab471d03ec" (UID: "bcc405ec-8acd-4fd7-bb9e-f3ab471d03ec"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:09:10 crc kubenswrapper[4754]: I0126 17:09:10.364477 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bcc405ec-8acd-4fd7-bb9e-f3ab471d03ec-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "bcc405ec-8acd-4fd7-bb9e-f3ab471d03ec" (UID: "bcc405ec-8acd-4fd7-bb9e-f3ab471d03ec"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:09:10 crc kubenswrapper[4754]: I0126 17:09:10.370077 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bcc405ec-8acd-4fd7-bb9e-f3ab471d03ec-config" (OuterVolumeSpecName: "config") pod "bcc405ec-8acd-4fd7-bb9e-f3ab471d03ec" (UID: "bcc405ec-8acd-4fd7-bb9e-f3ab471d03ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:09:10 crc kubenswrapper[4754]: I0126 17:09:10.389056 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-5d4bd978-rj87r" Jan 26 17:09:10 crc kubenswrapper[4754]: I0126 17:09:10.403939 4754 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bcc405ec-8acd-4fd7-bb9e-f3ab471d03ec-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Jan 26 17:09:10 crc kubenswrapper[4754]: I0126 17:09:10.403972 4754 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bcc405ec-8acd-4fd7-bb9e-f3ab471d03ec-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Jan 26 17:09:10 crc kubenswrapper[4754]: I0126 17:09:10.403986 4754 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bcc405ec-8acd-4fd7-bb9e-f3ab471d03ec-config\") on node \"crc\" DevicePath \"\"" Jan 26 17:09:10 crc kubenswrapper[4754]: I0126 17:09:10.403998 4754 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bcc405ec-8acd-4fd7-bb9e-f3ab471d03ec-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Jan 26 17:09:10 crc kubenswrapper[4754]: I0126 17:09:10.412816 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bcc405ec-8acd-4fd7-bb9e-f3ab471d03ec-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "bcc405ec-8acd-4fd7-bb9e-f3ab471d03ec" (UID: "bcc405ec-8acd-4fd7-bb9e-f3ab471d03ec"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:09:10 crc kubenswrapper[4754]: I0126 17:09:10.417101 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Jan 26 17:09:10 crc kubenswrapper[4754]: W0126 17:09:10.417838 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod19dee3c5_51dc_46be_8454_9c10c76b3655.slice/crio-269c6e802b6cb50034ea9b2f8fdf253bc950f98743a26d02c702290cd9f67b88 WatchSource:0}: Error finding container 269c6e802b6cb50034ea9b2f8fdf253bc950f98743a26d02c702290cd9f67b88: Status 404 returned error can't find the container with id 269c6e802b6cb50034ea9b2f8fdf253bc950f98743a26d02c702290cd9f67b88 Jan 26 17:09:10 crc kubenswrapper[4754]: I0126 17:09:10.456651 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-5d4bd978-rj87r" Jan 26 17:09:10 crc kubenswrapper[4754]: I0126 17:09:10.505810 4754 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bcc405ec-8acd-4fd7-bb9e-f3ab471d03ec-dns-svc\") on node \"crc\" DevicePath \"\"" Jan 26 17:09:10 crc kubenswrapper[4754]: I0126 17:09:10.535336 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-5d7f8b89fd-jwf5r"] Jan 26 17:09:10 crc kubenswrapper[4754]: I0126 17:09:10.535945 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-5d7f8b89fd-jwf5r" podUID="0dee3ce6-9657-4b8a-9034-7c744db2d209" containerName="barbican-api-log" containerID="cri-o://a5dcc43e1e1644072f8d0a7424cbbc5cfda3573dc5d056f0c4781edb61fa886e" gracePeriod=30 Jan 26 17:09:10 crc kubenswrapper[4754]: I0126 17:09:10.537242 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-5d7f8b89fd-jwf5r" podUID="0dee3ce6-9657-4b8a-9034-7c744db2d209" containerName="barbican-api" containerID="cri-o://6401776521d0c72d1ea925c21389233c039edc6ed1e4450e98139bea1afe8ec3" gracePeriod=30 Jan 26 17:09:10 crc kubenswrapper[4754]: I0126 17:09:10.549101 4754 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-5d7f8b89fd-jwf5r" podUID="0dee3ce6-9657-4b8a-9034-7c744db2d209" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.164:9311/healthcheck\": EOF" Jan 26 17:09:10 crc kubenswrapper[4754]: I0126 17:09:10.549145 4754 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-5d7f8b89fd-jwf5r" podUID="0dee3ce6-9657-4b8a-9034-7c744db2d209" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.164:9311/healthcheck\": EOF" Jan 26 17:09:10 crc kubenswrapper[4754]: I0126 17:09:10.549439 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-5d7f8b89fd-jwf5r" podUID="0dee3ce6-9657-4b8a-9034-7c744db2d209" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.164:9311/healthcheck\": EOF" Jan 26 17:09:10 crc kubenswrapper[4754]: I0126 17:09:10.549566 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-5d7f8b89fd-jwf5r" podUID="0dee3ce6-9657-4b8a-9034-7c744db2d209" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.164:9311/healthcheck\": EOF" Jan 26 17:09:10 crc kubenswrapper[4754]: I0126 17:09:10.853093 4754 generic.go:334] "Generic (PLEG): container finished" podID="0dee3ce6-9657-4b8a-9034-7c744db2d209" containerID="a5dcc43e1e1644072f8d0a7424cbbc5cfda3573dc5d056f0c4781edb61fa886e" exitCode=143 Jan 26 17:09:10 crc kubenswrapper[4754]: I0126 17:09:10.853321 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5d7f8b89fd-jwf5r" event={"ID":"0dee3ce6-9657-4b8a-9034-7c744db2d209","Type":"ContainerDied","Data":"a5dcc43e1e1644072f8d0a7424cbbc5cfda3573dc5d056f0c4781edb61fa886e"} Jan 26 17:09:10 crc kubenswrapper[4754]: I0126 17:09:10.860722 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"19dee3c5-51dc-46be-8454-9c10c76b3655","Type":"ContainerStarted","Data":"269c6e802b6cb50034ea9b2f8fdf253bc950f98743a26d02c702290cd9f67b88"} Jan 26 17:09:10 crc kubenswrapper[4754]: I0126 17:09:10.864964 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-rjqkz" event={"ID":"bcc405ec-8acd-4fd7-bb9e-f3ab471d03ec","Type":"ContainerDied","Data":"7dae99ce358a95b74a1485097aa09927ce38b696a146189337145849541693dd"} Jan 26 17:09:10 crc kubenswrapper[4754]: I0126 17:09:10.865032 4754 scope.go:117] "RemoveContainer" containerID="a08ed9215a80b3818664f82516ebea1b8ce063efca461fceaef02f735167f849" Jan 26 17:09:10 crc kubenswrapper[4754]: I0126 17:09:10.865222 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-rjqkz" Jan 26 17:09:10 crc kubenswrapper[4754]: I0126 17:09:10.876344 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-99649c64-sldzz" event={"ID":"0b004ba0-01f6-49f9-8d8d-56033dcd3533","Type":"ContainerStarted","Data":"6944798c33048f6aac35394a9094eb1a51faa90196d6965984c56de05b87c692"} Jan 26 17:09:10 crc kubenswrapper[4754]: I0126 17:09:10.898356 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-5677764859-tswcr" event={"ID":"d64625ff-6586-4c80-a720-b2febd49a966","Type":"ContainerStarted","Data":"5687276b7c0f8024033632f9e651cd93e780c79ea31419e1c20c382bb1d11d9e"} Jan 26 17:09:10 crc kubenswrapper[4754]: I0126 17:09:10.909348 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4c65192f-a05a-45e2-9d56-33ddeb265374","Type":"ContainerStarted","Data":"9283c476059ebf15ee0cfb420c7cb60dc5112b1663c3e2672d5ce9b1cc7cd9f9"} Jan 26 17:09:10 crc kubenswrapper[4754]: I0126 17:09:10.909448 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-85f95cb57b-99z67" Jan 26 17:09:11 crc kubenswrapper[4754]: I0126 17:09:11.060853 4754 scope.go:117] "RemoveContainer" containerID="ec3ffb21d050781a4c93354e0bc93031a4a8adfdc941863e27e609e89907b997" Jan 26 17:09:11 crc kubenswrapper[4754]: I0126 17:09:11.084299 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-rjqkz"] Jan 26 17:09:11 crc kubenswrapper[4754]: I0126 17:09:11.095788 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-rjqkz"] Jan 26 17:09:11 crc kubenswrapper[4754]: I0126 17:09:11.106938 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-85f95cb57b-99z67"] Jan 26 17:09:11 crc kubenswrapper[4754]: I0126 17:09:11.115692 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-85f95cb57b-99z67"] Jan 26 17:09:11 crc kubenswrapper[4754]: I0126 17:09:11.545859 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-577b5f64f6-qwc6n" Jan 26 17:09:11 crc kubenswrapper[4754]: I0126 17:09:11.582760 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-577b5f64f6-qwc6n" Jan 26 17:09:11 crc kubenswrapper[4754]: I0126 17:09:11.788856 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bcc405ec-8acd-4fd7-bb9e-f3ab471d03ec" path="/var/lib/kubelet/pods/bcc405ec-8acd-4fd7-bb9e-f3ab471d03ec/volumes" Jan 26 17:09:11 crc kubenswrapper[4754]: I0126 17:09:11.790550 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ea911a5c-f353-4da9-9900-bfcbb234a377" path="/var/lib/kubelet/pods/ea911a5c-f353-4da9-9900-bfcbb234a377/volumes" Jan 26 17:09:11 crc kubenswrapper[4754]: I0126 17:09:11.962292 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-99649c64-sldzz" event={"ID":"0b004ba0-01f6-49f9-8d8d-56033dcd3533","Type":"ContainerStarted","Data":"f86b1926bfaed987ee68274e092a2b0196105f0cd66a5b6cf51beab6afb7ff2b"} Jan 26 17:09:11 crc kubenswrapper[4754]: I0126 17:09:11.975342 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-5677764859-tswcr" event={"ID":"d64625ff-6586-4c80-a720-b2febd49a966","Type":"ContainerStarted","Data":"984e0332788b38cfe426a184be0a318c0e5d65716f195f20053ca2d4627e00be"} Jan 26 17:09:11 crc kubenswrapper[4754]: I0126 17:09:11.993528 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4c65192f-a05a-45e2-9d56-33ddeb265374","Type":"ContainerStarted","Data":"374dc6158046ff5aeb4527bdd32fd9636f239a18d5ba35caa78a117156d7fb06"} Jan 26 17:09:12 crc kubenswrapper[4754]: I0126 17:09:12.005407 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-99649c64-sldzz" podStartSLOduration=3.710653908 podStartE2EDuration="17.005383859s" podCreationTimestamp="2026-01-26 17:08:55 +0000 UTC" firstStartedPulling="2026-01-26 17:08:56.440146001 +0000 UTC m=+1302.964326435" lastFinishedPulling="2026-01-26 17:09:09.734875952 +0000 UTC m=+1316.259056386" observedRunningTime="2026-01-26 17:09:11.981157335 +0000 UTC m=+1318.505337799" watchObservedRunningTime="2026-01-26 17:09:12.005383859 +0000 UTC m=+1318.529564293" Jan 26 17:09:12 crc kubenswrapper[4754]: I0126 17:09:12.016174 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-5677764859-tswcr" podStartSLOduration=3.636085437 podStartE2EDuration="17.016153991s" podCreationTimestamp="2026-01-26 17:08:55 +0000 UTC" firstStartedPulling="2026-01-26 17:08:56.315210951 +0000 UTC m=+1302.839391385" lastFinishedPulling="2026-01-26 17:09:09.695279505 +0000 UTC m=+1316.219459939" observedRunningTime="2026-01-26 17:09:11.998876809 +0000 UTC m=+1318.523057243" watchObservedRunningTime="2026-01-26 17:09:12.016153991 +0000 UTC m=+1318.540334425" Jan 26 17:09:12 crc kubenswrapper[4754]: I0126 17:09:12.079781 4754 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-85f95cb57b-99z67" podUID="ea911a5c-f353-4da9-9900-bfcbb234a377" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.154:8443/dashboard/auth/login/?next=/dashboard/\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Jan 26 17:09:13 crc kubenswrapper[4754]: I0126 17:09:13.043803 4754 generic.go:334] "Generic (PLEG): container finished" podID="ab3f008d-78ab-4b73-962d-4447f312c9fa" containerID="eda7e173767be2f94247fe84877d1dc7a95556f64d1cc275cf4ccf7dd0a1ae64" exitCode=0 Jan 26 17:09:13 crc kubenswrapper[4754]: I0126 17:09:13.044089 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-7hskr" event={"ID":"ab3f008d-78ab-4b73-962d-4447f312c9fa","Type":"ContainerDied","Data":"eda7e173767be2f94247fe84877d1dc7a95556f64d1cc275cf4ccf7dd0a1ae64"} Jan 26 17:09:13 crc kubenswrapper[4754]: I0126 17:09:13.059232 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4c65192f-a05a-45e2-9d56-33ddeb265374","Type":"ContainerStarted","Data":"0bd046d181549657eae96bff37d02dbcc14075bc5395685c218279cc1485d7bd"} Jan 26 17:09:14 crc kubenswrapper[4754]: I0126 17:09:14.070546 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4c65192f-a05a-45e2-9d56-33ddeb265374","Type":"ContainerStarted","Data":"3e9828e079b47eb4e66edc9599b5bdd713959f79d23385e7543f9bcb4e6ef25e"} Jan 26 17:09:14 crc kubenswrapper[4754]: I0126 17:09:14.505208 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-7hskr" Jan 26 17:09:14 crc kubenswrapper[4754]: I0126 17:09:14.613701 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ab3f008d-78ab-4b73-962d-4447f312c9fa-config-data\") pod \"ab3f008d-78ab-4b73-962d-4447f312c9fa\" (UID: \"ab3f008d-78ab-4b73-962d-4447f312c9fa\") " Jan 26 17:09:14 crc kubenswrapper[4754]: I0126 17:09:14.613788 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ab3f008d-78ab-4b73-962d-4447f312c9fa-scripts\") pod \"ab3f008d-78ab-4b73-962d-4447f312c9fa\" (UID: \"ab3f008d-78ab-4b73-962d-4447f312c9fa\") " Jan 26 17:09:14 crc kubenswrapper[4754]: I0126 17:09:14.613816 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5gg2w\" (UniqueName: \"kubernetes.io/projected/ab3f008d-78ab-4b73-962d-4447f312c9fa-kube-api-access-5gg2w\") pod \"ab3f008d-78ab-4b73-962d-4447f312c9fa\" (UID: \"ab3f008d-78ab-4b73-962d-4447f312c9fa\") " Jan 26 17:09:14 crc kubenswrapper[4754]: I0126 17:09:14.614771 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab3f008d-78ab-4b73-962d-4447f312c9fa-combined-ca-bundle\") pod \"ab3f008d-78ab-4b73-962d-4447f312c9fa\" (UID: \"ab3f008d-78ab-4b73-962d-4447f312c9fa\") " Jan 26 17:09:14 crc kubenswrapper[4754]: I0126 17:09:14.614832 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ab3f008d-78ab-4b73-962d-4447f312c9fa-db-sync-config-data\") pod \"ab3f008d-78ab-4b73-962d-4447f312c9fa\" (UID: \"ab3f008d-78ab-4b73-962d-4447f312c9fa\") " Jan 26 17:09:14 crc kubenswrapper[4754]: I0126 17:09:14.615064 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ab3f008d-78ab-4b73-962d-4447f312c9fa-etc-machine-id\") pod \"ab3f008d-78ab-4b73-962d-4447f312c9fa\" (UID: \"ab3f008d-78ab-4b73-962d-4447f312c9fa\") " Jan 26 17:09:14 crc kubenswrapper[4754]: I0126 17:09:14.615614 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ab3f008d-78ab-4b73-962d-4447f312c9fa-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "ab3f008d-78ab-4b73-962d-4447f312c9fa" (UID: "ab3f008d-78ab-4b73-962d-4447f312c9fa"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 17:09:14 crc kubenswrapper[4754]: I0126 17:09:14.620334 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ab3f008d-78ab-4b73-962d-4447f312c9fa-scripts" (OuterVolumeSpecName: "scripts") pod "ab3f008d-78ab-4b73-962d-4447f312c9fa" (UID: "ab3f008d-78ab-4b73-962d-4447f312c9fa"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:09:14 crc kubenswrapper[4754]: I0126 17:09:14.629856 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ab3f008d-78ab-4b73-962d-4447f312c9fa-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "ab3f008d-78ab-4b73-962d-4447f312c9fa" (UID: "ab3f008d-78ab-4b73-962d-4447f312c9fa"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:09:14 crc kubenswrapper[4754]: I0126 17:09:14.630088 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ab3f008d-78ab-4b73-962d-4447f312c9fa-kube-api-access-5gg2w" (OuterVolumeSpecName: "kube-api-access-5gg2w") pod "ab3f008d-78ab-4b73-962d-4447f312c9fa" (UID: "ab3f008d-78ab-4b73-962d-4447f312c9fa"). InnerVolumeSpecName "kube-api-access-5gg2w". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:09:14 crc kubenswrapper[4754]: I0126 17:09:14.655024 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ab3f008d-78ab-4b73-962d-4447f312c9fa-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ab3f008d-78ab-4b73-962d-4447f312c9fa" (UID: "ab3f008d-78ab-4b73-962d-4447f312c9fa"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:09:14 crc kubenswrapper[4754]: I0126 17:09:14.703737 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ab3f008d-78ab-4b73-962d-4447f312c9fa-config-data" (OuterVolumeSpecName: "config-data") pod "ab3f008d-78ab-4b73-962d-4447f312c9fa" (UID: "ab3f008d-78ab-4b73-962d-4447f312c9fa"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:09:14 crc kubenswrapper[4754]: I0126 17:09:14.716706 4754 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ab3f008d-78ab-4b73-962d-4447f312c9fa-scripts\") on node \"crc\" DevicePath \"\"" Jan 26 17:09:14 crc kubenswrapper[4754]: I0126 17:09:14.716756 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5gg2w\" (UniqueName: \"kubernetes.io/projected/ab3f008d-78ab-4b73-962d-4447f312c9fa-kube-api-access-5gg2w\") on node \"crc\" DevicePath \"\"" Jan 26 17:09:14 crc kubenswrapper[4754]: I0126 17:09:14.716770 4754 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab3f008d-78ab-4b73-962d-4447f312c9fa-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 26 17:09:14 crc kubenswrapper[4754]: I0126 17:09:14.716781 4754 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ab3f008d-78ab-4b73-962d-4447f312c9fa-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Jan 26 17:09:14 crc kubenswrapper[4754]: I0126 17:09:14.716792 4754 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ab3f008d-78ab-4b73-962d-4447f312c9fa-etc-machine-id\") on node \"crc\" DevicePath \"\"" Jan 26 17:09:14 crc kubenswrapper[4754]: I0126 17:09:14.716801 4754 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ab3f008d-78ab-4b73-962d-4447f312c9fa-config-data\") on node \"crc\" DevicePath \"\"" Jan 26 17:09:15 crc kubenswrapper[4754]: I0126 17:09:15.091243 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-7hskr" event={"ID":"ab3f008d-78ab-4b73-962d-4447f312c9fa","Type":"ContainerDied","Data":"b5a4699e6575064c5f4124d122fdc57ab2a2e7cc8cfcf396dd86d91929f75182"} Jan 26 17:09:15 crc kubenswrapper[4754]: I0126 17:09:15.091286 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b5a4699e6575064c5f4124d122fdc57ab2a2e7cc8cfcf396dd86d91929f75182" Jan 26 17:09:15 crc kubenswrapper[4754]: I0126 17:09:15.091283 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-7hskr" Jan 26 17:09:15 crc kubenswrapper[4754]: I0126 17:09:15.360820 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Jan 26 17:09:15 crc kubenswrapper[4754]: E0126 17:09:15.361456 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bcc405ec-8acd-4fd7-bb9e-f3ab471d03ec" containerName="dnsmasq-dns" Jan 26 17:09:15 crc kubenswrapper[4754]: I0126 17:09:15.361471 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="bcc405ec-8acd-4fd7-bb9e-f3ab471d03ec" containerName="dnsmasq-dns" Jan 26 17:09:15 crc kubenswrapper[4754]: E0126 17:09:15.361492 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea911a5c-f353-4da9-9900-bfcbb234a377" containerName="horizon" Jan 26 17:09:15 crc kubenswrapper[4754]: I0126 17:09:15.361499 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea911a5c-f353-4da9-9900-bfcbb234a377" containerName="horizon" Jan 26 17:09:15 crc kubenswrapper[4754]: E0126 17:09:15.361537 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bcc405ec-8acd-4fd7-bb9e-f3ab471d03ec" containerName="init" Jan 26 17:09:15 crc kubenswrapper[4754]: I0126 17:09:15.361558 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="bcc405ec-8acd-4fd7-bb9e-f3ab471d03ec" containerName="init" Jan 26 17:09:15 crc kubenswrapper[4754]: E0126 17:09:15.361591 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ab3f008d-78ab-4b73-962d-4447f312c9fa" containerName="cinder-db-sync" Jan 26 17:09:15 crc kubenswrapper[4754]: I0126 17:09:15.361606 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="ab3f008d-78ab-4b73-962d-4447f312c9fa" containerName="cinder-db-sync" Jan 26 17:09:15 crc kubenswrapper[4754]: E0126 17:09:15.361624 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea911a5c-f353-4da9-9900-bfcbb234a377" containerName="horizon-log" Jan 26 17:09:15 crc kubenswrapper[4754]: I0126 17:09:15.361630 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea911a5c-f353-4da9-9900-bfcbb234a377" containerName="horizon-log" Jan 26 17:09:15 crc kubenswrapper[4754]: I0126 17:09:15.361843 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="bcc405ec-8acd-4fd7-bb9e-f3ab471d03ec" containerName="dnsmasq-dns" Jan 26 17:09:15 crc kubenswrapper[4754]: I0126 17:09:15.361864 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="ea911a5c-f353-4da9-9900-bfcbb234a377" containerName="horizon-log" Jan 26 17:09:15 crc kubenswrapper[4754]: I0126 17:09:15.361875 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="ab3f008d-78ab-4b73-962d-4447f312c9fa" containerName="cinder-db-sync" Jan 26 17:09:15 crc kubenswrapper[4754]: I0126 17:09:15.361886 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="ea911a5c-f353-4da9-9900-bfcbb234a377" containerName="horizon" Jan 26 17:09:15 crc kubenswrapper[4754]: I0126 17:09:15.363740 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Jan 26 17:09:15 crc kubenswrapper[4754]: I0126 17:09:15.370656 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-bp2xk" Jan 26 17:09:15 crc kubenswrapper[4754]: I0126 17:09:15.370732 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Jan 26 17:09:15 crc kubenswrapper[4754]: I0126 17:09:15.371984 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Jan 26 17:09:15 crc kubenswrapper[4754]: I0126 17:09:15.372284 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Jan 26 17:09:15 crc kubenswrapper[4754]: I0126 17:09:15.384100 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Jan 26 17:09:15 crc kubenswrapper[4754]: I0126 17:09:15.505544 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-795f4db4bc-h7mcp"] Jan 26 17:09:15 crc kubenswrapper[4754]: I0126 17:09:15.507401 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-795f4db4bc-h7mcp" Jan 26 17:09:15 crc kubenswrapper[4754]: I0126 17:09:15.526632 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-795f4db4bc-h7mcp"] Jan 26 17:09:15 crc kubenswrapper[4754]: I0126 17:09:15.528768 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6f82d0f5-47a4-4584-9ae9-046882fa8f61-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"6f82d0f5-47a4-4584-9ae9-046882fa8f61\") " pod="openstack/cinder-scheduler-0" Jan 26 17:09:15 crc kubenswrapper[4754]: I0126 17:09:15.528910 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6f82d0f5-47a4-4584-9ae9-046882fa8f61-config-data\") pod \"cinder-scheduler-0\" (UID: \"6f82d0f5-47a4-4584-9ae9-046882fa8f61\") " pod="openstack/cinder-scheduler-0" Jan 26 17:09:15 crc kubenswrapper[4754]: I0126 17:09:15.528951 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9hzfn\" (UniqueName: \"kubernetes.io/projected/6f82d0f5-47a4-4584-9ae9-046882fa8f61-kube-api-access-9hzfn\") pod \"cinder-scheduler-0\" (UID: \"6f82d0f5-47a4-4584-9ae9-046882fa8f61\") " pod="openstack/cinder-scheduler-0" Jan 26 17:09:15 crc kubenswrapper[4754]: I0126 17:09:15.528985 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f82d0f5-47a4-4584-9ae9-046882fa8f61-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"6f82d0f5-47a4-4584-9ae9-046882fa8f61\") " pod="openstack/cinder-scheduler-0" Jan 26 17:09:15 crc kubenswrapper[4754]: I0126 17:09:15.529010 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6f82d0f5-47a4-4584-9ae9-046882fa8f61-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"6f82d0f5-47a4-4584-9ae9-046882fa8f61\") " pod="openstack/cinder-scheduler-0" Jan 26 17:09:15 crc kubenswrapper[4754]: I0126 17:09:15.529038 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6f82d0f5-47a4-4584-9ae9-046882fa8f61-scripts\") pod \"cinder-scheduler-0\" (UID: \"6f82d0f5-47a4-4584-9ae9-046882fa8f61\") " pod="openstack/cinder-scheduler-0" Jan 26 17:09:15 crc kubenswrapper[4754]: I0126 17:09:15.631560 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/977ff39d-2c55-438b-b957-62609c4c2ece-ovsdbserver-nb\") pod \"dnsmasq-dns-795f4db4bc-h7mcp\" (UID: \"977ff39d-2c55-438b-b957-62609c4c2ece\") " pod="openstack/dnsmasq-dns-795f4db4bc-h7mcp" Jan 26 17:09:15 crc kubenswrapper[4754]: I0126 17:09:15.631680 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6f82d0f5-47a4-4584-9ae9-046882fa8f61-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"6f82d0f5-47a4-4584-9ae9-046882fa8f61\") " pod="openstack/cinder-scheduler-0" Jan 26 17:09:15 crc kubenswrapper[4754]: I0126 17:09:15.631727 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/977ff39d-2c55-438b-b957-62609c4c2ece-ovsdbserver-sb\") pod \"dnsmasq-dns-795f4db4bc-h7mcp\" (UID: \"977ff39d-2c55-438b-b957-62609c4c2ece\") " pod="openstack/dnsmasq-dns-795f4db4bc-h7mcp" Jan 26 17:09:15 crc kubenswrapper[4754]: I0126 17:09:15.631775 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/977ff39d-2c55-438b-b957-62609c4c2ece-dns-swift-storage-0\") pod \"dnsmasq-dns-795f4db4bc-h7mcp\" (UID: \"977ff39d-2c55-438b-b957-62609c4c2ece\") " pod="openstack/dnsmasq-dns-795f4db4bc-h7mcp" Jan 26 17:09:15 crc kubenswrapper[4754]: I0126 17:09:15.631887 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6f82d0f5-47a4-4584-9ae9-046882fa8f61-config-data\") pod \"cinder-scheduler-0\" (UID: \"6f82d0f5-47a4-4584-9ae9-046882fa8f61\") " pod="openstack/cinder-scheduler-0" Jan 26 17:09:15 crc kubenswrapper[4754]: I0126 17:09:15.631916 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wndl7\" (UniqueName: \"kubernetes.io/projected/977ff39d-2c55-438b-b957-62609c4c2ece-kube-api-access-wndl7\") pod \"dnsmasq-dns-795f4db4bc-h7mcp\" (UID: \"977ff39d-2c55-438b-b957-62609c4c2ece\") " pod="openstack/dnsmasq-dns-795f4db4bc-h7mcp" Jan 26 17:09:15 crc kubenswrapper[4754]: I0126 17:09:15.631941 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/977ff39d-2c55-438b-b957-62609c4c2ece-config\") pod \"dnsmasq-dns-795f4db4bc-h7mcp\" (UID: \"977ff39d-2c55-438b-b957-62609c4c2ece\") " pod="openstack/dnsmasq-dns-795f4db4bc-h7mcp" Jan 26 17:09:15 crc kubenswrapper[4754]: I0126 17:09:15.631971 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9hzfn\" (UniqueName: \"kubernetes.io/projected/6f82d0f5-47a4-4584-9ae9-046882fa8f61-kube-api-access-9hzfn\") pod \"cinder-scheduler-0\" (UID: \"6f82d0f5-47a4-4584-9ae9-046882fa8f61\") " pod="openstack/cinder-scheduler-0" Jan 26 17:09:15 crc kubenswrapper[4754]: I0126 17:09:15.631995 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/977ff39d-2c55-438b-b957-62609c4c2ece-dns-svc\") pod \"dnsmasq-dns-795f4db4bc-h7mcp\" (UID: \"977ff39d-2c55-438b-b957-62609c4c2ece\") " pod="openstack/dnsmasq-dns-795f4db4bc-h7mcp" Jan 26 17:09:15 crc kubenswrapper[4754]: I0126 17:09:15.632029 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f82d0f5-47a4-4584-9ae9-046882fa8f61-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"6f82d0f5-47a4-4584-9ae9-046882fa8f61\") " pod="openstack/cinder-scheduler-0" Jan 26 17:09:15 crc kubenswrapper[4754]: I0126 17:09:15.632054 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6f82d0f5-47a4-4584-9ae9-046882fa8f61-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"6f82d0f5-47a4-4584-9ae9-046882fa8f61\") " pod="openstack/cinder-scheduler-0" Jan 26 17:09:15 crc kubenswrapper[4754]: I0126 17:09:15.632088 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6f82d0f5-47a4-4584-9ae9-046882fa8f61-scripts\") pod \"cinder-scheduler-0\" (UID: \"6f82d0f5-47a4-4584-9ae9-046882fa8f61\") " pod="openstack/cinder-scheduler-0" Jan 26 17:09:15 crc kubenswrapper[4754]: I0126 17:09:15.635741 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6f82d0f5-47a4-4584-9ae9-046882fa8f61-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"6f82d0f5-47a4-4584-9ae9-046882fa8f61\") " pod="openstack/cinder-scheduler-0" Jan 26 17:09:15 crc kubenswrapper[4754]: I0126 17:09:15.642710 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6f82d0f5-47a4-4584-9ae9-046882fa8f61-scripts\") pod \"cinder-scheduler-0\" (UID: \"6f82d0f5-47a4-4584-9ae9-046882fa8f61\") " pod="openstack/cinder-scheduler-0" Jan 26 17:09:15 crc kubenswrapper[4754]: I0126 17:09:15.643899 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f82d0f5-47a4-4584-9ae9-046882fa8f61-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"6f82d0f5-47a4-4584-9ae9-046882fa8f61\") " pod="openstack/cinder-scheduler-0" Jan 26 17:09:15 crc kubenswrapper[4754]: I0126 17:09:15.645808 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6f82d0f5-47a4-4584-9ae9-046882fa8f61-config-data\") pod \"cinder-scheduler-0\" (UID: \"6f82d0f5-47a4-4584-9ae9-046882fa8f61\") " pod="openstack/cinder-scheduler-0" Jan 26 17:09:15 crc kubenswrapper[4754]: I0126 17:09:15.647043 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Jan 26 17:09:15 crc kubenswrapper[4754]: I0126 17:09:15.650345 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6f82d0f5-47a4-4584-9ae9-046882fa8f61-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"6f82d0f5-47a4-4584-9ae9-046882fa8f61\") " pod="openstack/cinder-scheduler-0" Jan 26 17:09:15 crc kubenswrapper[4754]: I0126 17:09:15.651020 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Jan 26 17:09:15 crc kubenswrapper[4754]: I0126 17:09:15.653534 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Jan 26 17:09:15 crc kubenswrapper[4754]: I0126 17:09:15.657281 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9hzfn\" (UniqueName: \"kubernetes.io/projected/6f82d0f5-47a4-4584-9ae9-046882fa8f61-kube-api-access-9hzfn\") pod \"cinder-scheduler-0\" (UID: \"6f82d0f5-47a4-4584-9ae9-046882fa8f61\") " pod="openstack/cinder-scheduler-0" Jan 26 17:09:15 crc kubenswrapper[4754]: I0126 17:09:15.657928 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Jan 26 17:09:15 crc kubenswrapper[4754]: I0126 17:09:15.683794 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Jan 26 17:09:15 crc kubenswrapper[4754]: I0126 17:09:15.733869 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wndl7\" (UniqueName: \"kubernetes.io/projected/977ff39d-2c55-438b-b957-62609c4c2ece-kube-api-access-wndl7\") pod \"dnsmasq-dns-795f4db4bc-h7mcp\" (UID: \"977ff39d-2c55-438b-b957-62609c4c2ece\") " pod="openstack/dnsmasq-dns-795f4db4bc-h7mcp" Jan 26 17:09:15 crc kubenswrapper[4754]: I0126 17:09:15.733932 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/977ff39d-2c55-438b-b957-62609c4c2ece-config\") pod \"dnsmasq-dns-795f4db4bc-h7mcp\" (UID: \"977ff39d-2c55-438b-b957-62609c4c2ece\") " pod="openstack/dnsmasq-dns-795f4db4bc-h7mcp" Jan 26 17:09:15 crc kubenswrapper[4754]: I0126 17:09:15.733966 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/54272c08-92b7-4512-8ee3-2a782926cda1-config-data-custom\") pod \"cinder-api-0\" (UID: \"54272c08-92b7-4512-8ee3-2a782926cda1\") " pod="openstack/cinder-api-0" Jan 26 17:09:15 crc kubenswrapper[4754]: I0126 17:09:15.733993 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-czd2j\" (UniqueName: \"kubernetes.io/projected/54272c08-92b7-4512-8ee3-2a782926cda1-kube-api-access-czd2j\") pod \"cinder-api-0\" (UID: \"54272c08-92b7-4512-8ee3-2a782926cda1\") " pod="openstack/cinder-api-0" Jan 26 17:09:15 crc kubenswrapper[4754]: I0126 17:09:15.734016 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/977ff39d-2c55-438b-b957-62609c4c2ece-dns-svc\") pod \"dnsmasq-dns-795f4db4bc-h7mcp\" (UID: \"977ff39d-2c55-438b-b957-62609c4c2ece\") " pod="openstack/dnsmasq-dns-795f4db4bc-h7mcp" Jan 26 17:09:15 crc kubenswrapper[4754]: I0126 17:09:15.734035 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/54272c08-92b7-4512-8ee3-2a782926cda1-scripts\") pod \"cinder-api-0\" (UID: \"54272c08-92b7-4512-8ee3-2a782926cda1\") " pod="openstack/cinder-api-0" Jan 26 17:09:15 crc kubenswrapper[4754]: I0126 17:09:15.734085 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/977ff39d-2c55-438b-b957-62609c4c2ece-ovsdbserver-nb\") pod \"dnsmasq-dns-795f4db4bc-h7mcp\" (UID: \"977ff39d-2c55-438b-b957-62609c4c2ece\") " pod="openstack/dnsmasq-dns-795f4db4bc-h7mcp" Jan 26 17:09:15 crc kubenswrapper[4754]: I0126 17:09:15.734295 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/977ff39d-2c55-438b-b957-62609c4c2ece-ovsdbserver-sb\") pod \"dnsmasq-dns-795f4db4bc-h7mcp\" (UID: \"977ff39d-2c55-438b-b957-62609c4c2ece\") " pod="openstack/dnsmasq-dns-795f4db4bc-h7mcp" Jan 26 17:09:15 crc kubenswrapper[4754]: I0126 17:09:15.734339 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/977ff39d-2c55-438b-b957-62609c4c2ece-dns-swift-storage-0\") pod \"dnsmasq-dns-795f4db4bc-h7mcp\" (UID: \"977ff39d-2c55-438b-b957-62609c4c2ece\") " pod="openstack/dnsmasq-dns-795f4db4bc-h7mcp" Jan 26 17:09:15 crc kubenswrapper[4754]: I0126 17:09:15.734437 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/54272c08-92b7-4512-8ee3-2a782926cda1-etc-machine-id\") pod \"cinder-api-0\" (UID: \"54272c08-92b7-4512-8ee3-2a782926cda1\") " pod="openstack/cinder-api-0" Jan 26 17:09:15 crc kubenswrapper[4754]: I0126 17:09:15.734462 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54272c08-92b7-4512-8ee3-2a782926cda1-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"54272c08-92b7-4512-8ee3-2a782926cda1\") " pod="openstack/cinder-api-0" Jan 26 17:09:15 crc kubenswrapper[4754]: I0126 17:09:15.734499 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/54272c08-92b7-4512-8ee3-2a782926cda1-config-data\") pod \"cinder-api-0\" (UID: \"54272c08-92b7-4512-8ee3-2a782926cda1\") " pod="openstack/cinder-api-0" Jan 26 17:09:15 crc kubenswrapper[4754]: I0126 17:09:15.734518 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/54272c08-92b7-4512-8ee3-2a782926cda1-logs\") pod \"cinder-api-0\" (UID: \"54272c08-92b7-4512-8ee3-2a782926cda1\") " pod="openstack/cinder-api-0" Jan 26 17:09:15 crc kubenswrapper[4754]: I0126 17:09:15.735246 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/977ff39d-2c55-438b-b957-62609c4c2ece-ovsdbserver-nb\") pod \"dnsmasq-dns-795f4db4bc-h7mcp\" (UID: \"977ff39d-2c55-438b-b957-62609c4c2ece\") " pod="openstack/dnsmasq-dns-795f4db4bc-h7mcp" Jan 26 17:09:15 crc kubenswrapper[4754]: I0126 17:09:15.735806 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/977ff39d-2c55-438b-b957-62609c4c2ece-config\") pod \"dnsmasq-dns-795f4db4bc-h7mcp\" (UID: \"977ff39d-2c55-438b-b957-62609c4c2ece\") " pod="openstack/dnsmasq-dns-795f4db4bc-h7mcp" Jan 26 17:09:15 crc kubenswrapper[4754]: I0126 17:09:15.736510 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/977ff39d-2c55-438b-b957-62609c4c2ece-ovsdbserver-sb\") pod \"dnsmasq-dns-795f4db4bc-h7mcp\" (UID: \"977ff39d-2c55-438b-b957-62609c4c2ece\") " pod="openstack/dnsmasq-dns-795f4db4bc-h7mcp" Jan 26 17:09:15 crc kubenswrapper[4754]: I0126 17:09:15.737442 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/977ff39d-2c55-438b-b957-62609c4c2ece-dns-svc\") pod \"dnsmasq-dns-795f4db4bc-h7mcp\" (UID: \"977ff39d-2c55-438b-b957-62609c4c2ece\") " pod="openstack/dnsmasq-dns-795f4db4bc-h7mcp" Jan 26 17:09:15 crc kubenswrapper[4754]: I0126 17:09:15.738284 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/977ff39d-2c55-438b-b957-62609c4c2ece-dns-swift-storage-0\") pod \"dnsmasq-dns-795f4db4bc-h7mcp\" (UID: \"977ff39d-2c55-438b-b957-62609c4c2ece\") " pod="openstack/dnsmasq-dns-795f4db4bc-h7mcp" Jan 26 17:09:15 crc kubenswrapper[4754]: I0126 17:09:15.752341 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wndl7\" (UniqueName: \"kubernetes.io/projected/977ff39d-2c55-438b-b957-62609c4c2ece-kube-api-access-wndl7\") pod \"dnsmasq-dns-795f4db4bc-h7mcp\" (UID: \"977ff39d-2c55-438b-b957-62609c4c2ece\") " pod="openstack/dnsmasq-dns-795f4db4bc-h7mcp" Jan 26 17:09:15 crc kubenswrapper[4754]: I0126 17:09:15.836492 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/54272c08-92b7-4512-8ee3-2a782926cda1-etc-machine-id\") pod \"cinder-api-0\" (UID: \"54272c08-92b7-4512-8ee3-2a782926cda1\") " pod="openstack/cinder-api-0" Jan 26 17:09:15 crc kubenswrapper[4754]: I0126 17:09:15.836856 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54272c08-92b7-4512-8ee3-2a782926cda1-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"54272c08-92b7-4512-8ee3-2a782926cda1\") " pod="openstack/cinder-api-0" Jan 26 17:09:15 crc kubenswrapper[4754]: I0126 17:09:15.836626 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/54272c08-92b7-4512-8ee3-2a782926cda1-etc-machine-id\") pod \"cinder-api-0\" (UID: \"54272c08-92b7-4512-8ee3-2a782926cda1\") " pod="openstack/cinder-api-0" Jan 26 17:09:15 crc kubenswrapper[4754]: I0126 17:09:15.837071 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/54272c08-92b7-4512-8ee3-2a782926cda1-config-data\") pod \"cinder-api-0\" (UID: \"54272c08-92b7-4512-8ee3-2a782926cda1\") " pod="openstack/cinder-api-0" Jan 26 17:09:15 crc kubenswrapper[4754]: I0126 17:09:15.837099 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/54272c08-92b7-4512-8ee3-2a782926cda1-logs\") pod \"cinder-api-0\" (UID: \"54272c08-92b7-4512-8ee3-2a782926cda1\") " pod="openstack/cinder-api-0" Jan 26 17:09:15 crc kubenswrapper[4754]: I0126 17:09:15.837170 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/54272c08-92b7-4512-8ee3-2a782926cda1-config-data-custom\") pod \"cinder-api-0\" (UID: \"54272c08-92b7-4512-8ee3-2a782926cda1\") " pod="openstack/cinder-api-0" Jan 26 17:09:15 crc kubenswrapper[4754]: I0126 17:09:15.837193 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-czd2j\" (UniqueName: \"kubernetes.io/projected/54272c08-92b7-4512-8ee3-2a782926cda1-kube-api-access-czd2j\") pod \"cinder-api-0\" (UID: \"54272c08-92b7-4512-8ee3-2a782926cda1\") " pod="openstack/cinder-api-0" Jan 26 17:09:15 crc kubenswrapper[4754]: I0126 17:09:15.837228 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/54272c08-92b7-4512-8ee3-2a782926cda1-scripts\") pod \"cinder-api-0\" (UID: \"54272c08-92b7-4512-8ee3-2a782926cda1\") " pod="openstack/cinder-api-0" Jan 26 17:09:15 crc kubenswrapper[4754]: I0126 17:09:15.840442 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/54272c08-92b7-4512-8ee3-2a782926cda1-logs\") pod \"cinder-api-0\" (UID: \"54272c08-92b7-4512-8ee3-2a782926cda1\") " pod="openstack/cinder-api-0" Jan 26 17:09:15 crc kubenswrapper[4754]: I0126 17:09:15.841087 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/54272c08-92b7-4512-8ee3-2a782926cda1-config-data-custom\") pod \"cinder-api-0\" (UID: \"54272c08-92b7-4512-8ee3-2a782926cda1\") " pod="openstack/cinder-api-0" Jan 26 17:09:15 crc kubenswrapper[4754]: I0126 17:09:15.841397 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-795f4db4bc-h7mcp" Jan 26 17:09:15 crc kubenswrapper[4754]: I0126 17:09:15.843035 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/54272c08-92b7-4512-8ee3-2a782926cda1-scripts\") pod \"cinder-api-0\" (UID: \"54272c08-92b7-4512-8ee3-2a782926cda1\") " pod="openstack/cinder-api-0" Jan 26 17:09:15 crc kubenswrapper[4754]: I0126 17:09:15.856895 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/54272c08-92b7-4512-8ee3-2a782926cda1-config-data\") pod \"cinder-api-0\" (UID: \"54272c08-92b7-4512-8ee3-2a782926cda1\") " pod="openstack/cinder-api-0" Jan 26 17:09:15 crc kubenswrapper[4754]: I0126 17:09:15.857009 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54272c08-92b7-4512-8ee3-2a782926cda1-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"54272c08-92b7-4512-8ee3-2a782926cda1\") " pod="openstack/cinder-api-0" Jan 26 17:09:15 crc kubenswrapper[4754]: I0126 17:09:15.858290 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-czd2j\" (UniqueName: \"kubernetes.io/projected/54272c08-92b7-4512-8ee3-2a782926cda1-kube-api-access-czd2j\") pod \"cinder-api-0\" (UID: \"54272c08-92b7-4512-8ee3-2a782926cda1\") " pod="openstack/cinder-api-0" Jan 26 17:09:16 crc kubenswrapper[4754]: I0126 17:09:16.117343 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Jan 26 17:09:16 crc kubenswrapper[4754]: I0126 17:09:16.118242 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4c65192f-a05a-45e2-9d56-33ddeb265374","Type":"ContainerStarted","Data":"50992540292cc9c045cd8bb4c9e58c8d373ef8d0a1454a811b960e86c77eec5d"} Jan 26 17:09:16 crc kubenswrapper[4754]: I0126 17:09:16.119242 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Jan 26 17:09:16 crc kubenswrapper[4754]: I0126 17:09:16.153006 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=10.546331968 podStartE2EDuration="15.152988149s" podCreationTimestamp="2026-01-26 17:09:01 +0000 UTC" firstStartedPulling="2026-01-26 17:09:10.298923303 +0000 UTC m=+1316.823103727" lastFinishedPulling="2026-01-26 17:09:14.905579474 +0000 UTC m=+1321.429759908" observedRunningTime="2026-01-26 17:09:16.144993859 +0000 UTC m=+1322.669174293" watchObservedRunningTime="2026-01-26 17:09:16.152988149 +0000 UTC m=+1322.677168583" Jan 26 17:09:16 crc kubenswrapper[4754]: I0126 17:09:16.285944 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Jan 26 17:09:16 crc kubenswrapper[4754]: I0126 17:09:16.293922 4754 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-5d7f8b89fd-jwf5r" podUID="0dee3ce6-9657-4b8a-9034-7c744db2d209" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.164:9311/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Jan 26 17:09:16 crc kubenswrapper[4754]: I0126 17:09:16.294371 4754 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-5d7f8b89fd-jwf5r" podUID="0dee3ce6-9657-4b8a-9034-7c744db2d209" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.164:9311/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Jan 26 17:09:16 crc kubenswrapper[4754]: I0126 17:09:16.503059 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-795f4db4bc-h7mcp"] Jan 26 17:09:16 crc kubenswrapper[4754]: I0126 17:09:16.729212 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Jan 26 17:09:17 crc kubenswrapper[4754]: I0126 17:09:17.013002 4754 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-5d7f8b89fd-jwf5r" podUID="0dee3ce6-9657-4b8a-9034-7c744db2d209" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.164:9311/healthcheck\": read tcp 10.217.0.2:52554->10.217.0.164:9311: read: connection reset by peer" Jan 26 17:09:17 crc kubenswrapper[4754]: I0126 17:09:17.013050 4754 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-5d7f8b89fd-jwf5r" podUID="0dee3ce6-9657-4b8a-9034-7c744db2d209" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.164:9311/healthcheck\": read tcp 10.217.0.2:52540->10.217.0.164:9311: read: connection reset by peer" Jan 26 17:09:17 crc kubenswrapper[4754]: I0126 17:09:17.131898 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"6f82d0f5-47a4-4584-9ae9-046882fa8f61","Type":"ContainerStarted","Data":"ec92e2a2d83dd9be92b1f6fd8f1523b2351ac7fb16413d424503e4d0ed37f8b9"} Jan 26 17:09:17 crc kubenswrapper[4754]: I0126 17:09:17.154069 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"54272c08-92b7-4512-8ee3-2a782926cda1","Type":"ContainerStarted","Data":"ef80ca84ce39ac6f9b5f7493ec39dbf65d993038fc52b5bd85fdf98575a1dbe3"} Jan 26 17:09:17 crc kubenswrapper[4754]: I0126 17:09:17.156015 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-795f4db4bc-h7mcp" event={"ID":"977ff39d-2c55-438b-b957-62609c4c2ece","Type":"ContainerStarted","Data":"bf3e959a5f40da261aa7cf7b87b3564bafac97e7808c4387dbbb133f67fc522f"} Jan 26 17:09:17 crc kubenswrapper[4754]: I0126 17:09:17.621365 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Jan 26 17:09:18 crc kubenswrapper[4754]: I0126 17:09:18.166638 4754 generic.go:334] "Generic (PLEG): container finished" podID="0dee3ce6-9657-4b8a-9034-7c744db2d209" containerID="6401776521d0c72d1ea925c21389233c039edc6ed1e4450e98139bea1afe8ec3" exitCode=0 Jan 26 17:09:18 crc kubenswrapper[4754]: I0126 17:09:18.166701 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5d7f8b89fd-jwf5r" event={"ID":"0dee3ce6-9657-4b8a-9034-7c744db2d209","Type":"ContainerDied","Data":"6401776521d0c72d1ea925c21389233c039edc6ed1e4450e98139bea1afe8ec3"} Jan 26 17:09:18 crc kubenswrapper[4754]: I0126 17:09:18.170355 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"54272c08-92b7-4512-8ee3-2a782926cda1","Type":"ContainerStarted","Data":"e38b379246c733fe550b402977dd4a807b523d7d988b7a00913ef758633e2e36"} Jan 26 17:09:18 crc kubenswrapper[4754]: I0126 17:09:18.172623 4754 generic.go:334] "Generic (PLEG): container finished" podID="977ff39d-2c55-438b-b957-62609c4c2ece" containerID="38b3ccb1c0bbed7e3bde1802c479d558b8daa522cae6af0fb1c2ed14fdc80d28" exitCode=0 Jan 26 17:09:18 crc kubenswrapper[4754]: I0126 17:09:18.172684 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-795f4db4bc-h7mcp" event={"ID":"977ff39d-2c55-438b-b957-62609c4c2ece","Type":"ContainerDied","Data":"38b3ccb1c0bbed7e3bde1802c479d558b8daa522cae6af0fb1c2ed14fdc80d28"} Jan 26 17:09:23 crc kubenswrapper[4754]: I0126 17:09:23.819337 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5d7f8b89fd-jwf5r" Jan 26 17:09:23 crc kubenswrapper[4754]: I0126 17:09:23.910763 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0dee3ce6-9657-4b8a-9034-7c744db2d209-config-data\") pod \"0dee3ce6-9657-4b8a-9034-7c744db2d209\" (UID: \"0dee3ce6-9657-4b8a-9034-7c744db2d209\") " Jan 26 17:09:23 crc kubenswrapper[4754]: I0126 17:09:23.911109 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0dee3ce6-9657-4b8a-9034-7c744db2d209-combined-ca-bundle\") pod \"0dee3ce6-9657-4b8a-9034-7c744db2d209\" (UID: \"0dee3ce6-9657-4b8a-9034-7c744db2d209\") " Jan 26 17:09:23 crc kubenswrapper[4754]: I0126 17:09:23.911140 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0dee3ce6-9657-4b8a-9034-7c744db2d209-logs\") pod \"0dee3ce6-9657-4b8a-9034-7c744db2d209\" (UID: \"0dee3ce6-9657-4b8a-9034-7c744db2d209\") " Jan 26 17:09:23 crc kubenswrapper[4754]: I0126 17:09:23.911198 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4sjpt\" (UniqueName: \"kubernetes.io/projected/0dee3ce6-9657-4b8a-9034-7c744db2d209-kube-api-access-4sjpt\") pod \"0dee3ce6-9657-4b8a-9034-7c744db2d209\" (UID: \"0dee3ce6-9657-4b8a-9034-7c744db2d209\") " Jan 26 17:09:23 crc kubenswrapper[4754]: I0126 17:09:23.911403 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0dee3ce6-9657-4b8a-9034-7c744db2d209-config-data-custom\") pod \"0dee3ce6-9657-4b8a-9034-7c744db2d209\" (UID: \"0dee3ce6-9657-4b8a-9034-7c744db2d209\") " Jan 26 17:09:23 crc kubenswrapper[4754]: I0126 17:09:23.911762 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0dee3ce6-9657-4b8a-9034-7c744db2d209-logs" (OuterVolumeSpecName: "logs") pod "0dee3ce6-9657-4b8a-9034-7c744db2d209" (UID: "0dee3ce6-9657-4b8a-9034-7c744db2d209"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:09:23 crc kubenswrapper[4754]: I0126 17:09:23.912148 4754 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0dee3ce6-9657-4b8a-9034-7c744db2d209-logs\") on node \"crc\" DevicePath \"\"" Jan 26 17:09:23 crc kubenswrapper[4754]: I0126 17:09:23.920513 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0dee3ce6-9657-4b8a-9034-7c744db2d209-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "0dee3ce6-9657-4b8a-9034-7c744db2d209" (UID: "0dee3ce6-9657-4b8a-9034-7c744db2d209"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:09:23 crc kubenswrapper[4754]: I0126 17:09:23.920624 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0dee3ce6-9657-4b8a-9034-7c744db2d209-kube-api-access-4sjpt" (OuterVolumeSpecName: "kube-api-access-4sjpt") pod "0dee3ce6-9657-4b8a-9034-7c744db2d209" (UID: "0dee3ce6-9657-4b8a-9034-7c744db2d209"). InnerVolumeSpecName "kube-api-access-4sjpt". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:09:24 crc kubenswrapper[4754]: I0126 17:09:24.021370 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4sjpt\" (UniqueName: \"kubernetes.io/projected/0dee3ce6-9657-4b8a-9034-7c744db2d209-kube-api-access-4sjpt\") on node \"crc\" DevicePath \"\"" Jan 26 17:09:24 crc kubenswrapper[4754]: I0126 17:09:24.021407 4754 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0dee3ce6-9657-4b8a-9034-7c744db2d209-config-data-custom\") on node \"crc\" DevicePath \"\"" Jan 26 17:09:24 crc kubenswrapper[4754]: I0126 17:09:24.026469 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0dee3ce6-9657-4b8a-9034-7c744db2d209-config-data" (OuterVolumeSpecName: "config-data") pod "0dee3ce6-9657-4b8a-9034-7c744db2d209" (UID: "0dee3ce6-9657-4b8a-9034-7c744db2d209"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:09:24 crc kubenswrapper[4754]: I0126 17:09:24.028112 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0dee3ce6-9657-4b8a-9034-7c744db2d209-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0dee3ce6-9657-4b8a-9034-7c744db2d209" (UID: "0dee3ce6-9657-4b8a-9034-7c744db2d209"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:09:24 crc kubenswrapper[4754]: I0126 17:09:24.122640 4754 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0dee3ce6-9657-4b8a-9034-7c744db2d209-config-data\") on node \"crc\" DevicePath \"\"" Jan 26 17:09:24 crc kubenswrapper[4754]: I0126 17:09:24.122686 4754 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0dee3ce6-9657-4b8a-9034-7c744db2d209-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 26 17:09:24 crc kubenswrapper[4754]: I0126 17:09:24.220001 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-6959bf4485-bkv7n"] Jan 26 17:09:24 crc kubenswrapper[4754]: E0126 17:09:24.220485 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0dee3ce6-9657-4b8a-9034-7c744db2d209" containerName="barbican-api-log" Jan 26 17:09:24 crc kubenswrapper[4754]: I0126 17:09:24.220509 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="0dee3ce6-9657-4b8a-9034-7c744db2d209" containerName="barbican-api-log" Jan 26 17:09:24 crc kubenswrapper[4754]: E0126 17:09:24.220547 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0dee3ce6-9657-4b8a-9034-7c744db2d209" containerName="barbican-api" Jan 26 17:09:24 crc kubenswrapper[4754]: I0126 17:09:24.220556 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="0dee3ce6-9657-4b8a-9034-7c744db2d209" containerName="barbican-api" Jan 26 17:09:24 crc kubenswrapper[4754]: I0126 17:09:24.220797 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="0dee3ce6-9657-4b8a-9034-7c744db2d209" containerName="barbican-api" Jan 26 17:09:24 crc kubenswrapper[4754]: I0126 17:09:24.220823 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="0dee3ce6-9657-4b8a-9034-7c744db2d209" containerName="barbican-api-log" Jan 26 17:09:24 crc kubenswrapper[4754]: I0126 17:09:24.222002 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-6959bf4485-bkv7n" Jan 26 17:09:24 crc kubenswrapper[4754]: I0126 17:09:24.225327 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Jan 26 17:09:24 crc kubenswrapper[4754]: I0126 17:09:24.226353 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Jan 26 17:09:24 crc kubenswrapper[4754]: I0126 17:09:24.227135 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Jan 26 17:09:24 crc kubenswrapper[4754]: I0126 17:09:24.233992 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-6959bf4485-bkv7n"] Jan 26 17:09:24 crc kubenswrapper[4754]: I0126 17:09:24.264110 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5d7f8b89fd-jwf5r" event={"ID":"0dee3ce6-9657-4b8a-9034-7c744db2d209","Type":"ContainerDied","Data":"17f7bc11c09ec7d6307edb1ea875db377ca62b684188c883a065b30028bc2138"} Jan 26 17:09:24 crc kubenswrapper[4754]: I0126 17:09:24.264165 4754 scope.go:117] "RemoveContainer" containerID="6401776521d0c72d1ea925c21389233c039edc6ed1e4450e98139bea1afe8ec3" Jan 26 17:09:24 crc kubenswrapper[4754]: I0126 17:09:24.264201 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5d7f8b89fd-jwf5r" Jan 26 17:09:24 crc kubenswrapper[4754]: I0126 17:09:24.292448 4754 scope.go:117] "RemoveContainer" containerID="a5dcc43e1e1644072f8d0a7424cbbc5cfda3573dc5d056f0c4781edb61fa886e" Jan 26 17:09:24 crc kubenswrapper[4754]: I0126 17:09:24.301167 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-5d7f8b89fd-jwf5r"] Jan 26 17:09:24 crc kubenswrapper[4754]: I0126 17:09:24.311757 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-5d7f8b89fd-jwf5r"] Jan 26 17:09:24 crc kubenswrapper[4754]: I0126 17:09:24.328467 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1622eb04-be06-4e2a-90d6-27a58ac54e60-run-httpd\") pod \"swift-proxy-6959bf4485-bkv7n\" (UID: \"1622eb04-be06-4e2a-90d6-27a58ac54e60\") " pod="openstack/swift-proxy-6959bf4485-bkv7n" Jan 26 17:09:24 crc kubenswrapper[4754]: I0126 17:09:24.328549 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/1622eb04-be06-4e2a-90d6-27a58ac54e60-etc-swift\") pod \"swift-proxy-6959bf4485-bkv7n\" (UID: \"1622eb04-be06-4e2a-90d6-27a58ac54e60\") " pod="openstack/swift-proxy-6959bf4485-bkv7n" Jan 26 17:09:24 crc kubenswrapper[4754]: I0126 17:09:24.328751 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1622eb04-be06-4e2a-90d6-27a58ac54e60-public-tls-certs\") pod \"swift-proxy-6959bf4485-bkv7n\" (UID: \"1622eb04-be06-4e2a-90d6-27a58ac54e60\") " pod="openstack/swift-proxy-6959bf4485-bkv7n" Jan 26 17:09:24 crc kubenswrapper[4754]: I0126 17:09:24.328860 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1622eb04-be06-4e2a-90d6-27a58ac54e60-internal-tls-certs\") pod \"swift-proxy-6959bf4485-bkv7n\" (UID: \"1622eb04-be06-4e2a-90d6-27a58ac54e60\") " pod="openstack/swift-proxy-6959bf4485-bkv7n" Jan 26 17:09:24 crc kubenswrapper[4754]: I0126 17:09:24.328902 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1622eb04-be06-4e2a-90d6-27a58ac54e60-combined-ca-bundle\") pod \"swift-proxy-6959bf4485-bkv7n\" (UID: \"1622eb04-be06-4e2a-90d6-27a58ac54e60\") " pod="openstack/swift-proxy-6959bf4485-bkv7n" Jan 26 17:09:24 crc kubenswrapper[4754]: I0126 17:09:24.328943 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mhx8c\" (UniqueName: \"kubernetes.io/projected/1622eb04-be06-4e2a-90d6-27a58ac54e60-kube-api-access-mhx8c\") pod \"swift-proxy-6959bf4485-bkv7n\" (UID: \"1622eb04-be06-4e2a-90d6-27a58ac54e60\") " pod="openstack/swift-proxy-6959bf4485-bkv7n" Jan 26 17:09:24 crc kubenswrapper[4754]: I0126 17:09:24.329056 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1622eb04-be06-4e2a-90d6-27a58ac54e60-log-httpd\") pod \"swift-proxy-6959bf4485-bkv7n\" (UID: \"1622eb04-be06-4e2a-90d6-27a58ac54e60\") " pod="openstack/swift-proxy-6959bf4485-bkv7n" Jan 26 17:09:24 crc kubenswrapper[4754]: I0126 17:09:24.329091 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1622eb04-be06-4e2a-90d6-27a58ac54e60-config-data\") pod \"swift-proxy-6959bf4485-bkv7n\" (UID: \"1622eb04-be06-4e2a-90d6-27a58ac54e60\") " pod="openstack/swift-proxy-6959bf4485-bkv7n" Jan 26 17:09:24 crc kubenswrapper[4754]: I0126 17:09:24.430093 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mhx8c\" (UniqueName: \"kubernetes.io/projected/1622eb04-be06-4e2a-90d6-27a58ac54e60-kube-api-access-mhx8c\") pod \"swift-proxy-6959bf4485-bkv7n\" (UID: \"1622eb04-be06-4e2a-90d6-27a58ac54e60\") " pod="openstack/swift-proxy-6959bf4485-bkv7n" Jan 26 17:09:24 crc kubenswrapper[4754]: I0126 17:09:24.430159 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1622eb04-be06-4e2a-90d6-27a58ac54e60-log-httpd\") pod \"swift-proxy-6959bf4485-bkv7n\" (UID: \"1622eb04-be06-4e2a-90d6-27a58ac54e60\") " pod="openstack/swift-proxy-6959bf4485-bkv7n" Jan 26 17:09:24 crc kubenswrapper[4754]: I0126 17:09:24.430183 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1622eb04-be06-4e2a-90d6-27a58ac54e60-config-data\") pod \"swift-proxy-6959bf4485-bkv7n\" (UID: \"1622eb04-be06-4e2a-90d6-27a58ac54e60\") " pod="openstack/swift-proxy-6959bf4485-bkv7n" Jan 26 17:09:24 crc kubenswrapper[4754]: I0126 17:09:24.430224 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1622eb04-be06-4e2a-90d6-27a58ac54e60-run-httpd\") pod \"swift-proxy-6959bf4485-bkv7n\" (UID: \"1622eb04-be06-4e2a-90d6-27a58ac54e60\") " pod="openstack/swift-proxy-6959bf4485-bkv7n" Jan 26 17:09:24 crc kubenswrapper[4754]: I0126 17:09:24.430244 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/1622eb04-be06-4e2a-90d6-27a58ac54e60-etc-swift\") pod \"swift-proxy-6959bf4485-bkv7n\" (UID: \"1622eb04-be06-4e2a-90d6-27a58ac54e60\") " pod="openstack/swift-proxy-6959bf4485-bkv7n" Jan 26 17:09:24 crc kubenswrapper[4754]: I0126 17:09:24.430298 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1622eb04-be06-4e2a-90d6-27a58ac54e60-public-tls-certs\") pod \"swift-proxy-6959bf4485-bkv7n\" (UID: \"1622eb04-be06-4e2a-90d6-27a58ac54e60\") " pod="openstack/swift-proxy-6959bf4485-bkv7n" Jan 26 17:09:24 crc kubenswrapper[4754]: I0126 17:09:24.430328 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1622eb04-be06-4e2a-90d6-27a58ac54e60-internal-tls-certs\") pod \"swift-proxy-6959bf4485-bkv7n\" (UID: \"1622eb04-be06-4e2a-90d6-27a58ac54e60\") " pod="openstack/swift-proxy-6959bf4485-bkv7n" Jan 26 17:09:24 crc kubenswrapper[4754]: I0126 17:09:24.430345 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1622eb04-be06-4e2a-90d6-27a58ac54e60-combined-ca-bundle\") pod \"swift-proxy-6959bf4485-bkv7n\" (UID: \"1622eb04-be06-4e2a-90d6-27a58ac54e60\") " pod="openstack/swift-proxy-6959bf4485-bkv7n" Jan 26 17:09:24 crc kubenswrapper[4754]: I0126 17:09:24.432160 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1622eb04-be06-4e2a-90d6-27a58ac54e60-log-httpd\") pod \"swift-proxy-6959bf4485-bkv7n\" (UID: \"1622eb04-be06-4e2a-90d6-27a58ac54e60\") " pod="openstack/swift-proxy-6959bf4485-bkv7n" Jan 26 17:09:24 crc kubenswrapper[4754]: I0126 17:09:24.432224 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1622eb04-be06-4e2a-90d6-27a58ac54e60-run-httpd\") pod \"swift-proxy-6959bf4485-bkv7n\" (UID: \"1622eb04-be06-4e2a-90d6-27a58ac54e60\") " pod="openstack/swift-proxy-6959bf4485-bkv7n" Jan 26 17:09:24 crc kubenswrapper[4754]: I0126 17:09:24.435681 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/1622eb04-be06-4e2a-90d6-27a58ac54e60-etc-swift\") pod \"swift-proxy-6959bf4485-bkv7n\" (UID: \"1622eb04-be06-4e2a-90d6-27a58ac54e60\") " pod="openstack/swift-proxy-6959bf4485-bkv7n" Jan 26 17:09:24 crc kubenswrapper[4754]: I0126 17:09:24.436133 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1622eb04-be06-4e2a-90d6-27a58ac54e60-config-data\") pod \"swift-proxy-6959bf4485-bkv7n\" (UID: \"1622eb04-be06-4e2a-90d6-27a58ac54e60\") " pod="openstack/swift-proxy-6959bf4485-bkv7n" Jan 26 17:09:24 crc kubenswrapper[4754]: I0126 17:09:24.436591 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1622eb04-be06-4e2a-90d6-27a58ac54e60-public-tls-certs\") pod \"swift-proxy-6959bf4485-bkv7n\" (UID: \"1622eb04-be06-4e2a-90d6-27a58ac54e60\") " pod="openstack/swift-proxy-6959bf4485-bkv7n" Jan 26 17:09:24 crc kubenswrapper[4754]: I0126 17:09:24.440784 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1622eb04-be06-4e2a-90d6-27a58ac54e60-internal-tls-certs\") pod \"swift-proxy-6959bf4485-bkv7n\" (UID: \"1622eb04-be06-4e2a-90d6-27a58ac54e60\") " pod="openstack/swift-proxy-6959bf4485-bkv7n" Jan 26 17:09:24 crc kubenswrapper[4754]: I0126 17:09:24.449139 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mhx8c\" (UniqueName: \"kubernetes.io/projected/1622eb04-be06-4e2a-90d6-27a58ac54e60-kube-api-access-mhx8c\") pod \"swift-proxy-6959bf4485-bkv7n\" (UID: \"1622eb04-be06-4e2a-90d6-27a58ac54e60\") " pod="openstack/swift-proxy-6959bf4485-bkv7n" Jan 26 17:09:24 crc kubenswrapper[4754]: I0126 17:09:24.455161 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1622eb04-be06-4e2a-90d6-27a58ac54e60-combined-ca-bundle\") pod \"swift-proxy-6959bf4485-bkv7n\" (UID: \"1622eb04-be06-4e2a-90d6-27a58ac54e60\") " pod="openstack/swift-proxy-6959bf4485-bkv7n" Jan 26 17:09:24 crc kubenswrapper[4754]: I0126 17:09:24.529748 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Jan 26 17:09:24 crc kubenswrapper[4754]: I0126 17:09:24.530191 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="4a8fafa1-3d82-4a16-85e1-fd3116bc998a" containerName="glance-log" containerID="cri-o://68b9e43495925deaaeb41ff5279207dddd543bad6b2dd86201727f1e8bd1577d" gracePeriod=30 Jan 26 17:09:24 crc kubenswrapper[4754]: I0126 17:09:24.530357 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="4a8fafa1-3d82-4a16-85e1-fd3116bc998a" containerName="glance-httpd" containerID="cri-o://0c78645b52c9649d8e28a70685c6baf4ee822767bcf49eaf0c693d92ab3e3521" gracePeriod=30 Jan 26 17:09:24 crc kubenswrapper[4754]: I0126 17:09:24.568211 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-6959bf4485-bkv7n" Jan 26 17:09:24 crc kubenswrapper[4754]: I0126 17:09:24.956719 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Jan 26 17:09:24 crc kubenswrapper[4754]: I0126 17:09:24.957310 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4c65192f-a05a-45e2-9d56-33ddeb265374" containerName="ceilometer-central-agent" containerID="cri-o://374dc6158046ff5aeb4527bdd32fd9636f239a18d5ba35caa78a117156d7fb06" gracePeriod=30 Jan 26 17:09:24 crc kubenswrapper[4754]: I0126 17:09:24.957355 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4c65192f-a05a-45e2-9d56-33ddeb265374" containerName="proxy-httpd" containerID="cri-o://50992540292cc9c045cd8bb4c9e58c8d373ef8d0a1454a811b960e86c77eec5d" gracePeriod=30 Jan 26 17:09:24 crc kubenswrapper[4754]: I0126 17:09:24.957397 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4c65192f-a05a-45e2-9d56-33ddeb265374" containerName="ceilometer-notification-agent" containerID="cri-o://0bd046d181549657eae96bff37d02dbcc14075bc5395685c218279cc1485d7bd" gracePeriod=30 Jan 26 17:09:24 crc kubenswrapper[4754]: I0126 17:09:24.957399 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4c65192f-a05a-45e2-9d56-33ddeb265374" containerName="sg-core" containerID="cri-o://3e9828e079b47eb4e66edc9599b5bdd713959f79d23385e7543f9bcb4e6ef25e" gracePeriod=30 Jan 26 17:09:25 crc kubenswrapper[4754]: I0126 17:09:25.210301 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-6959bf4485-bkv7n"] Jan 26 17:09:25 crc kubenswrapper[4754]: W0126 17:09:25.211810 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1622eb04_be06_4e2a_90d6_27a58ac54e60.slice/crio-68de358117abea3ed1363a89ea041a8331328f94209be04f8e1fcdf3d10b68ba WatchSource:0}: Error finding container 68de358117abea3ed1363a89ea041a8331328f94209be04f8e1fcdf3d10b68ba: Status 404 returned error can't find the container with id 68de358117abea3ed1363a89ea041a8331328f94209be04f8e1fcdf3d10b68ba Jan 26 17:09:25 crc kubenswrapper[4754]: I0126 17:09:25.274460 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-795f4db4bc-h7mcp" event={"ID":"977ff39d-2c55-438b-b957-62609c4c2ece","Type":"ContainerStarted","Data":"9a6add83bff5d8df998c3ea59d3a4c3be4deffcc6f345db9dfe249b6f0c866f3"} Jan 26 17:09:25 crc kubenswrapper[4754]: I0126 17:09:25.274966 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-795f4db4bc-h7mcp" Jan 26 17:09:25 crc kubenswrapper[4754]: I0126 17:09:25.284287 4754 generic.go:334] "Generic (PLEG): container finished" podID="4c65192f-a05a-45e2-9d56-33ddeb265374" containerID="50992540292cc9c045cd8bb4c9e58c8d373ef8d0a1454a811b960e86c77eec5d" exitCode=0 Jan 26 17:09:25 crc kubenswrapper[4754]: I0126 17:09:25.284314 4754 generic.go:334] "Generic (PLEG): container finished" podID="4c65192f-a05a-45e2-9d56-33ddeb265374" containerID="3e9828e079b47eb4e66edc9599b5bdd713959f79d23385e7543f9bcb4e6ef25e" exitCode=2 Jan 26 17:09:25 crc kubenswrapper[4754]: I0126 17:09:25.284400 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4c65192f-a05a-45e2-9d56-33ddeb265374","Type":"ContainerDied","Data":"50992540292cc9c045cd8bb4c9e58c8d373ef8d0a1454a811b960e86c77eec5d"} Jan 26 17:09:25 crc kubenswrapper[4754]: I0126 17:09:25.284460 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4c65192f-a05a-45e2-9d56-33ddeb265374","Type":"ContainerDied","Data":"3e9828e079b47eb4e66edc9599b5bdd713959f79d23385e7543f9bcb4e6ef25e"} Jan 26 17:09:25 crc kubenswrapper[4754]: I0126 17:09:25.292467 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"19dee3c5-51dc-46be-8454-9c10c76b3655","Type":"ContainerStarted","Data":"c6d6bd4692f90629ac349f0a045b26074f85cdd87d863143d54053f67bf6cfb5"} Jan 26 17:09:25 crc kubenswrapper[4754]: I0126 17:09:25.295571 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"54272c08-92b7-4512-8ee3-2a782926cda1","Type":"ContainerStarted","Data":"ef3d40953204892aa3f23aa91daa5f3a836343670b1245db122f4e2e696ade5b"} Jan 26 17:09:25 crc kubenswrapper[4754]: I0126 17:09:25.296538 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-6959bf4485-bkv7n" event={"ID":"1622eb04-be06-4e2a-90d6-27a58ac54e60","Type":"ContainerStarted","Data":"68de358117abea3ed1363a89ea041a8331328f94209be04f8e1fcdf3d10b68ba"} Jan 26 17:09:25 crc kubenswrapper[4754]: I0126 17:09:25.303011 4754 generic.go:334] "Generic (PLEG): container finished" podID="4a8fafa1-3d82-4a16-85e1-fd3116bc998a" containerID="68b9e43495925deaaeb41ff5279207dddd543bad6b2dd86201727f1e8bd1577d" exitCode=143 Jan 26 17:09:25 crc kubenswrapper[4754]: I0126 17:09:25.303047 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"4a8fafa1-3d82-4a16-85e1-fd3116bc998a","Type":"ContainerDied","Data":"68b9e43495925deaaeb41ff5279207dddd543bad6b2dd86201727f1e8bd1577d"} Jan 26 17:09:25 crc kubenswrapper[4754]: I0126 17:09:25.323216 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-795f4db4bc-h7mcp" podStartSLOduration=10.323197704 podStartE2EDuration="10.323197704s" podCreationTimestamp="2026-01-26 17:09:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 17:09:25.322172307 +0000 UTC m=+1331.846352741" watchObservedRunningTime="2026-01-26 17:09:25.323197704 +0000 UTC m=+1331.847378138" Jan 26 17:09:25 crc kubenswrapper[4754]: I0126 17:09:25.366148 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=4.779512798 podStartE2EDuration="18.366126317s" podCreationTimestamp="2026-01-26 17:09:07 +0000 UTC" firstStartedPulling="2026-01-26 17:09:10.422807654 +0000 UTC m=+1316.946988088" lastFinishedPulling="2026-01-26 17:09:24.009421163 +0000 UTC m=+1330.533601607" observedRunningTime="2026-01-26 17:09:25.346985377 +0000 UTC m=+1331.871165811" watchObservedRunningTime="2026-01-26 17:09:25.366126317 +0000 UTC m=+1331.890306741" Jan 26 17:09:25 crc kubenswrapper[4754]: I0126 17:09:25.780320 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0dee3ce6-9657-4b8a-9034-7c744db2d209" path="/var/lib/kubelet/pods/0dee3ce6-9657-4b8a-9034-7c744db2d209/volumes" Jan 26 17:09:26 crc kubenswrapper[4754]: I0126 17:09:26.199686 4754 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-5d7f8b89fd-jwf5r" podUID="0dee3ce6-9657-4b8a-9034-7c744db2d209" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.164:9311/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Jan 26 17:09:26 crc kubenswrapper[4754]: I0126 17:09:26.199725 4754 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-5d7f8b89fd-jwf5r" podUID="0dee3ce6-9657-4b8a-9034-7c744db2d209" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.164:9311/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Jan 26 17:09:26 crc kubenswrapper[4754]: I0126 17:09:26.318649 4754 generic.go:334] "Generic (PLEG): container finished" podID="4c65192f-a05a-45e2-9d56-33ddeb265374" containerID="374dc6158046ff5aeb4527bdd32fd9636f239a18d5ba35caa78a117156d7fb06" exitCode=0 Jan 26 17:09:26 crc kubenswrapper[4754]: I0126 17:09:26.318694 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4c65192f-a05a-45e2-9d56-33ddeb265374","Type":"ContainerDied","Data":"374dc6158046ff5aeb4527bdd32fd9636f239a18d5ba35caa78a117156d7fb06"} Jan 26 17:09:26 crc kubenswrapper[4754]: I0126 17:09:26.321899 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"6f82d0f5-47a4-4584-9ae9-046882fa8f61","Type":"ContainerStarted","Data":"eb836410836ef755e634f1dddd915e0296b6426b6a75ab5c090ec925a000a6ba"} Jan 26 17:09:26 crc kubenswrapper[4754]: I0126 17:09:26.324416 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-6959bf4485-bkv7n" event={"ID":"1622eb04-be06-4e2a-90d6-27a58ac54e60","Type":"ContainerStarted","Data":"091a4bab3f52f3669cb33d8fb51f1995b5b4d0507a40119c43855fe0d509fbe6"} Jan 26 17:09:26 crc kubenswrapper[4754]: I0126 17:09:26.324913 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="54272c08-92b7-4512-8ee3-2a782926cda1" containerName="cinder-api" containerID="cri-o://ef3d40953204892aa3f23aa91daa5f3a836343670b1245db122f4e2e696ade5b" gracePeriod=30 Jan 26 17:09:26 crc kubenswrapper[4754]: I0126 17:09:26.325565 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="54272c08-92b7-4512-8ee3-2a782926cda1" containerName="cinder-api-log" containerID="cri-o://e38b379246c733fe550b402977dd4a807b523d7d988b7a00913ef758633e2e36" gracePeriod=30 Jan 26 17:09:26 crc kubenswrapper[4754]: I0126 17:09:26.349452 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=11.349432769 podStartE2EDuration="11.349432769s" podCreationTimestamp="2026-01-26 17:09:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 17:09:26.346935953 +0000 UTC m=+1332.871116387" watchObservedRunningTime="2026-01-26 17:09:26.349432769 +0000 UTC m=+1332.873613203" Jan 26 17:09:27 crc kubenswrapper[4754]: I0126 17:09:27.340912 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"6f82d0f5-47a4-4584-9ae9-046882fa8f61","Type":"ContainerStarted","Data":"66f684c6f13cd3e1ff23df6b19e556f74dc2dd34475bbf370be3391e82f922c0"} Jan 26 17:09:27 crc kubenswrapper[4754]: I0126 17:09:27.348284 4754 generic.go:334] "Generic (PLEG): container finished" podID="54272c08-92b7-4512-8ee3-2a782926cda1" containerID="ef3d40953204892aa3f23aa91daa5f3a836343670b1245db122f4e2e696ade5b" exitCode=0 Jan 26 17:09:27 crc kubenswrapper[4754]: I0126 17:09:27.348322 4754 generic.go:334] "Generic (PLEG): container finished" podID="54272c08-92b7-4512-8ee3-2a782926cda1" containerID="e38b379246c733fe550b402977dd4a807b523d7d988b7a00913ef758633e2e36" exitCode=143 Jan 26 17:09:27 crc kubenswrapper[4754]: I0126 17:09:27.348389 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"54272c08-92b7-4512-8ee3-2a782926cda1","Type":"ContainerDied","Data":"ef3d40953204892aa3f23aa91daa5f3a836343670b1245db122f4e2e696ade5b"} Jan 26 17:09:27 crc kubenswrapper[4754]: I0126 17:09:27.348415 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"54272c08-92b7-4512-8ee3-2a782926cda1","Type":"ContainerDied","Data":"e38b379246c733fe550b402977dd4a807b523d7d988b7a00913ef758633e2e36"} Jan 26 17:09:27 crc kubenswrapper[4754]: I0126 17:09:27.352936 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-6959bf4485-bkv7n" event={"ID":"1622eb04-be06-4e2a-90d6-27a58ac54e60","Type":"ContainerStarted","Data":"19a4efed16672e16f48a19b6c2723765bea2f0083ea3ae31b237bbb8daca2c6d"} Jan 26 17:09:27 crc kubenswrapper[4754]: I0126 17:09:27.353875 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-6959bf4485-bkv7n" Jan 26 17:09:27 crc kubenswrapper[4754]: I0126 17:09:27.353955 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-6959bf4485-bkv7n" Jan 26 17:09:27 crc kubenswrapper[4754]: I0126 17:09:27.365514 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=4.804531745 podStartE2EDuration="12.365496728s" podCreationTimestamp="2026-01-26 17:09:15 +0000 UTC" firstStartedPulling="2026-01-26 17:09:16.309298168 +0000 UTC m=+1322.833478602" lastFinishedPulling="2026-01-26 17:09:23.870263151 +0000 UTC m=+1330.394443585" observedRunningTime="2026-01-26 17:09:27.364921774 +0000 UTC m=+1333.889102218" watchObservedRunningTime="2026-01-26 17:09:27.365496728 +0000 UTC m=+1333.889677162" Jan 26 17:09:27 crc kubenswrapper[4754]: I0126 17:09:27.407116 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-6959bf4485-bkv7n" podStartSLOduration=3.407098187 podStartE2EDuration="3.407098187s" podCreationTimestamp="2026-01-26 17:09:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 17:09:27.389802815 +0000 UTC m=+1333.913983249" watchObservedRunningTime="2026-01-26 17:09:27.407098187 +0000 UTC m=+1333.931278621" Jan 26 17:09:27 crc kubenswrapper[4754]: I0126 17:09:27.589730 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Jan 26 17:09:27 crc kubenswrapper[4754]: I0126 17:09:27.701280 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/54272c08-92b7-4512-8ee3-2a782926cda1-scripts\") pod \"54272c08-92b7-4512-8ee3-2a782926cda1\" (UID: \"54272c08-92b7-4512-8ee3-2a782926cda1\") " Jan 26 17:09:27 crc kubenswrapper[4754]: I0126 17:09:27.701317 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/54272c08-92b7-4512-8ee3-2a782926cda1-etc-machine-id\") pod \"54272c08-92b7-4512-8ee3-2a782926cda1\" (UID: \"54272c08-92b7-4512-8ee3-2a782926cda1\") " Jan 26 17:09:27 crc kubenswrapper[4754]: I0126 17:09:27.701348 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54272c08-92b7-4512-8ee3-2a782926cda1-combined-ca-bundle\") pod \"54272c08-92b7-4512-8ee3-2a782926cda1\" (UID: \"54272c08-92b7-4512-8ee3-2a782926cda1\") " Jan 26 17:09:27 crc kubenswrapper[4754]: I0126 17:09:27.701373 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/54272c08-92b7-4512-8ee3-2a782926cda1-config-data\") pod \"54272c08-92b7-4512-8ee3-2a782926cda1\" (UID: \"54272c08-92b7-4512-8ee3-2a782926cda1\") " Jan 26 17:09:27 crc kubenswrapper[4754]: I0126 17:09:27.701394 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-czd2j\" (UniqueName: \"kubernetes.io/projected/54272c08-92b7-4512-8ee3-2a782926cda1-kube-api-access-czd2j\") pod \"54272c08-92b7-4512-8ee3-2a782926cda1\" (UID: \"54272c08-92b7-4512-8ee3-2a782926cda1\") " Jan 26 17:09:27 crc kubenswrapper[4754]: I0126 17:09:27.701438 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/54272c08-92b7-4512-8ee3-2a782926cda1-config-data-custom\") pod \"54272c08-92b7-4512-8ee3-2a782926cda1\" (UID: \"54272c08-92b7-4512-8ee3-2a782926cda1\") " Jan 26 17:09:27 crc kubenswrapper[4754]: I0126 17:09:27.701495 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/54272c08-92b7-4512-8ee3-2a782926cda1-logs\") pod \"54272c08-92b7-4512-8ee3-2a782926cda1\" (UID: \"54272c08-92b7-4512-8ee3-2a782926cda1\") " Jan 26 17:09:27 crc kubenswrapper[4754]: I0126 17:09:27.701728 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/54272c08-92b7-4512-8ee3-2a782926cda1-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "54272c08-92b7-4512-8ee3-2a782926cda1" (UID: "54272c08-92b7-4512-8ee3-2a782926cda1"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 17:09:27 crc kubenswrapper[4754]: I0126 17:09:27.702052 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/54272c08-92b7-4512-8ee3-2a782926cda1-logs" (OuterVolumeSpecName: "logs") pod "54272c08-92b7-4512-8ee3-2a782926cda1" (UID: "54272c08-92b7-4512-8ee3-2a782926cda1"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:09:27 crc kubenswrapper[4754]: I0126 17:09:27.702065 4754 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/54272c08-92b7-4512-8ee3-2a782926cda1-etc-machine-id\") on node \"crc\" DevicePath \"\"" Jan 26 17:09:27 crc kubenswrapper[4754]: I0126 17:09:27.716034 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/54272c08-92b7-4512-8ee3-2a782926cda1-kube-api-access-czd2j" (OuterVolumeSpecName: "kube-api-access-czd2j") pod "54272c08-92b7-4512-8ee3-2a782926cda1" (UID: "54272c08-92b7-4512-8ee3-2a782926cda1"). InnerVolumeSpecName "kube-api-access-czd2j". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:09:27 crc kubenswrapper[4754]: I0126 17:09:27.719011 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/54272c08-92b7-4512-8ee3-2a782926cda1-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "54272c08-92b7-4512-8ee3-2a782926cda1" (UID: "54272c08-92b7-4512-8ee3-2a782926cda1"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:09:27 crc kubenswrapper[4754]: I0126 17:09:27.722195 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/54272c08-92b7-4512-8ee3-2a782926cda1-scripts" (OuterVolumeSpecName: "scripts") pod "54272c08-92b7-4512-8ee3-2a782926cda1" (UID: "54272c08-92b7-4512-8ee3-2a782926cda1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:09:27 crc kubenswrapper[4754]: I0126 17:09:27.739010 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/54272c08-92b7-4512-8ee3-2a782926cda1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "54272c08-92b7-4512-8ee3-2a782926cda1" (UID: "54272c08-92b7-4512-8ee3-2a782926cda1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:09:27 crc kubenswrapper[4754]: I0126 17:09:27.769756 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/54272c08-92b7-4512-8ee3-2a782926cda1-config-data" (OuterVolumeSpecName: "config-data") pod "54272c08-92b7-4512-8ee3-2a782926cda1" (UID: "54272c08-92b7-4512-8ee3-2a782926cda1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:09:27 crc kubenswrapper[4754]: I0126 17:09:27.803422 4754 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/54272c08-92b7-4512-8ee3-2a782926cda1-config-data-custom\") on node \"crc\" DevicePath \"\"" Jan 26 17:09:27 crc kubenswrapper[4754]: I0126 17:09:27.803468 4754 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/54272c08-92b7-4512-8ee3-2a782926cda1-logs\") on node \"crc\" DevicePath \"\"" Jan 26 17:09:27 crc kubenswrapper[4754]: I0126 17:09:27.803481 4754 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/54272c08-92b7-4512-8ee3-2a782926cda1-scripts\") on node \"crc\" DevicePath \"\"" Jan 26 17:09:27 crc kubenswrapper[4754]: I0126 17:09:27.803494 4754 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54272c08-92b7-4512-8ee3-2a782926cda1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 26 17:09:27 crc kubenswrapper[4754]: I0126 17:09:27.803506 4754 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/54272c08-92b7-4512-8ee3-2a782926cda1-config-data\") on node \"crc\" DevicePath \"\"" Jan 26 17:09:27 crc kubenswrapper[4754]: I0126 17:09:27.803518 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-czd2j\" (UniqueName: \"kubernetes.io/projected/54272c08-92b7-4512-8ee3-2a782926cda1-kube-api-access-czd2j\") on node \"crc\" DevicePath \"\"" Jan 26 17:09:28 crc kubenswrapper[4754]: I0126 17:09:28.250972 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Jan 26 17:09:28 crc kubenswrapper[4754]: I0126 17:09:28.312313 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"4a8fafa1-3d82-4a16-85e1-fd3116bc998a\" (UID: \"4a8fafa1-3d82-4a16-85e1-fd3116bc998a\") " Jan 26 17:09:28 crc kubenswrapper[4754]: I0126 17:09:28.312418 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4a8fafa1-3d82-4a16-85e1-fd3116bc998a-config-data\") pod \"4a8fafa1-3d82-4a16-85e1-fd3116bc998a\" (UID: \"4a8fafa1-3d82-4a16-85e1-fd3116bc998a\") " Jan 26 17:09:28 crc kubenswrapper[4754]: I0126 17:09:28.312458 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4a8fafa1-3d82-4a16-85e1-fd3116bc998a-public-tls-certs\") pod \"4a8fafa1-3d82-4a16-85e1-fd3116bc998a\" (UID: \"4a8fafa1-3d82-4a16-85e1-fd3116bc998a\") " Jan 26 17:09:28 crc kubenswrapper[4754]: I0126 17:09:28.312553 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g62ms\" (UniqueName: \"kubernetes.io/projected/4a8fafa1-3d82-4a16-85e1-fd3116bc998a-kube-api-access-g62ms\") pod \"4a8fafa1-3d82-4a16-85e1-fd3116bc998a\" (UID: \"4a8fafa1-3d82-4a16-85e1-fd3116bc998a\") " Jan 26 17:09:28 crc kubenswrapper[4754]: I0126 17:09:28.312630 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4a8fafa1-3d82-4a16-85e1-fd3116bc998a-httpd-run\") pod \"4a8fafa1-3d82-4a16-85e1-fd3116bc998a\" (UID: \"4a8fafa1-3d82-4a16-85e1-fd3116bc998a\") " Jan 26 17:09:28 crc kubenswrapper[4754]: I0126 17:09:28.312765 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4a8fafa1-3d82-4a16-85e1-fd3116bc998a-scripts\") pod \"4a8fafa1-3d82-4a16-85e1-fd3116bc998a\" (UID: \"4a8fafa1-3d82-4a16-85e1-fd3116bc998a\") " Jan 26 17:09:28 crc kubenswrapper[4754]: I0126 17:09:28.312826 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a8fafa1-3d82-4a16-85e1-fd3116bc998a-combined-ca-bundle\") pod \"4a8fafa1-3d82-4a16-85e1-fd3116bc998a\" (UID: \"4a8fafa1-3d82-4a16-85e1-fd3116bc998a\") " Jan 26 17:09:28 crc kubenswrapper[4754]: I0126 17:09:28.312874 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4a8fafa1-3d82-4a16-85e1-fd3116bc998a-logs\") pod \"4a8fafa1-3d82-4a16-85e1-fd3116bc998a\" (UID: \"4a8fafa1-3d82-4a16-85e1-fd3116bc998a\") " Jan 26 17:09:28 crc kubenswrapper[4754]: I0126 17:09:28.313109 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4a8fafa1-3d82-4a16-85e1-fd3116bc998a-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "4a8fafa1-3d82-4a16-85e1-fd3116bc998a" (UID: "4a8fafa1-3d82-4a16-85e1-fd3116bc998a"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:09:28 crc kubenswrapper[4754]: I0126 17:09:28.313478 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4a8fafa1-3d82-4a16-85e1-fd3116bc998a-logs" (OuterVolumeSpecName: "logs") pod "4a8fafa1-3d82-4a16-85e1-fd3116bc998a" (UID: "4a8fafa1-3d82-4a16-85e1-fd3116bc998a"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:09:28 crc kubenswrapper[4754]: I0126 17:09:28.313587 4754 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4a8fafa1-3d82-4a16-85e1-fd3116bc998a-httpd-run\") on node \"crc\" DevicePath \"\"" Jan 26 17:09:28 crc kubenswrapper[4754]: I0126 17:09:28.342207 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4a8fafa1-3d82-4a16-85e1-fd3116bc998a-scripts" (OuterVolumeSpecName: "scripts") pod "4a8fafa1-3d82-4a16-85e1-fd3116bc998a" (UID: "4a8fafa1-3d82-4a16-85e1-fd3116bc998a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:09:28 crc kubenswrapper[4754]: I0126 17:09:28.360659 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4a8fafa1-3d82-4a16-85e1-fd3116bc998a-kube-api-access-g62ms" (OuterVolumeSpecName: "kube-api-access-g62ms") pod "4a8fafa1-3d82-4a16-85e1-fd3116bc998a" (UID: "4a8fafa1-3d82-4a16-85e1-fd3116bc998a"). InnerVolumeSpecName "kube-api-access-g62ms". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:09:28 crc kubenswrapper[4754]: I0126 17:09:28.370896 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage11-crc" (OuterVolumeSpecName: "glance") pod "4a8fafa1-3d82-4a16-85e1-fd3116bc998a" (UID: "4a8fafa1-3d82-4a16-85e1-fd3116bc998a"). InnerVolumeSpecName "local-storage11-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Jan 26 17:09:28 crc kubenswrapper[4754]: I0126 17:09:28.403983 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4a8fafa1-3d82-4a16-85e1-fd3116bc998a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4a8fafa1-3d82-4a16-85e1-fd3116bc998a" (UID: "4a8fafa1-3d82-4a16-85e1-fd3116bc998a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:09:28 crc kubenswrapper[4754]: I0126 17:09:28.417922 4754 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a8fafa1-3d82-4a16-85e1-fd3116bc998a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 26 17:09:28 crc kubenswrapper[4754]: I0126 17:09:28.417964 4754 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4a8fafa1-3d82-4a16-85e1-fd3116bc998a-logs\") on node \"crc\" DevicePath \"\"" Jan 26 17:09:28 crc kubenswrapper[4754]: I0126 17:09:28.418003 4754 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" " Jan 26 17:09:28 crc kubenswrapper[4754]: I0126 17:09:28.418016 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g62ms\" (UniqueName: \"kubernetes.io/projected/4a8fafa1-3d82-4a16-85e1-fd3116bc998a-kube-api-access-g62ms\") on node \"crc\" DevicePath \"\"" Jan 26 17:09:28 crc kubenswrapper[4754]: I0126 17:09:28.418055 4754 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4a8fafa1-3d82-4a16-85e1-fd3116bc998a-scripts\") on node \"crc\" DevicePath \"\"" Jan 26 17:09:28 crc kubenswrapper[4754]: I0126 17:09:28.423608 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"54272c08-92b7-4512-8ee3-2a782926cda1","Type":"ContainerDied","Data":"ef80ca84ce39ac6f9b5f7493ec39dbf65d993038fc52b5bd85fdf98575a1dbe3"} Jan 26 17:09:28 crc kubenswrapper[4754]: I0126 17:09:28.423685 4754 scope.go:117] "RemoveContainer" containerID="ef3d40953204892aa3f23aa91daa5f3a836343670b1245db122f4e2e696ade5b" Jan 26 17:09:28 crc kubenswrapper[4754]: I0126 17:09:28.423916 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Jan 26 17:09:28 crc kubenswrapper[4754]: I0126 17:09:28.433842 4754 generic.go:334] "Generic (PLEG): container finished" podID="4a8fafa1-3d82-4a16-85e1-fd3116bc998a" containerID="0c78645b52c9649d8e28a70685c6baf4ee822767bcf49eaf0c693d92ab3e3521" exitCode=0 Jan 26 17:09:28 crc kubenswrapper[4754]: I0126 17:09:28.434084 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"4a8fafa1-3d82-4a16-85e1-fd3116bc998a","Type":"ContainerDied","Data":"0c78645b52c9649d8e28a70685c6baf4ee822767bcf49eaf0c693d92ab3e3521"} Jan 26 17:09:28 crc kubenswrapper[4754]: I0126 17:09:28.434117 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"4a8fafa1-3d82-4a16-85e1-fd3116bc998a","Type":"ContainerDied","Data":"1816ede461917fc9aed1037d8a2a51d98ac1b2d9c969bae3690e1f12841c4111"} Jan 26 17:09:28 crc kubenswrapper[4754]: I0126 17:09:28.434155 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Jan 26 17:09:28 crc kubenswrapper[4754]: I0126 17:09:28.450433 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4a8fafa1-3d82-4a16-85e1-fd3116bc998a-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "4a8fafa1-3d82-4a16-85e1-fd3116bc998a" (UID: "4a8fafa1-3d82-4a16-85e1-fd3116bc998a"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:09:28 crc kubenswrapper[4754]: I0126 17:09:28.459864 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Jan 26 17:09:28 crc kubenswrapper[4754]: I0126 17:09:28.468754 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4a8fafa1-3d82-4a16-85e1-fd3116bc998a-config-data" (OuterVolumeSpecName: "config-data") pod "4a8fafa1-3d82-4a16-85e1-fd3116bc998a" (UID: "4a8fafa1-3d82-4a16-85e1-fd3116bc998a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:09:28 crc kubenswrapper[4754]: I0126 17:09:28.470849 4754 scope.go:117] "RemoveContainer" containerID="e38b379246c733fe550b402977dd4a807b523d7d988b7a00913ef758633e2e36" Jan 26 17:09:28 crc kubenswrapper[4754]: I0126 17:09:28.486532 4754 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage11-crc" (UniqueName: "kubernetes.io/local-volume/local-storage11-crc") on node "crc" Jan 26 17:09:28 crc kubenswrapper[4754]: I0126 17:09:28.490874 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Jan 26 17:09:28 crc kubenswrapper[4754]: I0126 17:09:28.498019 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Jan 26 17:09:28 crc kubenswrapper[4754]: E0126 17:09:28.498530 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4a8fafa1-3d82-4a16-85e1-fd3116bc998a" containerName="glance-log" Jan 26 17:09:28 crc kubenswrapper[4754]: I0126 17:09:28.498549 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="4a8fafa1-3d82-4a16-85e1-fd3116bc998a" containerName="glance-log" Jan 26 17:09:28 crc kubenswrapper[4754]: E0126 17:09:28.498564 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4a8fafa1-3d82-4a16-85e1-fd3116bc998a" containerName="glance-httpd" Jan 26 17:09:28 crc kubenswrapper[4754]: I0126 17:09:28.498571 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="4a8fafa1-3d82-4a16-85e1-fd3116bc998a" containerName="glance-httpd" Jan 26 17:09:28 crc kubenswrapper[4754]: E0126 17:09:28.498586 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54272c08-92b7-4512-8ee3-2a782926cda1" containerName="cinder-api-log" Jan 26 17:09:28 crc kubenswrapper[4754]: I0126 17:09:28.498592 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="54272c08-92b7-4512-8ee3-2a782926cda1" containerName="cinder-api-log" Jan 26 17:09:28 crc kubenswrapper[4754]: E0126 17:09:28.498599 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54272c08-92b7-4512-8ee3-2a782926cda1" containerName="cinder-api" Jan 26 17:09:28 crc kubenswrapper[4754]: I0126 17:09:28.498606 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="54272c08-92b7-4512-8ee3-2a782926cda1" containerName="cinder-api" Jan 26 17:09:28 crc kubenswrapper[4754]: I0126 17:09:28.498780 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="54272c08-92b7-4512-8ee3-2a782926cda1" containerName="cinder-api" Jan 26 17:09:28 crc kubenswrapper[4754]: I0126 17:09:28.498791 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="4a8fafa1-3d82-4a16-85e1-fd3116bc998a" containerName="glance-log" Jan 26 17:09:28 crc kubenswrapper[4754]: I0126 17:09:28.498809 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="54272c08-92b7-4512-8ee3-2a782926cda1" containerName="cinder-api-log" Jan 26 17:09:28 crc kubenswrapper[4754]: I0126 17:09:28.498823 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="4a8fafa1-3d82-4a16-85e1-fd3116bc998a" containerName="glance-httpd" Jan 26 17:09:28 crc kubenswrapper[4754]: I0126 17:09:28.499819 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Jan 26 17:09:28 crc kubenswrapper[4754]: I0126 17:09:28.503457 4754 scope.go:117] "RemoveContainer" containerID="0c78645b52c9649d8e28a70685c6baf4ee822767bcf49eaf0c693d92ab3e3521" Jan 26 17:09:28 crc kubenswrapper[4754]: I0126 17:09:28.503642 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Jan 26 17:09:28 crc kubenswrapper[4754]: I0126 17:09:28.503811 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Jan 26 17:09:28 crc kubenswrapper[4754]: I0126 17:09:28.504048 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Jan 26 17:09:28 crc kubenswrapper[4754]: I0126 17:09:28.509566 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Jan 26 17:09:28 crc kubenswrapper[4754]: I0126 17:09:28.522606 4754 reconciler_common.go:293] "Volume detached for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" DevicePath \"\"" Jan 26 17:09:28 crc kubenswrapper[4754]: I0126 17:09:28.522634 4754 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4a8fafa1-3d82-4a16-85e1-fd3116bc998a-config-data\") on node \"crc\" DevicePath \"\"" Jan 26 17:09:28 crc kubenswrapper[4754]: I0126 17:09:28.522643 4754 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4a8fafa1-3d82-4a16-85e1-fd3116bc998a-public-tls-certs\") on node \"crc\" DevicePath \"\"" Jan 26 17:09:28 crc kubenswrapper[4754]: I0126 17:09:28.546920 4754 scope.go:117] "RemoveContainer" containerID="68b9e43495925deaaeb41ff5279207dddd543bad6b2dd86201727f1e8bd1577d" Jan 26 17:09:28 crc kubenswrapper[4754]: I0126 17:09:28.568136 4754 scope.go:117] "RemoveContainer" containerID="0c78645b52c9649d8e28a70685c6baf4ee822767bcf49eaf0c693d92ab3e3521" Jan 26 17:09:28 crc kubenswrapper[4754]: E0126 17:09:28.568610 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0c78645b52c9649d8e28a70685c6baf4ee822767bcf49eaf0c693d92ab3e3521\": container with ID starting with 0c78645b52c9649d8e28a70685c6baf4ee822767bcf49eaf0c693d92ab3e3521 not found: ID does not exist" containerID="0c78645b52c9649d8e28a70685c6baf4ee822767bcf49eaf0c693d92ab3e3521" Jan 26 17:09:28 crc kubenswrapper[4754]: I0126 17:09:28.568724 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0c78645b52c9649d8e28a70685c6baf4ee822767bcf49eaf0c693d92ab3e3521"} err="failed to get container status \"0c78645b52c9649d8e28a70685c6baf4ee822767bcf49eaf0c693d92ab3e3521\": rpc error: code = NotFound desc = could not find container \"0c78645b52c9649d8e28a70685c6baf4ee822767bcf49eaf0c693d92ab3e3521\": container with ID starting with 0c78645b52c9649d8e28a70685c6baf4ee822767bcf49eaf0c693d92ab3e3521 not found: ID does not exist" Jan 26 17:09:28 crc kubenswrapper[4754]: I0126 17:09:28.568843 4754 scope.go:117] "RemoveContainer" containerID="68b9e43495925deaaeb41ff5279207dddd543bad6b2dd86201727f1e8bd1577d" Jan 26 17:09:28 crc kubenswrapper[4754]: E0126 17:09:28.569261 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"68b9e43495925deaaeb41ff5279207dddd543bad6b2dd86201727f1e8bd1577d\": container with ID starting with 68b9e43495925deaaeb41ff5279207dddd543bad6b2dd86201727f1e8bd1577d not found: ID does not exist" containerID="68b9e43495925deaaeb41ff5279207dddd543bad6b2dd86201727f1e8bd1577d" Jan 26 17:09:28 crc kubenswrapper[4754]: I0126 17:09:28.569312 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"68b9e43495925deaaeb41ff5279207dddd543bad6b2dd86201727f1e8bd1577d"} err="failed to get container status \"68b9e43495925deaaeb41ff5279207dddd543bad6b2dd86201727f1e8bd1577d\": rpc error: code = NotFound desc = could not find container \"68b9e43495925deaaeb41ff5279207dddd543bad6b2dd86201727f1e8bd1577d\": container with ID starting with 68b9e43495925deaaeb41ff5279207dddd543bad6b2dd86201727f1e8bd1577d not found: ID does not exist" Jan 26 17:09:28 crc kubenswrapper[4754]: I0126 17:09:28.623765 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/529bf77e-614d-4354-94bd-e6383f353920-config-data-custom\") pod \"cinder-api-0\" (UID: \"529bf77e-614d-4354-94bd-e6383f353920\") " pod="openstack/cinder-api-0" Jan 26 17:09:28 crc kubenswrapper[4754]: I0126 17:09:28.624016 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/529bf77e-614d-4354-94bd-e6383f353920-public-tls-certs\") pod \"cinder-api-0\" (UID: \"529bf77e-614d-4354-94bd-e6383f353920\") " pod="openstack/cinder-api-0" Jan 26 17:09:28 crc kubenswrapper[4754]: I0126 17:09:28.624140 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s2v55\" (UniqueName: \"kubernetes.io/projected/529bf77e-614d-4354-94bd-e6383f353920-kube-api-access-s2v55\") pod \"cinder-api-0\" (UID: \"529bf77e-614d-4354-94bd-e6383f353920\") " pod="openstack/cinder-api-0" Jan 26 17:09:28 crc kubenswrapper[4754]: I0126 17:09:28.624238 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/529bf77e-614d-4354-94bd-e6383f353920-scripts\") pod \"cinder-api-0\" (UID: \"529bf77e-614d-4354-94bd-e6383f353920\") " pod="openstack/cinder-api-0" Jan 26 17:09:28 crc kubenswrapper[4754]: I0126 17:09:28.624398 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/529bf77e-614d-4354-94bd-e6383f353920-etc-machine-id\") pod \"cinder-api-0\" (UID: \"529bf77e-614d-4354-94bd-e6383f353920\") " pod="openstack/cinder-api-0" Jan 26 17:09:28 crc kubenswrapper[4754]: I0126 17:09:28.624476 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/529bf77e-614d-4354-94bd-e6383f353920-logs\") pod \"cinder-api-0\" (UID: \"529bf77e-614d-4354-94bd-e6383f353920\") " pod="openstack/cinder-api-0" Jan 26 17:09:28 crc kubenswrapper[4754]: I0126 17:09:28.624558 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/529bf77e-614d-4354-94bd-e6383f353920-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"529bf77e-614d-4354-94bd-e6383f353920\") " pod="openstack/cinder-api-0" Jan 26 17:09:28 crc kubenswrapper[4754]: I0126 17:09:28.624685 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/529bf77e-614d-4354-94bd-e6383f353920-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"529bf77e-614d-4354-94bd-e6383f353920\") " pod="openstack/cinder-api-0" Jan 26 17:09:28 crc kubenswrapper[4754]: I0126 17:09:28.624784 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/529bf77e-614d-4354-94bd-e6383f353920-config-data\") pod \"cinder-api-0\" (UID: \"529bf77e-614d-4354-94bd-e6383f353920\") " pod="openstack/cinder-api-0" Jan 26 17:09:28 crc kubenswrapper[4754]: I0126 17:09:28.727013 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/529bf77e-614d-4354-94bd-e6383f353920-config-data-custom\") pod \"cinder-api-0\" (UID: \"529bf77e-614d-4354-94bd-e6383f353920\") " pod="openstack/cinder-api-0" Jan 26 17:09:28 crc kubenswrapper[4754]: I0126 17:09:28.727053 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/529bf77e-614d-4354-94bd-e6383f353920-public-tls-certs\") pod \"cinder-api-0\" (UID: \"529bf77e-614d-4354-94bd-e6383f353920\") " pod="openstack/cinder-api-0" Jan 26 17:09:28 crc kubenswrapper[4754]: I0126 17:09:28.727076 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2v55\" (UniqueName: \"kubernetes.io/projected/529bf77e-614d-4354-94bd-e6383f353920-kube-api-access-s2v55\") pod \"cinder-api-0\" (UID: \"529bf77e-614d-4354-94bd-e6383f353920\") " pod="openstack/cinder-api-0" Jan 26 17:09:28 crc kubenswrapper[4754]: I0126 17:09:28.727105 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/529bf77e-614d-4354-94bd-e6383f353920-scripts\") pod \"cinder-api-0\" (UID: \"529bf77e-614d-4354-94bd-e6383f353920\") " pod="openstack/cinder-api-0" Jan 26 17:09:28 crc kubenswrapper[4754]: I0126 17:09:28.727968 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/529bf77e-614d-4354-94bd-e6383f353920-etc-machine-id\") pod \"cinder-api-0\" (UID: \"529bf77e-614d-4354-94bd-e6383f353920\") " pod="openstack/cinder-api-0" Jan 26 17:09:28 crc kubenswrapper[4754]: I0126 17:09:28.727986 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/529bf77e-614d-4354-94bd-e6383f353920-logs\") pod \"cinder-api-0\" (UID: \"529bf77e-614d-4354-94bd-e6383f353920\") " pod="openstack/cinder-api-0" Jan 26 17:09:28 crc kubenswrapper[4754]: I0126 17:09:28.728005 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/529bf77e-614d-4354-94bd-e6383f353920-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"529bf77e-614d-4354-94bd-e6383f353920\") " pod="openstack/cinder-api-0" Jan 26 17:09:28 crc kubenswrapper[4754]: I0126 17:09:28.728047 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/529bf77e-614d-4354-94bd-e6383f353920-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"529bf77e-614d-4354-94bd-e6383f353920\") " pod="openstack/cinder-api-0" Jan 26 17:09:28 crc kubenswrapper[4754]: I0126 17:09:28.728070 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/529bf77e-614d-4354-94bd-e6383f353920-config-data\") pod \"cinder-api-0\" (UID: \"529bf77e-614d-4354-94bd-e6383f353920\") " pod="openstack/cinder-api-0" Jan 26 17:09:28 crc kubenswrapper[4754]: I0126 17:09:28.728316 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/529bf77e-614d-4354-94bd-e6383f353920-logs\") pod \"cinder-api-0\" (UID: \"529bf77e-614d-4354-94bd-e6383f353920\") " pod="openstack/cinder-api-0" Jan 26 17:09:28 crc kubenswrapper[4754]: I0126 17:09:28.728369 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/529bf77e-614d-4354-94bd-e6383f353920-etc-machine-id\") pod \"cinder-api-0\" (UID: \"529bf77e-614d-4354-94bd-e6383f353920\") " pod="openstack/cinder-api-0" Jan 26 17:09:28 crc kubenswrapper[4754]: I0126 17:09:28.732435 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/529bf77e-614d-4354-94bd-e6383f353920-public-tls-certs\") pod \"cinder-api-0\" (UID: \"529bf77e-614d-4354-94bd-e6383f353920\") " pod="openstack/cinder-api-0" Jan 26 17:09:28 crc kubenswrapper[4754]: I0126 17:09:28.732550 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/529bf77e-614d-4354-94bd-e6383f353920-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"529bf77e-614d-4354-94bd-e6383f353920\") " pod="openstack/cinder-api-0" Jan 26 17:09:28 crc kubenswrapper[4754]: I0126 17:09:28.734897 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/529bf77e-614d-4354-94bd-e6383f353920-config-data\") pod \"cinder-api-0\" (UID: \"529bf77e-614d-4354-94bd-e6383f353920\") " pod="openstack/cinder-api-0" Jan 26 17:09:28 crc kubenswrapper[4754]: I0126 17:09:28.735908 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/529bf77e-614d-4354-94bd-e6383f353920-scripts\") pod \"cinder-api-0\" (UID: \"529bf77e-614d-4354-94bd-e6383f353920\") " pod="openstack/cinder-api-0" Jan 26 17:09:28 crc kubenswrapper[4754]: I0126 17:09:28.736166 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/529bf77e-614d-4354-94bd-e6383f353920-config-data-custom\") pod \"cinder-api-0\" (UID: \"529bf77e-614d-4354-94bd-e6383f353920\") " pod="openstack/cinder-api-0" Jan 26 17:09:28 crc kubenswrapper[4754]: I0126 17:09:28.736566 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/529bf77e-614d-4354-94bd-e6383f353920-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"529bf77e-614d-4354-94bd-e6383f353920\") " pod="openstack/cinder-api-0" Jan 26 17:09:28 crc kubenswrapper[4754]: I0126 17:09:28.746989 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2v55\" (UniqueName: \"kubernetes.io/projected/529bf77e-614d-4354-94bd-e6383f353920-kube-api-access-s2v55\") pod \"cinder-api-0\" (UID: \"529bf77e-614d-4354-94bd-e6383f353920\") " pod="openstack/cinder-api-0" Jan 26 17:09:28 crc kubenswrapper[4754]: I0126 17:09:28.829458 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Jan 26 17:09:28 crc kubenswrapper[4754]: I0126 17:09:28.848582 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Jan 26 17:09:28 crc kubenswrapper[4754]: I0126 17:09:28.858427 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Jan 26 17:09:28 crc kubenswrapper[4754]: I0126 17:09:28.877692 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Jan 26 17:09:28 crc kubenswrapper[4754]: I0126 17:09:28.880574 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Jan 26 17:09:28 crc kubenswrapper[4754]: I0126 17:09:28.890681 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Jan 26 17:09:28 crc kubenswrapper[4754]: I0126 17:09:28.890921 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Jan 26 17:09:28 crc kubenswrapper[4754]: I0126 17:09:28.907893 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Jan 26 17:09:28 crc kubenswrapper[4754]: I0126 17:09:28.931778 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/724e114a-1ace-4455-846f-d7ab65c593f1-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"724e114a-1ace-4455-846f-d7ab65c593f1\") " pod="openstack/glance-default-external-api-0" Jan 26 17:09:28 crc kubenswrapper[4754]: I0126 17:09:28.931860 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mxbmg\" (UniqueName: \"kubernetes.io/projected/724e114a-1ace-4455-846f-d7ab65c593f1-kube-api-access-mxbmg\") pod \"glance-default-external-api-0\" (UID: \"724e114a-1ace-4455-846f-d7ab65c593f1\") " pod="openstack/glance-default-external-api-0" Jan 26 17:09:28 crc kubenswrapper[4754]: I0126 17:09:28.932056 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"724e114a-1ace-4455-846f-d7ab65c593f1\") " pod="openstack/glance-default-external-api-0" Jan 26 17:09:28 crc kubenswrapper[4754]: I0126 17:09:28.932144 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/724e114a-1ace-4455-846f-d7ab65c593f1-logs\") pod \"glance-default-external-api-0\" (UID: \"724e114a-1ace-4455-846f-d7ab65c593f1\") " pod="openstack/glance-default-external-api-0" Jan 26 17:09:28 crc kubenswrapper[4754]: I0126 17:09:28.932252 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/724e114a-1ace-4455-846f-d7ab65c593f1-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"724e114a-1ace-4455-846f-d7ab65c593f1\") " pod="openstack/glance-default-external-api-0" Jan 26 17:09:28 crc kubenswrapper[4754]: I0126 17:09:28.932416 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/724e114a-1ace-4455-846f-d7ab65c593f1-config-data\") pod \"glance-default-external-api-0\" (UID: \"724e114a-1ace-4455-846f-d7ab65c593f1\") " pod="openstack/glance-default-external-api-0" Jan 26 17:09:28 crc kubenswrapper[4754]: I0126 17:09:28.932502 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/724e114a-1ace-4455-846f-d7ab65c593f1-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"724e114a-1ace-4455-846f-d7ab65c593f1\") " pod="openstack/glance-default-external-api-0" Jan 26 17:09:28 crc kubenswrapper[4754]: I0126 17:09:28.932543 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/724e114a-1ace-4455-846f-d7ab65c593f1-scripts\") pod \"glance-default-external-api-0\" (UID: \"724e114a-1ace-4455-846f-d7ab65c593f1\") " pod="openstack/glance-default-external-api-0" Jan 26 17:09:29 crc kubenswrapper[4754]: I0126 17:09:29.035387 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/724e114a-1ace-4455-846f-d7ab65c593f1-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"724e114a-1ace-4455-846f-d7ab65c593f1\") " pod="openstack/glance-default-external-api-0" Jan 26 17:09:29 crc kubenswrapper[4754]: I0126 17:09:29.035699 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mxbmg\" (UniqueName: \"kubernetes.io/projected/724e114a-1ace-4455-846f-d7ab65c593f1-kube-api-access-mxbmg\") pod \"glance-default-external-api-0\" (UID: \"724e114a-1ace-4455-846f-d7ab65c593f1\") " pod="openstack/glance-default-external-api-0" Jan 26 17:09:29 crc kubenswrapper[4754]: I0126 17:09:29.035729 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"724e114a-1ace-4455-846f-d7ab65c593f1\") " pod="openstack/glance-default-external-api-0" Jan 26 17:09:29 crc kubenswrapper[4754]: I0126 17:09:29.035754 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/724e114a-1ace-4455-846f-d7ab65c593f1-logs\") pod \"glance-default-external-api-0\" (UID: \"724e114a-1ace-4455-846f-d7ab65c593f1\") " pod="openstack/glance-default-external-api-0" Jan 26 17:09:29 crc kubenswrapper[4754]: I0126 17:09:29.035798 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/724e114a-1ace-4455-846f-d7ab65c593f1-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"724e114a-1ace-4455-846f-d7ab65c593f1\") " pod="openstack/glance-default-external-api-0" Jan 26 17:09:29 crc kubenswrapper[4754]: I0126 17:09:29.035830 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/724e114a-1ace-4455-846f-d7ab65c593f1-config-data\") pod \"glance-default-external-api-0\" (UID: \"724e114a-1ace-4455-846f-d7ab65c593f1\") " pod="openstack/glance-default-external-api-0" Jan 26 17:09:29 crc kubenswrapper[4754]: I0126 17:09:29.035851 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/724e114a-1ace-4455-846f-d7ab65c593f1-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"724e114a-1ace-4455-846f-d7ab65c593f1\") " pod="openstack/glance-default-external-api-0" Jan 26 17:09:29 crc kubenswrapper[4754]: I0126 17:09:29.035870 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/724e114a-1ace-4455-846f-d7ab65c593f1-scripts\") pod \"glance-default-external-api-0\" (UID: \"724e114a-1ace-4455-846f-d7ab65c593f1\") " pod="openstack/glance-default-external-api-0" Jan 26 17:09:29 crc kubenswrapper[4754]: I0126 17:09:29.037377 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/724e114a-1ace-4455-846f-d7ab65c593f1-logs\") pod \"glance-default-external-api-0\" (UID: \"724e114a-1ace-4455-846f-d7ab65c593f1\") " pod="openstack/glance-default-external-api-0" Jan 26 17:09:29 crc kubenswrapper[4754]: I0126 17:09:29.037623 4754 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"724e114a-1ace-4455-846f-d7ab65c593f1\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/glance-default-external-api-0" Jan 26 17:09:29 crc kubenswrapper[4754]: I0126 17:09:29.041124 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/724e114a-1ace-4455-846f-d7ab65c593f1-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"724e114a-1ace-4455-846f-d7ab65c593f1\") " pod="openstack/glance-default-external-api-0" Jan 26 17:09:29 crc kubenswrapper[4754]: I0126 17:09:29.042290 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/724e114a-1ace-4455-846f-d7ab65c593f1-config-data\") pod \"glance-default-external-api-0\" (UID: \"724e114a-1ace-4455-846f-d7ab65c593f1\") " pod="openstack/glance-default-external-api-0" Jan 26 17:09:29 crc kubenswrapper[4754]: I0126 17:09:29.042496 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/724e114a-1ace-4455-846f-d7ab65c593f1-scripts\") pod \"glance-default-external-api-0\" (UID: \"724e114a-1ace-4455-846f-d7ab65c593f1\") " pod="openstack/glance-default-external-api-0" Jan 26 17:09:29 crc kubenswrapper[4754]: I0126 17:09:29.043021 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/724e114a-1ace-4455-846f-d7ab65c593f1-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"724e114a-1ace-4455-846f-d7ab65c593f1\") " pod="openstack/glance-default-external-api-0" Jan 26 17:09:29 crc kubenswrapper[4754]: I0126 17:09:29.045042 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/724e114a-1ace-4455-846f-d7ab65c593f1-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"724e114a-1ace-4455-846f-d7ab65c593f1\") " pod="openstack/glance-default-external-api-0" Jan 26 17:09:29 crc kubenswrapper[4754]: I0126 17:09:29.061111 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mxbmg\" (UniqueName: \"kubernetes.io/projected/724e114a-1ace-4455-846f-d7ab65c593f1-kube-api-access-mxbmg\") pod \"glance-default-external-api-0\" (UID: \"724e114a-1ace-4455-846f-d7ab65c593f1\") " pod="openstack/glance-default-external-api-0" Jan 26 17:09:29 crc kubenswrapper[4754]: I0126 17:09:29.081153 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"724e114a-1ace-4455-846f-d7ab65c593f1\") " pod="openstack/glance-default-external-api-0" Jan 26 17:09:29 crc kubenswrapper[4754]: I0126 17:09:29.284567 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Jan 26 17:09:29 crc kubenswrapper[4754]: I0126 17:09:29.386981 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Jan 26 17:09:29 crc kubenswrapper[4754]: I0126 17:09:29.447326 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"529bf77e-614d-4354-94bd-e6383f353920","Type":"ContainerStarted","Data":"6f46e829e7a0d956fb3a009296368b2e1c4a2fb4dba8591c362b9bdef6195e30"} Jan 26 17:09:29 crc kubenswrapper[4754]: I0126 17:09:29.783602 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4a8fafa1-3d82-4a16-85e1-fd3116bc998a" path="/var/lib/kubelet/pods/4a8fafa1-3d82-4a16-85e1-fd3116bc998a/volumes" Jan 26 17:09:29 crc kubenswrapper[4754]: I0126 17:09:29.784712 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="54272c08-92b7-4512-8ee3-2a782926cda1" path="/var/lib/kubelet/pods/54272c08-92b7-4512-8ee3-2a782926cda1/volumes" Jan 26 17:09:29 crc kubenswrapper[4754]: I0126 17:09:29.871591 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Jan 26 17:09:30 crc kubenswrapper[4754]: I0126 17:09:30.459401 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"724e114a-1ace-4455-846f-d7ab65c593f1","Type":"ContainerStarted","Data":"ac8d3af1e4227279f86966fe4360df11750bab44e1c6da721c9b30b910c1db2e"} Jan 26 17:09:30 crc kubenswrapper[4754]: I0126 17:09:30.462510 4754 generic.go:334] "Generic (PLEG): container finished" podID="4c65192f-a05a-45e2-9d56-33ddeb265374" containerID="0bd046d181549657eae96bff37d02dbcc14075bc5395685c218279cc1485d7bd" exitCode=0 Jan 26 17:09:30 crc kubenswrapper[4754]: I0126 17:09:30.462548 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4c65192f-a05a-45e2-9d56-33ddeb265374","Type":"ContainerDied","Data":"0bd046d181549657eae96bff37d02dbcc14075bc5395685c218279cc1485d7bd"} Jan 26 17:09:30 crc kubenswrapper[4754]: I0126 17:09:30.685039 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Jan 26 17:09:30 crc kubenswrapper[4754]: I0126 17:09:30.843834 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-795f4db4bc-h7mcp" Jan 26 17:09:30 crc kubenswrapper[4754]: I0126 17:09:30.902123 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Jan 26 17:09:30 crc kubenswrapper[4754]: I0126 17:09:30.917774 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-586bdc5f9-lqn4x"] Jan 26 17:09:30 crc kubenswrapper[4754]: I0126 17:09:30.918049 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-586bdc5f9-lqn4x" podUID="057f7e48-aac2-407d-9fd4-555b5f8c6dad" containerName="dnsmasq-dns" containerID="cri-o://2e5ddf125f0ef1060a21de205d16388b079b15d704493d09a9b6fa0a3bc82c2d" gracePeriod=10 Jan 26 17:09:31 crc kubenswrapper[4754]: I0126 17:09:31.476866 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"724e114a-1ace-4455-846f-d7ab65c593f1","Type":"ContainerStarted","Data":"7b8205884cbb7b40b61ffd2cf9c911868d89a80509508559360938b6f6d8e870"} Jan 26 17:09:31 crc kubenswrapper[4754]: I0126 17:09:31.478871 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"529bf77e-614d-4354-94bd-e6383f353920","Type":"ContainerStarted","Data":"d8a1530caa8f60a651566b2691a5e31434ab0d0afadcee954feb9ba937f1d51d"} Jan 26 17:09:31 crc kubenswrapper[4754]: I0126 17:09:31.534378 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Jan 26 17:09:31 crc kubenswrapper[4754]: I0126 17:09:31.757436 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 26 17:09:31 crc kubenswrapper[4754]: I0126 17:09:31.916707 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4c65192f-a05a-45e2-9d56-33ddeb265374-run-httpd\") pod \"4c65192f-a05a-45e2-9d56-33ddeb265374\" (UID: \"4c65192f-a05a-45e2-9d56-33ddeb265374\") " Jan 26 17:09:31 crc kubenswrapper[4754]: I0126 17:09:31.917016 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4vqll\" (UniqueName: \"kubernetes.io/projected/4c65192f-a05a-45e2-9d56-33ddeb265374-kube-api-access-4vqll\") pod \"4c65192f-a05a-45e2-9d56-33ddeb265374\" (UID: \"4c65192f-a05a-45e2-9d56-33ddeb265374\") " Jan 26 17:09:31 crc kubenswrapper[4754]: I0126 17:09:31.917034 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4c65192f-a05a-45e2-9d56-33ddeb265374-sg-core-conf-yaml\") pod \"4c65192f-a05a-45e2-9d56-33ddeb265374\" (UID: \"4c65192f-a05a-45e2-9d56-33ddeb265374\") " Jan 26 17:09:31 crc kubenswrapper[4754]: I0126 17:09:31.917165 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4c65192f-a05a-45e2-9d56-33ddeb265374-scripts\") pod \"4c65192f-a05a-45e2-9d56-33ddeb265374\" (UID: \"4c65192f-a05a-45e2-9d56-33ddeb265374\") " Jan 26 17:09:31 crc kubenswrapper[4754]: I0126 17:09:31.917196 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4c65192f-a05a-45e2-9d56-33ddeb265374-log-httpd\") pod \"4c65192f-a05a-45e2-9d56-33ddeb265374\" (UID: \"4c65192f-a05a-45e2-9d56-33ddeb265374\") " Jan 26 17:09:31 crc kubenswrapper[4754]: I0126 17:09:31.917313 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c65192f-a05a-45e2-9d56-33ddeb265374-config-data\") pod \"4c65192f-a05a-45e2-9d56-33ddeb265374\" (UID: \"4c65192f-a05a-45e2-9d56-33ddeb265374\") " Jan 26 17:09:31 crc kubenswrapper[4754]: I0126 17:09:31.917357 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c65192f-a05a-45e2-9d56-33ddeb265374-combined-ca-bundle\") pod \"4c65192f-a05a-45e2-9d56-33ddeb265374\" (UID: \"4c65192f-a05a-45e2-9d56-33ddeb265374\") " Jan 26 17:09:31 crc kubenswrapper[4754]: I0126 17:09:31.918575 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4c65192f-a05a-45e2-9d56-33ddeb265374-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "4c65192f-a05a-45e2-9d56-33ddeb265374" (UID: "4c65192f-a05a-45e2-9d56-33ddeb265374"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:09:31 crc kubenswrapper[4754]: I0126 17:09:31.918956 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4c65192f-a05a-45e2-9d56-33ddeb265374-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "4c65192f-a05a-45e2-9d56-33ddeb265374" (UID: "4c65192f-a05a-45e2-9d56-33ddeb265374"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:09:31 crc kubenswrapper[4754]: I0126 17:09:31.933643 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4c65192f-a05a-45e2-9d56-33ddeb265374-kube-api-access-4vqll" (OuterVolumeSpecName: "kube-api-access-4vqll") pod "4c65192f-a05a-45e2-9d56-33ddeb265374" (UID: "4c65192f-a05a-45e2-9d56-33ddeb265374"). InnerVolumeSpecName "kube-api-access-4vqll". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:09:31 crc kubenswrapper[4754]: I0126 17:09:31.942055 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c65192f-a05a-45e2-9d56-33ddeb265374-scripts" (OuterVolumeSpecName: "scripts") pod "4c65192f-a05a-45e2-9d56-33ddeb265374" (UID: "4c65192f-a05a-45e2-9d56-33ddeb265374"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:09:31 crc kubenswrapper[4754]: I0126 17:09:31.943096 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Jan 26 17:09:31 crc kubenswrapper[4754]: I0126 17:09:31.943346 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="428912eb-4aed-4926-8b00-f716a43a4399" containerName="glance-log" containerID="cri-o://17a24ca2bd71c5f9883d7d6763e457a9ad1791b53a390081e373f23d8c1cc0af" gracePeriod=30 Jan 26 17:09:31 crc kubenswrapper[4754]: I0126 17:09:31.943803 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="428912eb-4aed-4926-8b00-f716a43a4399" containerName="glance-httpd" containerID="cri-o://34521a84db431d82e60898db22e047282e63e4072af221baedb12427e8c15b10" gracePeriod=30 Jan 26 17:09:31 crc kubenswrapper[4754]: I0126 17:09:31.994802 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c65192f-a05a-45e2-9d56-33ddeb265374-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "4c65192f-a05a-45e2-9d56-33ddeb265374" (UID: "4c65192f-a05a-45e2-9d56-33ddeb265374"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:09:32 crc kubenswrapper[4754]: I0126 17:09:32.020100 4754 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4c65192f-a05a-45e2-9d56-33ddeb265374-run-httpd\") on node \"crc\" DevicePath \"\"" Jan 26 17:09:32 crc kubenswrapper[4754]: I0126 17:09:32.020152 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4vqll\" (UniqueName: \"kubernetes.io/projected/4c65192f-a05a-45e2-9d56-33ddeb265374-kube-api-access-4vqll\") on node \"crc\" DevicePath \"\"" Jan 26 17:09:32 crc kubenswrapper[4754]: I0126 17:09:32.020167 4754 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4c65192f-a05a-45e2-9d56-33ddeb265374-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Jan 26 17:09:32 crc kubenswrapper[4754]: I0126 17:09:32.020178 4754 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4c65192f-a05a-45e2-9d56-33ddeb265374-scripts\") on node \"crc\" DevicePath \"\"" Jan 26 17:09:32 crc kubenswrapper[4754]: I0126 17:09:32.020188 4754 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4c65192f-a05a-45e2-9d56-33ddeb265374-log-httpd\") on node \"crc\" DevicePath \"\"" Jan 26 17:09:32 crc kubenswrapper[4754]: I0126 17:09:32.021909 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c65192f-a05a-45e2-9d56-33ddeb265374-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4c65192f-a05a-45e2-9d56-33ddeb265374" (UID: "4c65192f-a05a-45e2-9d56-33ddeb265374"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:09:32 crc kubenswrapper[4754]: I0126 17:09:32.112444 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c65192f-a05a-45e2-9d56-33ddeb265374-config-data" (OuterVolumeSpecName: "config-data") pod "4c65192f-a05a-45e2-9d56-33ddeb265374" (UID: "4c65192f-a05a-45e2-9d56-33ddeb265374"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:09:32 crc kubenswrapper[4754]: I0126 17:09:32.122788 4754 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c65192f-a05a-45e2-9d56-33ddeb265374-config-data\") on node \"crc\" DevicePath \"\"" Jan 26 17:09:32 crc kubenswrapper[4754]: I0126 17:09:32.122850 4754 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c65192f-a05a-45e2-9d56-33ddeb265374-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 26 17:09:32 crc kubenswrapper[4754]: I0126 17:09:32.282207 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-586bdc5f9-lqn4x" Jan 26 17:09:32 crc kubenswrapper[4754]: I0126 17:09:32.439188 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/057f7e48-aac2-407d-9fd4-555b5f8c6dad-ovsdbserver-sb\") pod \"057f7e48-aac2-407d-9fd4-555b5f8c6dad\" (UID: \"057f7e48-aac2-407d-9fd4-555b5f8c6dad\") " Jan 26 17:09:32 crc kubenswrapper[4754]: I0126 17:09:32.439335 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/057f7e48-aac2-407d-9fd4-555b5f8c6dad-dns-swift-storage-0\") pod \"057f7e48-aac2-407d-9fd4-555b5f8c6dad\" (UID: \"057f7e48-aac2-407d-9fd4-555b5f8c6dad\") " Jan 26 17:09:32 crc kubenswrapper[4754]: I0126 17:09:32.439383 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/057f7e48-aac2-407d-9fd4-555b5f8c6dad-dns-svc\") pod \"057f7e48-aac2-407d-9fd4-555b5f8c6dad\" (UID: \"057f7e48-aac2-407d-9fd4-555b5f8c6dad\") " Jan 26 17:09:32 crc kubenswrapper[4754]: I0126 17:09:32.439417 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/057f7e48-aac2-407d-9fd4-555b5f8c6dad-config\") pod \"057f7e48-aac2-407d-9fd4-555b5f8c6dad\" (UID: \"057f7e48-aac2-407d-9fd4-555b5f8c6dad\") " Jan 26 17:09:32 crc kubenswrapper[4754]: I0126 17:09:32.439435 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jh76b\" (UniqueName: \"kubernetes.io/projected/057f7e48-aac2-407d-9fd4-555b5f8c6dad-kube-api-access-jh76b\") pod \"057f7e48-aac2-407d-9fd4-555b5f8c6dad\" (UID: \"057f7e48-aac2-407d-9fd4-555b5f8c6dad\") " Jan 26 17:09:32 crc kubenswrapper[4754]: I0126 17:09:32.439476 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/057f7e48-aac2-407d-9fd4-555b5f8c6dad-ovsdbserver-nb\") pod \"057f7e48-aac2-407d-9fd4-555b5f8c6dad\" (UID: \"057f7e48-aac2-407d-9fd4-555b5f8c6dad\") " Jan 26 17:09:32 crc kubenswrapper[4754]: I0126 17:09:32.475830 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/057f7e48-aac2-407d-9fd4-555b5f8c6dad-kube-api-access-jh76b" (OuterVolumeSpecName: "kube-api-access-jh76b") pod "057f7e48-aac2-407d-9fd4-555b5f8c6dad" (UID: "057f7e48-aac2-407d-9fd4-555b5f8c6dad"). InnerVolumeSpecName "kube-api-access-jh76b". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:09:32 crc kubenswrapper[4754]: I0126 17:09:32.519352 4754 generic.go:334] "Generic (PLEG): container finished" podID="428912eb-4aed-4926-8b00-f716a43a4399" containerID="17a24ca2bd71c5f9883d7d6763e457a9ad1791b53a390081e373f23d8c1cc0af" exitCode=143 Jan 26 17:09:32 crc kubenswrapper[4754]: I0126 17:09:32.519445 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"428912eb-4aed-4926-8b00-f716a43a4399","Type":"ContainerDied","Data":"17a24ca2bd71c5f9883d7d6763e457a9ad1791b53a390081e373f23d8c1cc0af"} Jan 26 17:09:32 crc kubenswrapper[4754]: I0126 17:09:32.527055 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 26 17:09:32 crc kubenswrapper[4754]: I0126 17:09:32.527185 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4c65192f-a05a-45e2-9d56-33ddeb265374","Type":"ContainerDied","Data":"9283c476059ebf15ee0cfb420c7cb60dc5112b1663c3e2672d5ce9b1cc7cd9f9"} Jan 26 17:09:32 crc kubenswrapper[4754]: I0126 17:09:32.527267 4754 scope.go:117] "RemoveContainer" containerID="50992540292cc9c045cd8bb4c9e58c8d373ef8d0a1454a811b960e86c77eec5d" Jan 26 17:09:32 crc kubenswrapper[4754]: I0126 17:09:32.535420 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"529bf77e-614d-4354-94bd-e6383f353920","Type":"ContainerStarted","Data":"da937657b37eb7d77699778822245a49a989ff98fa811c1a53220eb0c49df6e1"} Jan 26 17:09:32 crc kubenswrapper[4754]: I0126 17:09:32.536358 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Jan 26 17:09:32 crc kubenswrapper[4754]: I0126 17:09:32.541123 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jh76b\" (UniqueName: \"kubernetes.io/projected/057f7e48-aac2-407d-9fd4-555b5f8c6dad-kube-api-access-jh76b\") on node \"crc\" DevicePath \"\"" Jan 26 17:09:32 crc kubenswrapper[4754]: I0126 17:09:32.543883 4754 generic.go:334] "Generic (PLEG): container finished" podID="057f7e48-aac2-407d-9fd4-555b5f8c6dad" containerID="2e5ddf125f0ef1060a21de205d16388b079b15d704493d09a9b6fa0a3bc82c2d" exitCode=0 Jan 26 17:09:32 crc kubenswrapper[4754]: I0126 17:09:32.543962 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-586bdc5f9-lqn4x" Jan 26 17:09:32 crc kubenswrapper[4754]: I0126 17:09:32.544036 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-586bdc5f9-lqn4x" event={"ID":"057f7e48-aac2-407d-9fd4-555b5f8c6dad","Type":"ContainerDied","Data":"2e5ddf125f0ef1060a21de205d16388b079b15d704493d09a9b6fa0a3bc82c2d"} Jan 26 17:09:32 crc kubenswrapper[4754]: I0126 17:09:32.544077 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-586bdc5f9-lqn4x" event={"ID":"057f7e48-aac2-407d-9fd4-555b5f8c6dad","Type":"ContainerDied","Data":"1d797d02e0a18f2e8fd1c14ae47ae138d6e9f382b84be194f3b2a811d63f5c2e"} Jan 26 17:09:32 crc kubenswrapper[4754]: I0126 17:09:32.544085 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="6f82d0f5-47a4-4584-9ae9-046882fa8f61" containerName="cinder-scheduler" containerID="cri-o://eb836410836ef755e634f1dddd915e0296b6426b6a75ab5c090ec925a000a6ba" gracePeriod=30 Jan 26 17:09:32 crc kubenswrapper[4754]: I0126 17:09:32.544247 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="6f82d0f5-47a4-4584-9ae9-046882fa8f61" containerName="probe" containerID="cri-o://66f684c6f13cd3e1ff23df6b19e556f74dc2dd34475bbf370be3391e82f922c0" gracePeriod=30 Jan 26 17:09:32 crc kubenswrapper[4754]: I0126 17:09:32.548830 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/057f7e48-aac2-407d-9fd4-555b5f8c6dad-config" (OuterVolumeSpecName: "config") pod "057f7e48-aac2-407d-9fd4-555b5f8c6dad" (UID: "057f7e48-aac2-407d-9fd4-555b5f8c6dad"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:09:32 crc kubenswrapper[4754]: I0126 17:09:32.566141 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/057f7e48-aac2-407d-9fd4-555b5f8c6dad-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "057f7e48-aac2-407d-9fd4-555b5f8c6dad" (UID: "057f7e48-aac2-407d-9fd4-555b5f8c6dad"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:09:32 crc kubenswrapper[4754]: I0126 17:09:32.575093 4754 scope.go:117] "RemoveContainer" containerID="3e9828e079b47eb4e66edc9599b5bdd713959f79d23385e7543f9bcb4e6ef25e" Jan 26 17:09:32 crc kubenswrapper[4754]: I0126 17:09:32.580259 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=4.580241674 podStartE2EDuration="4.580241674s" podCreationTimestamp="2026-01-26 17:09:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 17:09:32.5663556 +0000 UTC m=+1339.090536024" watchObservedRunningTime="2026-01-26 17:09:32.580241674 +0000 UTC m=+1339.104422108" Jan 26 17:09:32 crc kubenswrapper[4754]: I0126 17:09:32.607956 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/057f7e48-aac2-407d-9fd4-555b5f8c6dad-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "057f7e48-aac2-407d-9fd4-555b5f8c6dad" (UID: "057f7e48-aac2-407d-9fd4-555b5f8c6dad"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:09:32 crc kubenswrapper[4754]: I0126 17:09:32.612699 4754 scope.go:117] "RemoveContainer" containerID="0bd046d181549657eae96bff37d02dbcc14075bc5395685c218279cc1485d7bd" Jan 26 17:09:32 crc kubenswrapper[4754]: I0126 17:09:32.617410 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/057f7e48-aac2-407d-9fd4-555b5f8c6dad-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "057f7e48-aac2-407d-9fd4-555b5f8c6dad" (UID: "057f7e48-aac2-407d-9fd4-555b5f8c6dad"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:09:32 crc kubenswrapper[4754]: I0126 17:09:32.619108 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Jan 26 17:09:32 crc kubenswrapper[4754]: I0126 17:09:32.621442 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/057f7e48-aac2-407d-9fd4-555b5f8c6dad-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "057f7e48-aac2-407d-9fd4-555b5f8c6dad" (UID: "057f7e48-aac2-407d-9fd4-555b5f8c6dad"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:09:32 crc kubenswrapper[4754]: I0126 17:09:32.640517 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Jan 26 17:09:32 crc kubenswrapper[4754]: I0126 17:09:32.644040 4754 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/057f7e48-aac2-407d-9fd4-555b5f8c6dad-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Jan 26 17:09:32 crc kubenswrapper[4754]: I0126 17:09:32.644067 4754 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/057f7e48-aac2-407d-9fd4-555b5f8c6dad-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Jan 26 17:09:32 crc kubenswrapper[4754]: I0126 17:09:32.644076 4754 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/057f7e48-aac2-407d-9fd4-555b5f8c6dad-dns-svc\") on node \"crc\" DevicePath \"\"" Jan 26 17:09:32 crc kubenswrapper[4754]: I0126 17:09:32.644085 4754 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/057f7e48-aac2-407d-9fd4-555b5f8c6dad-config\") on node \"crc\" DevicePath \"\"" Jan 26 17:09:32 crc kubenswrapper[4754]: I0126 17:09:32.644093 4754 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/057f7e48-aac2-407d-9fd4-555b5f8c6dad-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Jan 26 17:09:32 crc kubenswrapper[4754]: I0126 17:09:32.658720 4754 scope.go:117] "RemoveContainer" containerID="374dc6158046ff5aeb4527bdd32fd9636f239a18d5ba35caa78a117156d7fb06" Jan 26 17:09:32 crc kubenswrapper[4754]: I0126 17:09:32.662700 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Jan 26 17:09:32 crc kubenswrapper[4754]: E0126 17:09:32.663194 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="057f7e48-aac2-407d-9fd4-555b5f8c6dad" containerName="dnsmasq-dns" Jan 26 17:09:32 crc kubenswrapper[4754]: I0126 17:09:32.663217 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="057f7e48-aac2-407d-9fd4-555b5f8c6dad" containerName="dnsmasq-dns" Jan 26 17:09:32 crc kubenswrapper[4754]: E0126 17:09:32.663317 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c65192f-a05a-45e2-9d56-33ddeb265374" containerName="ceilometer-notification-agent" Jan 26 17:09:32 crc kubenswrapper[4754]: I0126 17:09:32.663333 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c65192f-a05a-45e2-9d56-33ddeb265374" containerName="ceilometer-notification-agent" Jan 26 17:09:32 crc kubenswrapper[4754]: E0126 17:09:32.663348 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c65192f-a05a-45e2-9d56-33ddeb265374" containerName="sg-core" Jan 26 17:09:32 crc kubenswrapper[4754]: I0126 17:09:32.663357 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c65192f-a05a-45e2-9d56-33ddeb265374" containerName="sg-core" Jan 26 17:09:32 crc kubenswrapper[4754]: E0126 17:09:32.663369 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c65192f-a05a-45e2-9d56-33ddeb265374" containerName="proxy-httpd" Jan 26 17:09:32 crc kubenswrapper[4754]: I0126 17:09:32.663376 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c65192f-a05a-45e2-9d56-33ddeb265374" containerName="proxy-httpd" Jan 26 17:09:32 crc kubenswrapper[4754]: E0126 17:09:32.663393 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="057f7e48-aac2-407d-9fd4-555b5f8c6dad" containerName="init" Jan 26 17:09:32 crc kubenswrapper[4754]: I0126 17:09:32.663401 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="057f7e48-aac2-407d-9fd4-555b5f8c6dad" containerName="init" Jan 26 17:09:32 crc kubenswrapper[4754]: E0126 17:09:32.663416 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c65192f-a05a-45e2-9d56-33ddeb265374" containerName="ceilometer-central-agent" Jan 26 17:09:32 crc kubenswrapper[4754]: I0126 17:09:32.663427 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c65192f-a05a-45e2-9d56-33ddeb265374" containerName="ceilometer-central-agent" Jan 26 17:09:32 crc kubenswrapper[4754]: I0126 17:09:32.663638 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="4c65192f-a05a-45e2-9d56-33ddeb265374" containerName="ceilometer-notification-agent" Jan 26 17:09:32 crc kubenswrapper[4754]: I0126 17:09:32.663682 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="4c65192f-a05a-45e2-9d56-33ddeb265374" containerName="ceilometer-central-agent" Jan 26 17:09:32 crc kubenswrapper[4754]: I0126 17:09:32.663700 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="4c65192f-a05a-45e2-9d56-33ddeb265374" containerName="proxy-httpd" Jan 26 17:09:32 crc kubenswrapper[4754]: I0126 17:09:32.663711 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="057f7e48-aac2-407d-9fd4-555b5f8c6dad" containerName="dnsmasq-dns" Jan 26 17:09:32 crc kubenswrapper[4754]: I0126 17:09:32.663733 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="4c65192f-a05a-45e2-9d56-33ddeb265374" containerName="sg-core" Jan 26 17:09:32 crc kubenswrapper[4754]: I0126 17:09:32.665749 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 26 17:09:32 crc kubenswrapper[4754]: I0126 17:09:32.668024 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Jan 26 17:09:32 crc kubenswrapper[4754]: I0126 17:09:32.668358 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Jan 26 17:09:32 crc kubenswrapper[4754]: I0126 17:09:32.672008 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Jan 26 17:09:32 crc kubenswrapper[4754]: I0126 17:09:32.685943 4754 scope.go:117] "RemoveContainer" containerID="2e5ddf125f0ef1060a21de205d16388b079b15d704493d09a9b6fa0a3bc82c2d" Jan 26 17:09:32 crc kubenswrapper[4754]: I0126 17:09:32.708696 4754 scope.go:117] "RemoveContainer" containerID="cf88012f2cc61bc7a32fa7efc83cfc617efba3a5eaf348518664d834f71460ab" Jan 26 17:09:32 crc kubenswrapper[4754]: I0126 17:09:32.728952 4754 scope.go:117] "RemoveContainer" containerID="2e5ddf125f0ef1060a21de205d16388b079b15d704493d09a9b6fa0a3bc82c2d" Jan 26 17:09:32 crc kubenswrapper[4754]: E0126 17:09:32.729474 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2e5ddf125f0ef1060a21de205d16388b079b15d704493d09a9b6fa0a3bc82c2d\": container with ID starting with 2e5ddf125f0ef1060a21de205d16388b079b15d704493d09a9b6fa0a3bc82c2d not found: ID does not exist" containerID="2e5ddf125f0ef1060a21de205d16388b079b15d704493d09a9b6fa0a3bc82c2d" Jan 26 17:09:32 crc kubenswrapper[4754]: I0126 17:09:32.729503 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2e5ddf125f0ef1060a21de205d16388b079b15d704493d09a9b6fa0a3bc82c2d"} err="failed to get container status \"2e5ddf125f0ef1060a21de205d16388b079b15d704493d09a9b6fa0a3bc82c2d\": rpc error: code = NotFound desc = could not find container \"2e5ddf125f0ef1060a21de205d16388b079b15d704493d09a9b6fa0a3bc82c2d\": container with ID starting with 2e5ddf125f0ef1060a21de205d16388b079b15d704493d09a9b6fa0a3bc82c2d not found: ID does not exist" Jan 26 17:09:32 crc kubenswrapper[4754]: I0126 17:09:32.729523 4754 scope.go:117] "RemoveContainer" containerID="cf88012f2cc61bc7a32fa7efc83cfc617efba3a5eaf348518664d834f71460ab" Jan 26 17:09:32 crc kubenswrapper[4754]: E0126 17:09:32.729797 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cf88012f2cc61bc7a32fa7efc83cfc617efba3a5eaf348518664d834f71460ab\": container with ID starting with cf88012f2cc61bc7a32fa7efc83cfc617efba3a5eaf348518664d834f71460ab not found: ID does not exist" containerID="cf88012f2cc61bc7a32fa7efc83cfc617efba3a5eaf348518664d834f71460ab" Jan 26 17:09:32 crc kubenswrapper[4754]: I0126 17:09:32.729843 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cf88012f2cc61bc7a32fa7efc83cfc617efba3a5eaf348518664d834f71460ab"} err="failed to get container status \"cf88012f2cc61bc7a32fa7efc83cfc617efba3a5eaf348518664d834f71460ab\": rpc error: code = NotFound desc = could not find container \"cf88012f2cc61bc7a32fa7efc83cfc617efba3a5eaf348518664d834f71460ab\": container with ID starting with cf88012f2cc61bc7a32fa7efc83cfc617efba3a5eaf348518664d834f71460ab not found: ID does not exist" Jan 26 17:09:32 crc kubenswrapper[4754]: I0126 17:09:32.847373 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b63306d0-0c02-4691-81c7-60846f3c3fdc-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b63306d0-0c02-4691-81c7-60846f3c3fdc\") " pod="openstack/ceilometer-0" Jan 26 17:09:32 crc kubenswrapper[4754]: I0126 17:09:32.847426 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b63306d0-0c02-4691-81c7-60846f3c3fdc-config-data\") pod \"ceilometer-0\" (UID: \"b63306d0-0c02-4691-81c7-60846f3c3fdc\") " pod="openstack/ceilometer-0" Jan 26 17:09:32 crc kubenswrapper[4754]: I0126 17:09:32.847461 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b63306d0-0c02-4691-81c7-60846f3c3fdc-log-httpd\") pod \"ceilometer-0\" (UID: \"b63306d0-0c02-4691-81c7-60846f3c3fdc\") " pod="openstack/ceilometer-0" Jan 26 17:09:32 crc kubenswrapper[4754]: I0126 17:09:32.847533 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jpqgs\" (UniqueName: \"kubernetes.io/projected/b63306d0-0c02-4691-81c7-60846f3c3fdc-kube-api-access-jpqgs\") pod \"ceilometer-0\" (UID: \"b63306d0-0c02-4691-81c7-60846f3c3fdc\") " pod="openstack/ceilometer-0" Jan 26 17:09:32 crc kubenswrapper[4754]: I0126 17:09:32.847610 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b63306d0-0c02-4691-81c7-60846f3c3fdc-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b63306d0-0c02-4691-81c7-60846f3c3fdc\") " pod="openstack/ceilometer-0" Jan 26 17:09:32 crc kubenswrapper[4754]: I0126 17:09:32.847692 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b63306d0-0c02-4691-81c7-60846f3c3fdc-scripts\") pod \"ceilometer-0\" (UID: \"b63306d0-0c02-4691-81c7-60846f3c3fdc\") " pod="openstack/ceilometer-0" Jan 26 17:09:32 crc kubenswrapper[4754]: I0126 17:09:32.847763 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b63306d0-0c02-4691-81c7-60846f3c3fdc-run-httpd\") pod \"ceilometer-0\" (UID: \"b63306d0-0c02-4691-81c7-60846f3c3fdc\") " pod="openstack/ceilometer-0" Jan 26 17:09:32 crc kubenswrapper[4754]: I0126 17:09:32.876960 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-586bdc5f9-lqn4x"] Jan 26 17:09:32 crc kubenswrapper[4754]: I0126 17:09:32.886820 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-586bdc5f9-lqn4x"] Jan 26 17:09:32 crc kubenswrapper[4754]: I0126 17:09:32.949406 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b63306d0-0c02-4691-81c7-60846f3c3fdc-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b63306d0-0c02-4691-81c7-60846f3c3fdc\") " pod="openstack/ceilometer-0" Jan 26 17:09:32 crc kubenswrapper[4754]: I0126 17:09:32.950227 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b63306d0-0c02-4691-81c7-60846f3c3fdc-config-data\") pod \"ceilometer-0\" (UID: \"b63306d0-0c02-4691-81c7-60846f3c3fdc\") " pod="openstack/ceilometer-0" Jan 26 17:09:32 crc kubenswrapper[4754]: I0126 17:09:32.950288 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b63306d0-0c02-4691-81c7-60846f3c3fdc-log-httpd\") pod \"ceilometer-0\" (UID: \"b63306d0-0c02-4691-81c7-60846f3c3fdc\") " pod="openstack/ceilometer-0" Jan 26 17:09:32 crc kubenswrapper[4754]: I0126 17:09:32.950394 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jpqgs\" (UniqueName: \"kubernetes.io/projected/b63306d0-0c02-4691-81c7-60846f3c3fdc-kube-api-access-jpqgs\") pod \"ceilometer-0\" (UID: \"b63306d0-0c02-4691-81c7-60846f3c3fdc\") " pod="openstack/ceilometer-0" Jan 26 17:09:32 crc kubenswrapper[4754]: I0126 17:09:32.950515 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b63306d0-0c02-4691-81c7-60846f3c3fdc-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b63306d0-0c02-4691-81c7-60846f3c3fdc\") " pod="openstack/ceilometer-0" Jan 26 17:09:32 crc kubenswrapper[4754]: I0126 17:09:32.950618 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b63306d0-0c02-4691-81c7-60846f3c3fdc-scripts\") pod \"ceilometer-0\" (UID: \"b63306d0-0c02-4691-81c7-60846f3c3fdc\") " pod="openstack/ceilometer-0" Jan 26 17:09:32 crc kubenswrapper[4754]: I0126 17:09:32.950834 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b63306d0-0c02-4691-81c7-60846f3c3fdc-log-httpd\") pod \"ceilometer-0\" (UID: \"b63306d0-0c02-4691-81c7-60846f3c3fdc\") " pod="openstack/ceilometer-0" Jan 26 17:09:32 crc kubenswrapper[4754]: I0126 17:09:32.950840 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b63306d0-0c02-4691-81c7-60846f3c3fdc-run-httpd\") pod \"ceilometer-0\" (UID: \"b63306d0-0c02-4691-81c7-60846f3c3fdc\") " pod="openstack/ceilometer-0" Jan 26 17:09:32 crc kubenswrapper[4754]: I0126 17:09:32.951353 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b63306d0-0c02-4691-81c7-60846f3c3fdc-run-httpd\") pod \"ceilometer-0\" (UID: \"b63306d0-0c02-4691-81c7-60846f3c3fdc\") " pod="openstack/ceilometer-0" Jan 26 17:09:32 crc kubenswrapper[4754]: I0126 17:09:32.954370 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b63306d0-0c02-4691-81c7-60846f3c3fdc-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b63306d0-0c02-4691-81c7-60846f3c3fdc\") " pod="openstack/ceilometer-0" Jan 26 17:09:32 crc kubenswrapper[4754]: I0126 17:09:32.954751 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b63306d0-0c02-4691-81c7-60846f3c3fdc-scripts\") pod \"ceilometer-0\" (UID: \"b63306d0-0c02-4691-81c7-60846f3c3fdc\") " pod="openstack/ceilometer-0" Jan 26 17:09:32 crc kubenswrapper[4754]: I0126 17:09:32.955694 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b63306d0-0c02-4691-81c7-60846f3c3fdc-config-data\") pod \"ceilometer-0\" (UID: \"b63306d0-0c02-4691-81c7-60846f3c3fdc\") " pod="openstack/ceilometer-0" Jan 26 17:09:32 crc kubenswrapper[4754]: I0126 17:09:32.957526 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b63306d0-0c02-4691-81c7-60846f3c3fdc-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b63306d0-0c02-4691-81c7-60846f3c3fdc\") " pod="openstack/ceilometer-0" Jan 26 17:09:32 crc kubenswrapper[4754]: I0126 17:09:32.969569 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jpqgs\" (UniqueName: \"kubernetes.io/projected/b63306d0-0c02-4691-81c7-60846f3c3fdc-kube-api-access-jpqgs\") pod \"ceilometer-0\" (UID: \"b63306d0-0c02-4691-81c7-60846f3c3fdc\") " pod="openstack/ceilometer-0" Jan 26 17:09:32 crc kubenswrapper[4754]: I0126 17:09:32.987464 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 26 17:09:33 crc kubenswrapper[4754]: I0126 17:09:33.454395 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Jan 26 17:09:33 crc kubenswrapper[4754]: I0126 17:09:33.557068 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b63306d0-0c02-4691-81c7-60846f3c3fdc","Type":"ContainerStarted","Data":"cbf2ae3def1abc7c686fb44f45c9e78d9557a69bc9f2ab121c6b1e8b279357ac"} Jan 26 17:09:33 crc kubenswrapper[4754]: I0126 17:09:33.559120 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"724e114a-1ace-4455-846f-d7ab65c593f1","Type":"ContainerStarted","Data":"e9fde76dcc9632388852f9ee52946d83993af81dfe234deb09a4181bb86ebfda"} Jan 26 17:09:33 crc kubenswrapper[4754]: I0126 17:09:33.562949 4754 generic.go:334] "Generic (PLEG): container finished" podID="6f82d0f5-47a4-4584-9ae9-046882fa8f61" containerID="66f684c6f13cd3e1ff23df6b19e556f74dc2dd34475bbf370be3391e82f922c0" exitCode=0 Jan 26 17:09:33 crc kubenswrapper[4754]: I0126 17:09:33.563006 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"6f82d0f5-47a4-4584-9ae9-046882fa8f61","Type":"ContainerDied","Data":"66f684c6f13cd3e1ff23df6b19e556f74dc2dd34475bbf370be3391e82f922c0"} Jan 26 17:09:33 crc kubenswrapper[4754]: I0126 17:09:33.586585 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=5.586560458 podStartE2EDuration="5.586560458s" podCreationTimestamp="2026-01-26 17:09:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 17:09:33.576651358 +0000 UTC m=+1340.100831792" watchObservedRunningTime="2026-01-26 17:09:33.586560458 +0000 UTC m=+1340.110740892" Jan 26 17:09:33 crc kubenswrapper[4754]: I0126 17:09:33.779287 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="057f7e48-aac2-407d-9fd4-555b5f8c6dad" path="/var/lib/kubelet/pods/057f7e48-aac2-407d-9fd4-555b5f8c6dad/volumes" Jan 26 17:09:33 crc kubenswrapper[4754]: I0126 17:09:33.779980 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4c65192f-a05a-45e2-9d56-33ddeb265374" path="/var/lib/kubelet/pods/4c65192f-a05a-45e2-9d56-33ddeb265374/volumes" Jan 26 17:09:34 crc kubenswrapper[4754]: I0126 17:09:34.573320 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-6959bf4485-bkv7n" Jan 26 17:09:34 crc kubenswrapper[4754]: I0126 17:09:34.573715 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-6959bf4485-bkv7n" Jan 26 17:09:35 crc kubenswrapper[4754]: I0126 17:09:35.111469 4754 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/glance-default-internal-api-0" podUID="428912eb-4aed-4926-8b00-f716a43a4399" containerName="glance-httpd" probeResult="failure" output="Get \"https://10.217.0.158:9292/healthcheck\": read tcp 10.217.0.2:50760->10.217.0.158:9292: read: connection reset by peer" Jan 26 17:09:35 crc kubenswrapper[4754]: I0126 17:09:35.111515 4754 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/glance-default-internal-api-0" podUID="428912eb-4aed-4926-8b00-f716a43a4399" containerName="glance-log" probeResult="failure" output="Get \"https://10.217.0.158:9292/healthcheck\": read tcp 10.217.0.2:50762->10.217.0.158:9292: read: connection reset by peer" Jan 26 17:09:35 crc kubenswrapper[4754]: I0126 17:09:35.578947 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b63306d0-0c02-4691-81c7-60846f3c3fdc","Type":"ContainerStarted","Data":"6dbb6881c0cdc0f402b9772f369d7222270b21d8425d63e6c42a80db5944dcde"} Jan 26 17:09:35 crc kubenswrapper[4754]: I0126 17:09:35.581456 4754 generic.go:334] "Generic (PLEG): container finished" podID="428912eb-4aed-4926-8b00-f716a43a4399" containerID="34521a84db431d82e60898db22e047282e63e4072af221baedb12427e8c15b10" exitCode=0 Jan 26 17:09:35 crc kubenswrapper[4754]: I0126 17:09:35.581561 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"428912eb-4aed-4926-8b00-f716a43a4399","Type":"ContainerDied","Data":"34521a84db431d82e60898db22e047282e63e4072af221baedb12427e8c15b10"} Jan 26 17:09:36 crc kubenswrapper[4754]: I0126 17:09:36.484027 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Jan 26 17:09:36 crc kubenswrapper[4754]: I0126 17:09:36.592398 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"428912eb-4aed-4926-8b00-f716a43a4399","Type":"ContainerDied","Data":"e4ef61878ff787d988699a25e243588a87ca9d9e98ae9807c4af1e5a053dfa12"} Jan 26 17:09:36 crc kubenswrapper[4754]: I0126 17:09:36.592456 4754 scope.go:117] "RemoveContainer" containerID="34521a84db431d82e60898db22e047282e63e4072af221baedb12427e8c15b10" Jan 26 17:09:36 crc kubenswrapper[4754]: I0126 17:09:36.592470 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Jan 26 17:09:36 crc kubenswrapper[4754]: I0126 17:09:36.615243 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"428912eb-4aed-4926-8b00-f716a43a4399\" (UID: \"428912eb-4aed-4926-8b00-f716a43a4399\") " Jan 26 17:09:36 crc kubenswrapper[4754]: I0126 17:09:36.615296 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/428912eb-4aed-4926-8b00-f716a43a4399-combined-ca-bundle\") pod \"428912eb-4aed-4926-8b00-f716a43a4399\" (UID: \"428912eb-4aed-4926-8b00-f716a43a4399\") " Jan 26 17:09:36 crc kubenswrapper[4754]: I0126 17:09:36.615369 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/428912eb-4aed-4926-8b00-f716a43a4399-httpd-run\") pod \"428912eb-4aed-4926-8b00-f716a43a4399\" (UID: \"428912eb-4aed-4926-8b00-f716a43a4399\") " Jan 26 17:09:36 crc kubenswrapper[4754]: I0126 17:09:36.615427 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/428912eb-4aed-4926-8b00-f716a43a4399-internal-tls-certs\") pod \"428912eb-4aed-4926-8b00-f716a43a4399\" (UID: \"428912eb-4aed-4926-8b00-f716a43a4399\") " Jan 26 17:09:36 crc kubenswrapper[4754]: I0126 17:09:36.615442 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/428912eb-4aed-4926-8b00-f716a43a4399-scripts\") pod \"428912eb-4aed-4926-8b00-f716a43a4399\" (UID: \"428912eb-4aed-4926-8b00-f716a43a4399\") " Jan 26 17:09:36 crc kubenswrapper[4754]: I0126 17:09:36.615480 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r5q6t\" (UniqueName: \"kubernetes.io/projected/428912eb-4aed-4926-8b00-f716a43a4399-kube-api-access-r5q6t\") pod \"428912eb-4aed-4926-8b00-f716a43a4399\" (UID: \"428912eb-4aed-4926-8b00-f716a43a4399\") " Jan 26 17:09:36 crc kubenswrapper[4754]: I0126 17:09:36.615498 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/428912eb-4aed-4926-8b00-f716a43a4399-logs\") pod \"428912eb-4aed-4926-8b00-f716a43a4399\" (UID: \"428912eb-4aed-4926-8b00-f716a43a4399\") " Jan 26 17:09:36 crc kubenswrapper[4754]: I0126 17:09:36.615524 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/428912eb-4aed-4926-8b00-f716a43a4399-config-data\") pod \"428912eb-4aed-4926-8b00-f716a43a4399\" (UID: \"428912eb-4aed-4926-8b00-f716a43a4399\") " Jan 26 17:09:36 crc kubenswrapper[4754]: I0126 17:09:36.615803 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/428912eb-4aed-4926-8b00-f716a43a4399-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "428912eb-4aed-4926-8b00-f716a43a4399" (UID: "428912eb-4aed-4926-8b00-f716a43a4399"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:09:36 crc kubenswrapper[4754]: I0126 17:09:36.615903 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/428912eb-4aed-4926-8b00-f716a43a4399-logs" (OuterVolumeSpecName: "logs") pod "428912eb-4aed-4926-8b00-f716a43a4399" (UID: "428912eb-4aed-4926-8b00-f716a43a4399"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:09:36 crc kubenswrapper[4754]: I0126 17:09:36.623355 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "glance") pod "428912eb-4aed-4926-8b00-f716a43a4399" (UID: "428912eb-4aed-4926-8b00-f716a43a4399"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Jan 26 17:09:36 crc kubenswrapper[4754]: I0126 17:09:36.623349 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/428912eb-4aed-4926-8b00-f716a43a4399-scripts" (OuterVolumeSpecName: "scripts") pod "428912eb-4aed-4926-8b00-f716a43a4399" (UID: "428912eb-4aed-4926-8b00-f716a43a4399"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:09:36 crc kubenswrapper[4754]: I0126 17:09:36.623466 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/428912eb-4aed-4926-8b00-f716a43a4399-kube-api-access-r5q6t" (OuterVolumeSpecName: "kube-api-access-r5q6t") pod "428912eb-4aed-4926-8b00-f716a43a4399" (UID: "428912eb-4aed-4926-8b00-f716a43a4399"). InnerVolumeSpecName "kube-api-access-r5q6t". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:09:36 crc kubenswrapper[4754]: I0126 17:09:36.623586 4754 scope.go:117] "RemoveContainer" containerID="17a24ca2bd71c5f9883d7d6763e457a9ad1791b53a390081e373f23d8c1cc0af" Jan 26 17:09:36 crc kubenswrapper[4754]: I0126 17:09:36.647771 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/428912eb-4aed-4926-8b00-f716a43a4399-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "428912eb-4aed-4926-8b00-f716a43a4399" (UID: "428912eb-4aed-4926-8b00-f716a43a4399"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:09:36 crc kubenswrapper[4754]: I0126 17:09:36.670030 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/428912eb-4aed-4926-8b00-f716a43a4399-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "428912eb-4aed-4926-8b00-f716a43a4399" (UID: "428912eb-4aed-4926-8b00-f716a43a4399"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:09:36 crc kubenswrapper[4754]: I0126 17:09:36.678862 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/428912eb-4aed-4926-8b00-f716a43a4399-config-data" (OuterVolumeSpecName: "config-data") pod "428912eb-4aed-4926-8b00-f716a43a4399" (UID: "428912eb-4aed-4926-8b00-f716a43a4399"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:09:36 crc kubenswrapper[4754]: I0126 17:09:36.717846 4754 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Jan 26 17:09:36 crc kubenswrapper[4754]: I0126 17:09:36.717877 4754 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/428912eb-4aed-4926-8b00-f716a43a4399-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 26 17:09:36 crc kubenswrapper[4754]: I0126 17:09:36.717888 4754 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/428912eb-4aed-4926-8b00-f716a43a4399-httpd-run\") on node \"crc\" DevicePath \"\"" Jan 26 17:09:36 crc kubenswrapper[4754]: I0126 17:09:36.717897 4754 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/428912eb-4aed-4926-8b00-f716a43a4399-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Jan 26 17:09:36 crc kubenswrapper[4754]: I0126 17:09:36.717905 4754 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/428912eb-4aed-4926-8b00-f716a43a4399-scripts\") on node \"crc\" DevicePath \"\"" Jan 26 17:09:36 crc kubenswrapper[4754]: I0126 17:09:36.717914 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r5q6t\" (UniqueName: \"kubernetes.io/projected/428912eb-4aed-4926-8b00-f716a43a4399-kube-api-access-r5q6t\") on node \"crc\" DevicePath \"\"" Jan 26 17:09:36 crc kubenswrapper[4754]: I0126 17:09:36.717925 4754 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/428912eb-4aed-4926-8b00-f716a43a4399-logs\") on node \"crc\" DevicePath \"\"" Jan 26 17:09:36 crc kubenswrapper[4754]: I0126 17:09:36.717932 4754 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/428912eb-4aed-4926-8b00-f716a43a4399-config-data\") on node \"crc\" DevicePath \"\"" Jan 26 17:09:36 crc kubenswrapper[4754]: I0126 17:09:36.739527 4754 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Jan 26 17:09:36 crc kubenswrapper[4754]: I0126 17:09:36.819551 4754 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Jan 26 17:09:36 crc kubenswrapper[4754]: I0126 17:09:36.928358 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Jan 26 17:09:36 crc kubenswrapper[4754]: I0126 17:09:36.935042 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Jan 26 17:09:36 crc kubenswrapper[4754]: I0126 17:09:36.961144 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Jan 26 17:09:36 crc kubenswrapper[4754]: E0126 17:09:36.961539 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="428912eb-4aed-4926-8b00-f716a43a4399" containerName="glance-httpd" Jan 26 17:09:36 crc kubenswrapper[4754]: I0126 17:09:36.961563 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="428912eb-4aed-4926-8b00-f716a43a4399" containerName="glance-httpd" Jan 26 17:09:36 crc kubenswrapper[4754]: E0126 17:09:36.961592 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="428912eb-4aed-4926-8b00-f716a43a4399" containerName="glance-log" Jan 26 17:09:36 crc kubenswrapper[4754]: I0126 17:09:36.961600 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="428912eb-4aed-4926-8b00-f716a43a4399" containerName="glance-log" Jan 26 17:09:36 crc kubenswrapper[4754]: I0126 17:09:36.961781 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="428912eb-4aed-4926-8b00-f716a43a4399" containerName="glance-httpd" Jan 26 17:09:36 crc kubenswrapper[4754]: I0126 17:09:36.961809 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="428912eb-4aed-4926-8b00-f716a43a4399" containerName="glance-log" Jan 26 17:09:36 crc kubenswrapper[4754]: I0126 17:09:36.962719 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Jan 26 17:09:36 crc kubenswrapper[4754]: I0126 17:09:36.964649 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Jan 26 17:09:36 crc kubenswrapper[4754]: I0126 17:09:36.967872 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Jan 26 17:09:36 crc kubenswrapper[4754]: I0126 17:09:36.976979 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Jan 26 17:09:37 crc kubenswrapper[4754]: I0126 17:09:37.023839 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c0decb48-8392-4a9a-a253-82b24b4f07ef-logs\") pod \"glance-default-internal-api-0\" (UID: \"c0decb48-8392-4a9a-a253-82b24b4f07ef\") " pod="openstack/glance-default-internal-api-0" Jan 26 17:09:37 crc kubenswrapper[4754]: I0126 17:09:37.024129 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c0decb48-8392-4a9a-a253-82b24b4f07ef-config-data\") pod \"glance-default-internal-api-0\" (UID: \"c0decb48-8392-4a9a-a253-82b24b4f07ef\") " pod="openstack/glance-default-internal-api-0" Jan 26 17:09:37 crc kubenswrapper[4754]: I0126 17:09:37.024226 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c0decb48-8392-4a9a-a253-82b24b4f07ef-scripts\") pod \"glance-default-internal-api-0\" (UID: \"c0decb48-8392-4a9a-a253-82b24b4f07ef\") " pod="openstack/glance-default-internal-api-0" Jan 26 17:09:37 crc kubenswrapper[4754]: I0126 17:09:37.024304 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c0decb48-8392-4a9a-a253-82b24b4f07ef-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"c0decb48-8392-4a9a-a253-82b24b4f07ef\") " pod="openstack/glance-default-internal-api-0" Jan 26 17:09:37 crc kubenswrapper[4754]: I0126 17:09:37.024374 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c0decb48-8392-4a9a-a253-82b24b4f07ef-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"c0decb48-8392-4a9a-a253-82b24b4f07ef\") " pod="openstack/glance-default-internal-api-0" Jan 26 17:09:37 crc kubenswrapper[4754]: I0126 17:09:37.024447 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vxv68\" (UniqueName: \"kubernetes.io/projected/c0decb48-8392-4a9a-a253-82b24b4f07ef-kube-api-access-vxv68\") pod \"glance-default-internal-api-0\" (UID: \"c0decb48-8392-4a9a-a253-82b24b4f07ef\") " pod="openstack/glance-default-internal-api-0" Jan 26 17:09:37 crc kubenswrapper[4754]: I0126 17:09:37.024547 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c0decb48-8392-4a9a-a253-82b24b4f07ef-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"c0decb48-8392-4a9a-a253-82b24b4f07ef\") " pod="openstack/glance-default-internal-api-0" Jan 26 17:09:37 crc kubenswrapper[4754]: I0126 17:09:37.024618 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"c0decb48-8392-4a9a-a253-82b24b4f07ef\") " pod="openstack/glance-default-internal-api-0" Jan 26 17:09:37 crc kubenswrapper[4754]: I0126 17:09:37.126622 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c0decb48-8392-4a9a-a253-82b24b4f07ef-logs\") pod \"glance-default-internal-api-0\" (UID: \"c0decb48-8392-4a9a-a253-82b24b4f07ef\") " pod="openstack/glance-default-internal-api-0" Jan 26 17:09:37 crc kubenswrapper[4754]: I0126 17:09:37.126884 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c0decb48-8392-4a9a-a253-82b24b4f07ef-config-data\") pod \"glance-default-internal-api-0\" (UID: \"c0decb48-8392-4a9a-a253-82b24b4f07ef\") " pod="openstack/glance-default-internal-api-0" Jan 26 17:09:37 crc kubenswrapper[4754]: I0126 17:09:37.126978 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c0decb48-8392-4a9a-a253-82b24b4f07ef-scripts\") pod \"glance-default-internal-api-0\" (UID: \"c0decb48-8392-4a9a-a253-82b24b4f07ef\") " pod="openstack/glance-default-internal-api-0" Jan 26 17:09:37 crc kubenswrapper[4754]: I0126 17:09:37.127097 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c0decb48-8392-4a9a-a253-82b24b4f07ef-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"c0decb48-8392-4a9a-a253-82b24b4f07ef\") " pod="openstack/glance-default-internal-api-0" Jan 26 17:09:37 crc kubenswrapper[4754]: I0126 17:09:37.127198 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c0decb48-8392-4a9a-a253-82b24b4f07ef-logs\") pod \"glance-default-internal-api-0\" (UID: \"c0decb48-8392-4a9a-a253-82b24b4f07ef\") " pod="openstack/glance-default-internal-api-0" Jan 26 17:09:37 crc kubenswrapper[4754]: I0126 17:09:37.127205 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c0decb48-8392-4a9a-a253-82b24b4f07ef-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"c0decb48-8392-4a9a-a253-82b24b4f07ef\") " pod="openstack/glance-default-internal-api-0" Jan 26 17:09:37 crc kubenswrapper[4754]: I0126 17:09:37.127262 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vxv68\" (UniqueName: \"kubernetes.io/projected/c0decb48-8392-4a9a-a253-82b24b4f07ef-kube-api-access-vxv68\") pod \"glance-default-internal-api-0\" (UID: \"c0decb48-8392-4a9a-a253-82b24b4f07ef\") " pod="openstack/glance-default-internal-api-0" Jan 26 17:09:37 crc kubenswrapper[4754]: I0126 17:09:37.127308 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c0decb48-8392-4a9a-a253-82b24b4f07ef-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"c0decb48-8392-4a9a-a253-82b24b4f07ef\") " pod="openstack/glance-default-internal-api-0" Jan 26 17:09:37 crc kubenswrapper[4754]: I0126 17:09:37.127349 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"c0decb48-8392-4a9a-a253-82b24b4f07ef\") " pod="openstack/glance-default-internal-api-0" Jan 26 17:09:37 crc kubenswrapper[4754]: I0126 17:09:37.127528 4754 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"c0decb48-8392-4a9a-a253-82b24b4f07ef\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/glance-default-internal-api-0" Jan 26 17:09:37 crc kubenswrapper[4754]: I0126 17:09:37.128742 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c0decb48-8392-4a9a-a253-82b24b4f07ef-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"c0decb48-8392-4a9a-a253-82b24b4f07ef\") " pod="openstack/glance-default-internal-api-0" Jan 26 17:09:37 crc kubenswrapper[4754]: I0126 17:09:37.129100 4754 patch_prober.go:28] interesting pod/machine-config-daemon-x65wv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 26 17:09:37 crc kubenswrapper[4754]: I0126 17:09:37.129147 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 26 17:09:37 crc kubenswrapper[4754]: I0126 17:09:37.132309 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c0decb48-8392-4a9a-a253-82b24b4f07ef-config-data\") pod \"glance-default-internal-api-0\" (UID: \"c0decb48-8392-4a9a-a253-82b24b4f07ef\") " pod="openstack/glance-default-internal-api-0" Jan 26 17:09:37 crc kubenswrapper[4754]: I0126 17:09:37.133557 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c0decb48-8392-4a9a-a253-82b24b4f07ef-scripts\") pod \"glance-default-internal-api-0\" (UID: \"c0decb48-8392-4a9a-a253-82b24b4f07ef\") " pod="openstack/glance-default-internal-api-0" Jan 26 17:09:37 crc kubenswrapper[4754]: I0126 17:09:37.133766 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c0decb48-8392-4a9a-a253-82b24b4f07ef-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"c0decb48-8392-4a9a-a253-82b24b4f07ef\") " pod="openstack/glance-default-internal-api-0" Jan 26 17:09:37 crc kubenswrapper[4754]: I0126 17:09:37.134335 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c0decb48-8392-4a9a-a253-82b24b4f07ef-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"c0decb48-8392-4a9a-a253-82b24b4f07ef\") " pod="openstack/glance-default-internal-api-0" Jan 26 17:09:37 crc kubenswrapper[4754]: I0126 17:09:37.143628 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vxv68\" (UniqueName: \"kubernetes.io/projected/c0decb48-8392-4a9a-a253-82b24b4f07ef-kube-api-access-vxv68\") pod \"glance-default-internal-api-0\" (UID: \"c0decb48-8392-4a9a-a253-82b24b4f07ef\") " pod="openstack/glance-default-internal-api-0" Jan 26 17:09:37 crc kubenswrapper[4754]: I0126 17:09:37.162505 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"c0decb48-8392-4a9a-a253-82b24b4f07ef\") " pod="openstack/glance-default-internal-api-0" Jan 26 17:09:37 crc kubenswrapper[4754]: I0126 17:09:37.284641 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Jan 26 17:09:37 crc kubenswrapper[4754]: I0126 17:09:37.780491 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="428912eb-4aed-4926-8b00-f716a43a4399" path="/var/lib/kubelet/pods/428912eb-4aed-4926-8b00-f716a43a4399/volumes" Jan 26 17:09:37 crc kubenswrapper[4754]: I0126 17:09:37.839536 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Jan 26 17:09:38 crc kubenswrapper[4754]: I0126 17:09:38.645499 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"c0decb48-8392-4a9a-a253-82b24b4f07ef","Type":"ContainerStarted","Data":"05b561cdfa840b822e6a60374a93d87dad867e17f82484134c17dca4b7646bcb"} Jan 26 17:09:38 crc kubenswrapper[4754]: I0126 17:09:38.646290 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"c0decb48-8392-4a9a-a253-82b24b4f07ef","Type":"ContainerStarted","Data":"8b6ee85240c62ac7067c60888af37a2630810ae8ad3abe3d399c208b6f3fec68"} Jan 26 17:09:38 crc kubenswrapper[4754]: I0126 17:09:38.661876 4754 generic.go:334] "Generic (PLEG): container finished" podID="6f82d0f5-47a4-4584-9ae9-046882fa8f61" containerID="eb836410836ef755e634f1dddd915e0296b6426b6a75ab5c090ec925a000a6ba" exitCode=0 Jan 26 17:09:38 crc kubenswrapper[4754]: I0126 17:09:38.661928 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"6f82d0f5-47a4-4584-9ae9-046882fa8f61","Type":"ContainerDied","Data":"eb836410836ef755e634f1dddd915e0296b6426b6a75ab5c090ec925a000a6ba"} Jan 26 17:09:39 crc kubenswrapper[4754]: I0126 17:09:39.098158 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Jan 26 17:09:39 crc kubenswrapper[4754]: I0126 17:09:39.155364 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Jan 26 17:09:39 crc kubenswrapper[4754]: I0126 17:09:39.169915 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6f82d0f5-47a4-4584-9ae9-046882fa8f61-scripts\") pod \"6f82d0f5-47a4-4584-9ae9-046882fa8f61\" (UID: \"6f82d0f5-47a4-4584-9ae9-046882fa8f61\") " Jan 26 17:09:39 crc kubenswrapper[4754]: I0126 17:09:39.170034 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6f82d0f5-47a4-4584-9ae9-046882fa8f61-etc-machine-id\") pod \"6f82d0f5-47a4-4584-9ae9-046882fa8f61\" (UID: \"6f82d0f5-47a4-4584-9ae9-046882fa8f61\") " Jan 26 17:09:39 crc kubenswrapper[4754]: I0126 17:09:39.170064 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9hzfn\" (UniqueName: \"kubernetes.io/projected/6f82d0f5-47a4-4584-9ae9-046882fa8f61-kube-api-access-9hzfn\") pod \"6f82d0f5-47a4-4584-9ae9-046882fa8f61\" (UID: \"6f82d0f5-47a4-4584-9ae9-046882fa8f61\") " Jan 26 17:09:39 crc kubenswrapper[4754]: I0126 17:09:39.170083 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6f82d0f5-47a4-4584-9ae9-046882fa8f61-config-data-custom\") pod \"6f82d0f5-47a4-4584-9ae9-046882fa8f61\" (UID: \"6f82d0f5-47a4-4584-9ae9-046882fa8f61\") " Jan 26 17:09:39 crc kubenswrapper[4754]: I0126 17:09:39.170190 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6f82d0f5-47a4-4584-9ae9-046882fa8f61-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "6f82d0f5-47a4-4584-9ae9-046882fa8f61" (UID: "6f82d0f5-47a4-4584-9ae9-046882fa8f61"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 17:09:39 crc kubenswrapper[4754]: I0126 17:09:39.171038 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6f82d0f5-47a4-4584-9ae9-046882fa8f61-config-data\") pod \"6f82d0f5-47a4-4584-9ae9-046882fa8f61\" (UID: \"6f82d0f5-47a4-4584-9ae9-046882fa8f61\") " Jan 26 17:09:39 crc kubenswrapper[4754]: I0126 17:09:39.171104 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f82d0f5-47a4-4584-9ae9-046882fa8f61-combined-ca-bundle\") pod \"6f82d0f5-47a4-4584-9ae9-046882fa8f61\" (UID: \"6f82d0f5-47a4-4584-9ae9-046882fa8f61\") " Jan 26 17:09:39 crc kubenswrapper[4754]: I0126 17:09:39.171814 4754 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6f82d0f5-47a4-4584-9ae9-046882fa8f61-etc-machine-id\") on node \"crc\" DevicePath \"\"" Jan 26 17:09:39 crc kubenswrapper[4754]: I0126 17:09:39.177870 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6f82d0f5-47a4-4584-9ae9-046882fa8f61-scripts" (OuterVolumeSpecName: "scripts") pod "6f82d0f5-47a4-4584-9ae9-046882fa8f61" (UID: "6f82d0f5-47a4-4584-9ae9-046882fa8f61"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:09:39 crc kubenswrapper[4754]: I0126 17:09:39.177892 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6f82d0f5-47a4-4584-9ae9-046882fa8f61-kube-api-access-9hzfn" (OuterVolumeSpecName: "kube-api-access-9hzfn") pod "6f82d0f5-47a4-4584-9ae9-046882fa8f61" (UID: "6f82d0f5-47a4-4584-9ae9-046882fa8f61"). InnerVolumeSpecName "kube-api-access-9hzfn". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:09:39 crc kubenswrapper[4754]: I0126 17:09:39.179935 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6f82d0f5-47a4-4584-9ae9-046882fa8f61-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "6f82d0f5-47a4-4584-9ae9-046882fa8f61" (UID: "6f82d0f5-47a4-4584-9ae9-046882fa8f61"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:09:39 crc kubenswrapper[4754]: I0126 17:09:39.240137 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6f82d0f5-47a4-4584-9ae9-046882fa8f61-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6f82d0f5-47a4-4584-9ae9-046882fa8f61" (UID: "6f82d0f5-47a4-4584-9ae9-046882fa8f61"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:09:39 crc kubenswrapper[4754]: I0126 17:09:39.273734 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9hzfn\" (UniqueName: \"kubernetes.io/projected/6f82d0f5-47a4-4584-9ae9-046882fa8f61-kube-api-access-9hzfn\") on node \"crc\" DevicePath \"\"" Jan 26 17:09:39 crc kubenswrapper[4754]: I0126 17:09:39.273769 4754 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6f82d0f5-47a4-4584-9ae9-046882fa8f61-config-data-custom\") on node \"crc\" DevicePath \"\"" Jan 26 17:09:39 crc kubenswrapper[4754]: I0126 17:09:39.273778 4754 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f82d0f5-47a4-4584-9ae9-046882fa8f61-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 26 17:09:39 crc kubenswrapper[4754]: I0126 17:09:39.273787 4754 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6f82d0f5-47a4-4584-9ae9-046882fa8f61-scripts\") on node \"crc\" DevicePath \"\"" Jan 26 17:09:39 crc kubenswrapper[4754]: I0126 17:09:39.282346 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6f82d0f5-47a4-4584-9ae9-046882fa8f61-config-data" (OuterVolumeSpecName: "config-data") pod "6f82d0f5-47a4-4584-9ae9-046882fa8f61" (UID: "6f82d0f5-47a4-4584-9ae9-046882fa8f61"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:09:39 crc kubenswrapper[4754]: I0126 17:09:39.284907 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Jan 26 17:09:39 crc kubenswrapper[4754]: I0126 17:09:39.284956 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Jan 26 17:09:39 crc kubenswrapper[4754]: I0126 17:09:39.361605 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Jan 26 17:09:39 crc kubenswrapper[4754]: I0126 17:09:39.368374 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Jan 26 17:09:39 crc kubenswrapper[4754]: I0126 17:09:39.375498 4754 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6f82d0f5-47a4-4584-9ae9-046882fa8f61-config-data\") on node \"crc\" DevicePath \"\"" Jan 26 17:09:39 crc kubenswrapper[4754]: I0126 17:09:39.671471 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Jan 26 17:09:39 crc kubenswrapper[4754]: I0126 17:09:39.671469 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"6f82d0f5-47a4-4584-9ae9-046882fa8f61","Type":"ContainerDied","Data":"ec92e2a2d83dd9be92b1f6fd8f1523b2351ac7fb16413d424503e4d0ed37f8b9"} Jan 26 17:09:39 crc kubenswrapper[4754]: I0126 17:09:39.671815 4754 scope.go:117] "RemoveContainer" containerID="66f684c6f13cd3e1ff23df6b19e556f74dc2dd34475bbf370be3391e82f922c0" Jan 26 17:09:39 crc kubenswrapper[4754]: I0126 17:09:39.673578 4754 generic.go:334] "Generic (PLEG): container finished" podID="7d5a6ee0-79d3-4112-a180-d2211fee09f6" containerID="d07947f249ed866995ee2e62e47cf9e3718926322c0dbaebe8e36d9013ee73f5" exitCode=0 Jan 26 17:09:39 crc kubenswrapper[4754]: I0126 17:09:39.673645 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-dzfm9" event={"ID":"7d5a6ee0-79d3-4112-a180-d2211fee09f6","Type":"ContainerDied","Data":"d07947f249ed866995ee2e62e47cf9e3718926322c0dbaebe8e36d9013ee73f5"} Jan 26 17:09:39 crc kubenswrapper[4754]: I0126 17:09:39.676651 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b63306d0-0c02-4691-81c7-60846f3c3fdc","Type":"ContainerStarted","Data":"8a78ded38db9ecaec0f9d2ab7c0ab888e5569d1c3ae81d30bda16a0bc370ff37"} Jan 26 17:09:39 crc kubenswrapper[4754]: I0126 17:09:39.676704 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b63306d0-0c02-4691-81c7-60846f3c3fdc","Type":"ContainerStarted","Data":"2670cc06722132104a1eed159fd13b54f7d18af3a482dc2d56fc11d1182a52d6"} Jan 26 17:09:39 crc kubenswrapper[4754]: I0126 17:09:39.679215 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"c0decb48-8392-4a9a-a253-82b24b4f07ef","Type":"ContainerStarted","Data":"ae3c3bb46794957eeadef3721fa719aabb093339c15cd80980f4fd23bb00cc22"} Jan 26 17:09:39 crc kubenswrapper[4754]: I0126 17:09:39.679450 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Jan 26 17:09:39 crc kubenswrapper[4754]: I0126 17:09:39.679495 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Jan 26 17:09:39 crc kubenswrapper[4754]: I0126 17:09:39.707176 4754 scope.go:117] "RemoveContainer" containerID="eb836410836ef755e634f1dddd915e0296b6426b6a75ab5c090ec925a000a6ba" Jan 26 17:09:39 crc kubenswrapper[4754]: I0126 17:09:39.735249 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=3.735232123 podStartE2EDuration="3.735232123s" podCreationTimestamp="2026-01-26 17:09:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 17:09:39.733911408 +0000 UTC m=+1346.258091842" watchObservedRunningTime="2026-01-26 17:09:39.735232123 +0000 UTC m=+1346.259412557" Jan 26 17:09:39 crc kubenswrapper[4754]: I0126 17:09:39.756801 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Jan 26 17:09:39 crc kubenswrapper[4754]: I0126 17:09:39.780070 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Jan 26 17:09:39 crc kubenswrapper[4754]: I0126 17:09:39.800100 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Jan 26 17:09:39 crc kubenswrapper[4754]: E0126 17:09:39.800487 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6f82d0f5-47a4-4584-9ae9-046882fa8f61" containerName="cinder-scheduler" Jan 26 17:09:39 crc kubenswrapper[4754]: I0126 17:09:39.800507 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="6f82d0f5-47a4-4584-9ae9-046882fa8f61" containerName="cinder-scheduler" Jan 26 17:09:39 crc kubenswrapper[4754]: E0126 17:09:39.800535 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6f82d0f5-47a4-4584-9ae9-046882fa8f61" containerName="probe" Jan 26 17:09:39 crc kubenswrapper[4754]: I0126 17:09:39.800542 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="6f82d0f5-47a4-4584-9ae9-046882fa8f61" containerName="probe" Jan 26 17:09:39 crc kubenswrapper[4754]: I0126 17:09:39.800732 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="6f82d0f5-47a4-4584-9ae9-046882fa8f61" containerName="probe" Jan 26 17:09:39 crc kubenswrapper[4754]: I0126 17:09:39.800749 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="6f82d0f5-47a4-4584-9ae9-046882fa8f61" containerName="cinder-scheduler" Jan 26 17:09:39 crc kubenswrapper[4754]: I0126 17:09:39.801694 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Jan 26 17:09:39 crc kubenswrapper[4754]: I0126 17:09:39.808783 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Jan 26 17:09:39 crc kubenswrapper[4754]: I0126 17:09:39.856473 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Jan 26 17:09:39 crc kubenswrapper[4754]: I0126 17:09:39.884503 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6147c61d-7878-41b8-8ce6-a165c9f03ede-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"6147c61d-7878-41b8-8ce6-a165c9f03ede\") " pod="openstack/cinder-scheduler-0" Jan 26 17:09:39 crc kubenswrapper[4754]: I0126 17:09:39.884553 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6147c61d-7878-41b8-8ce6-a165c9f03ede-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"6147c61d-7878-41b8-8ce6-a165c9f03ede\") " pod="openstack/cinder-scheduler-0" Jan 26 17:09:39 crc kubenswrapper[4754]: I0126 17:09:39.884596 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6147c61d-7878-41b8-8ce6-a165c9f03ede-scripts\") pod \"cinder-scheduler-0\" (UID: \"6147c61d-7878-41b8-8ce6-a165c9f03ede\") " pod="openstack/cinder-scheduler-0" Jan 26 17:09:39 crc kubenswrapper[4754]: I0126 17:09:39.885107 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6147c61d-7878-41b8-8ce6-a165c9f03ede-config-data\") pod \"cinder-scheduler-0\" (UID: \"6147c61d-7878-41b8-8ce6-a165c9f03ede\") " pod="openstack/cinder-scheduler-0" Jan 26 17:09:39 crc kubenswrapper[4754]: I0126 17:09:39.885233 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h2hp9\" (UniqueName: \"kubernetes.io/projected/6147c61d-7878-41b8-8ce6-a165c9f03ede-kube-api-access-h2hp9\") pod \"cinder-scheduler-0\" (UID: \"6147c61d-7878-41b8-8ce6-a165c9f03ede\") " pod="openstack/cinder-scheduler-0" Jan 26 17:09:39 crc kubenswrapper[4754]: I0126 17:09:39.885356 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6147c61d-7878-41b8-8ce6-a165c9f03ede-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"6147c61d-7878-41b8-8ce6-a165c9f03ede\") " pod="openstack/cinder-scheduler-0" Jan 26 17:09:39 crc kubenswrapper[4754]: I0126 17:09:39.987078 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6147c61d-7878-41b8-8ce6-a165c9f03ede-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"6147c61d-7878-41b8-8ce6-a165c9f03ede\") " pod="openstack/cinder-scheduler-0" Jan 26 17:09:39 crc kubenswrapper[4754]: I0126 17:09:39.987133 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6147c61d-7878-41b8-8ce6-a165c9f03ede-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"6147c61d-7878-41b8-8ce6-a165c9f03ede\") " pod="openstack/cinder-scheduler-0" Jan 26 17:09:39 crc kubenswrapper[4754]: I0126 17:09:39.987190 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6147c61d-7878-41b8-8ce6-a165c9f03ede-scripts\") pod \"cinder-scheduler-0\" (UID: \"6147c61d-7878-41b8-8ce6-a165c9f03ede\") " pod="openstack/cinder-scheduler-0" Jan 26 17:09:39 crc kubenswrapper[4754]: I0126 17:09:39.987269 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6147c61d-7878-41b8-8ce6-a165c9f03ede-config-data\") pod \"cinder-scheduler-0\" (UID: \"6147c61d-7878-41b8-8ce6-a165c9f03ede\") " pod="openstack/cinder-scheduler-0" Jan 26 17:09:39 crc kubenswrapper[4754]: I0126 17:09:39.987308 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6147c61d-7878-41b8-8ce6-a165c9f03ede-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"6147c61d-7878-41b8-8ce6-a165c9f03ede\") " pod="openstack/cinder-scheduler-0" Jan 26 17:09:39 crc kubenswrapper[4754]: I0126 17:09:39.987340 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h2hp9\" (UniqueName: \"kubernetes.io/projected/6147c61d-7878-41b8-8ce6-a165c9f03ede-kube-api-access-h2hp9\") pod \"cinder-scheduler-0\" (UID: \"6147c61d-7878-41b8-8ce6-a165c9f03ede\") " pod="openstack/cinder-scheduler-0" Jan 26 17:09:39 crc kubenswrapper[4754]: I0126 17:09:39.987500 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6147c61d-7878-41b8-8ce6-a165c9f03ede-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"6147c61d-7878-41b8-8ce6-a165c9f03ede\") " pod="openstack/cinder-scheduler-0" Jan 26 17:09:39 crc kubenswrapper[4754]: I0126 17:09:39.993999 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6147c61d-7878-41b8-8ce6-a165c9f03ede-config-data\") pod \"cinder-scheduler-0\" (UID: \"6147c61d-7878-41b8-8ce6-a165c9f03ede\") " pod="openstack/cinder-scheduler-0" Jan 26 17:09:39 crc kubenswrapper[4754]: I0126 17:09:39.994306 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6147c61d-7878-41b8-8ce6-a165c9f03ede-scripts\") pod \"cinder-scheduler-0\" (UID: \"6147c61d-7878-41b8-8ce6-a165c9f03ede\") " pod="openstack/cinder-scheduler-0" Jan 26 17:09:40 crc kubenswrapper[4754]: I0126 17:09:40.001305 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6147c61d-7878-41b8-8ce6-a165c9f03ede-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"6147c61d-7878-41b8-8ce6-a165c9f03ede\") " pod="openstack/cinder-scheduler-0" Jan 26 17:09:40 crc kubenswrapper[4754]: I0126 17:09:40.005479 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6147c61d-7878-41b8-8ce6-a165c9f03ede-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"6147c61d-7878-41b8-8ce6-a165c9f03ede\") " pod="openstack/cinder-scheduler-0" Jan 26 17:09:40 crc kubenswrapper[4754]: I0126 17:09:40.018270 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h2hp9\" (UniqueName: \"kubernetes.io/projected/6147c61d-7878-41b8-8ce6-a165c9f03ede-kube-api-access-h2hp9\") pod \"cinder-scheduler-0\" (UID: \"6147c61d-7878-41b8-8ce6-a165c9f03ede\") " pod="openstack/cinder-scheduler-0" Jan 26 17:09:40 crc kubenswrapper[4754]: I0126 17:09:40.146257 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Jan 26 17:09:40 crc kubenswrapper[4754]: I0126 17:09:40.667349 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Jan 26 17:09:40 crc kubenswrapper[4754]: I0126 17:09:40.691949 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"6147c61d-7878-41b8-8ce6-a165c9f03ede","Type":"ContainerStarted","Data":"705bc166b72c9a02abdde1198d2778a4e966caa1baccab818db467cab8d79520"} Jan 26 17:09:41 crc kubenswrapper[4754]: I0126 17:09:41.176479 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-dzfm9" Jan 26 17:09:41 crc kubenswrapper[4754]: I0126 17:09:41.327492 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d5a6ee0-79d3-4112-a180-d2211fee09f6-combined-ca-bundle\") pod \"7d5a6ee0-79d3-4112-a180-d2211fee09f6\" (UID: \"7d5a6ee0-79d3-4112-a180-d2211fee09f6\") " Jan 26 17:09:41 crc kubenswrapper[4754]: I0126 17:09:41.327563 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/7d5a6ee0-79d3-4112-a180-d2211fee09f6-config\") pod \"7d5a6ee0-79d3-4112-a180-d2211fee09f6\" (UID: \"7d5a6ee0-79d3-4112-a180-d2211fee09f6\") " Jan 26 17:09:41 crc kubenswrapper[4754]: I0126 17:09:41.327715 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ptw8x\" (UniqueName: \"kubernetes.io/projected/7d5a6ee0-79d3-4112-a180-d2211fee09f6-kube-api-access-ptw8x\") pod \"7d5a6ee0-79d3-4112-a180-d2211fee09f6\" (UID: \"7d5a6ee0-79d3-4112-a180-d2211fee09f6\") " Jan 26 17:09:41 crc kubenswrapper[4754]: I0126 17:09:41.334949 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7d5a6ee0-79d3-4112-a180-d2211fee09f6-kube-api-access-ptw8x" (OuterVolumeSpecName: "kube-api-access-ptw8x") pod "7d5a6ee0-79d3-4112-a180-d2211fee09f6" (UID: "7d5a6ee0-79d3-4112-a180-d2211fee09f6"). InnerVolumeSpecName "kube-api-access-ptw8x". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:09:41 crc kubenswrapper[4754]: I0126 17:09:41.362020 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7d5a6ee0-79d3-4112-a180-d2211fee09f6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7d5a6ee0-79d3-4112-a180-d2211fee09f6" (UID: "7d5a6ee0-79d3-4112-a180-d2211fee09f6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:09:41 crc kubenswrapper[4754]: I0126 17:09:41.363058 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7d5a6ee0-79d3-4112-a180-d2211fee09f6-config" (OuterVolumeSpecName: "config") pod "7d5a6ee0-79d3-4112-a180-d2211fee09f6" (UID: "7d5a6ee0-79d3-4112-a180-d2211fee09f6"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:09:41 crc kubenswrapper[4754]: I0126 17:09:41.430340 4754 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d5a6ee0-79d3-4112-a180-d2211fee09f6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 26 17:09:41 crc kubenswrapper[4754]: I0126 17:09:41.430388 4754 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/7d5a6ee0-79d3-4112-a180-d2211fee09f6-config\") on node \"crc\" DevicePath \"\"" Jan 26 17:09:41 crc kubenswrapper[4754]: I0126 17:09:41.430400 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ptw8x\" (UniqueName: \"kubernetes.io/projected/7d5a6ee0-79d3-4112-a180-d2211fee09f6-kube-api-access-ptw8x\") on node \"crc\" DevicePath \"\"" Jan 26 17:09:41 crc kubenswrapper[4754]: I0126 17:09:41.521762 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Jan 26 17:09:41 crc kubenswrapper[4754]: I0126 17:09:41.718134 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-dzfm9" event={"ID":"7d5a6ee0-79d3-4112-a180-d2211fee09f6","Type":"ContainerDied","Data":"64712fda37a9c52e0bb5053258072c3db8b94658f03933e4e85b50bda7af4cac"} Jan 26 17:09:41 crc kubenswrapper[4754]: I0126 17:09:41.718171 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="64712fda37a9c52e0bb5053258072c3db8b94658f03933e4e85b50bda7af4cac" Jan 26 17:09:41 crc kubenswrapper[4754]: I0126 17:09:41.718225 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-dzfm9" Jan 26 17:09:41 crc kubenswrapper[4754]: I0126 17:09:41.776482 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6f82d0f5-47a4-4584-9ae9-046882fa8f61" path="/var/lib/kubelet/pods/6f82d0f5-47a4-4584-9ae9-046882fa8f61/volumes" Jan 26 17:09:41 crc kubenswrapper[4754]: I0126 17:09:41.900958 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Jan 26 17:09:41 crc kubenswrapper[4754]: I0126 17:09:41.901060 4754 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Jan 26 17:09:41 crc kubenswrapper[4754]: I0126 17:09:41.903057 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Jan 26 17:09:44 crc kubenswrapper[4754]: I0126 17:09:44.297376 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-tnfmz"] Jan 26 17:09:44 crc kubenswrapper[4754]: E0126 17:09:44.298304 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d5a6ee0-79d3-4112-a180-d2211fee09f6" containerName="neutron-db-sync" Jan 26 17:09:44 crc kubenswrapper[4754]: I0126 17:09:44.298322 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d5a6ee0-79d3-4112-a180-d2211fee09f6" containerName="neutron-db-sync" Jan 26 17:09:44 crc kubenswrapper[4754]: I0126 17:09:44.298593 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="7d5a6ee0-79d3-4112-a180-d2211fee09f6" containerName="neutron-db-sync" Jan 26 17:09:44 crc kubenswrapper[4754]: I0126 17:09:44.306924 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-tnfmz" Jan 26 17:09:44 crc kubenswrapper[4754]: I0126 17:09:44.381579 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kr2nr\" (UniqueName: \"kubernetes.io/projected/952095c1-837b-4ff5-b61c-ce1e42102b56-kube-api-access-kr2nr\") pod \"dnsmasq-dns-5c9776ccc5-tnfmz\" (UID: \"952095c1-837b-4ff5-b61c-ce1e42102b56\") " pod="openstack/dnsmasq-dns-5c9776ccc5-tnfmz" Jan 26 17:09:44 crc kubenswrapper[4754]: I0126 17:09:44.381634 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/952095c1-837b-4ff5-b61c-ce1e42102b56-ovsdbserver-sb\") pod \"dnsmasq-dns-5c9776ccc5-tnfmz\" (UID: \"952095c1-837b-4ff5-b61c-ce1e42102b56\") " pod="openstack/dnsmasq-dns-5c9776ccc5-tnfmz" Jan 26 17:09:44 crc kubenswrapper[4754]: I0126 17:09:44.381704 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/952095c1-837b-4ff5-b61c-ce1e42102b56-ovsdbserver-nb\") pod \"dnsmasq-dns-5c9776ccc5-tnfmz\" (UID: \"952095c1-837b-4ff5-b61c-ce1e42102b56\") " pod="openstack/dnsmasq-dns-5c9776ccc5-tnfmz" Jan 26 17:09:44 crc kubenswrapper[4754]: I0126 17:09:44.381724 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/952095c1-837b-4ff5-b61c-ce1e42102b56-config\") pod \"dnsmasq-dns-5c9776ccc5-tnfmz\" (UID: \"952095c1-837b-4ff5-b61c-ce1e42102b56\") " pod="openstack/dnsmasq-dns-5c9776ccc5-tnfmz" Jan 26 17:09:44 crc kubenswrapper[4754]: I0126 17:09:44.381780 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/952095c1-837b-4ff5-b61c-ce1e42102b56-dns-svc\") pod \"dnsmasq-dns-5c9776ccc5-tnfmz\" (UID: \"952095c1-837b-4ff5-b61c-ce1e42102b56\") " pod="openstack/dnsmasq-dns-5c9776ccc5-tnfmz" Jan 26 17:09:44 crc kubenswrapper[4754]: I0126 17:09:44.381825 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/952095c1-837b-4ff5-b61c-ce1e42102b56-dns-swift-storage-0\") pod \"dnsmasq-dns-5c9776ccc5-tnfmz\" (UID: \"952095c1-837b-4ff5-b61c-ce1e42102b56\") " pod="openstack/dnsmasq-dns-5c9776ccc5-tnfmz" Jan 26 17:09:44 crc kubenswrapper[4754]: I0126 17:09:44.389707 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-tnfmz"] Jan 26 17:09:44 crc kubenswrapper[4754]: I0126 17:09:44.409734 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-55c9b8478b-6brjv"] Jan 26 17:09:44 crc kubenswrapper[4754]: I0126 17:09:44.413949 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-55c9b8478b-6brjv" Jan 26 17:09:44 crc kubenswrapper[4754]: I0126 17:09:44.418516 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Jan 26 17:09:44 crc kubenswrapper[4754]: I0126 17:09:44.418754 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-9z5cf" Jan 26 17:09:44 crc kubenswrapper[4754]: I0126 17:09:44.418856 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Jan 26 17:09:44 crc kubenswrapper[4754]: I0126 17:09:44.418796 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Jan 26 17:09:44 crc kubenswrapper[4754]: I0126 17:09:44.438518 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-55c9b8478b-6brjv"] Jan 26 17:09:44 crc kubenswrapper[4754]: I0126 17:09:44.483290 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/952095c1-837b-4ff5-b61c-ce1e42102b56-dns-swift-storage-0\") pod \"dnsmasq-dns-5c9776ccc5-tnfmz\" (UID: \"952095c1-837b-4ff5-b61c-ce1e42102b56\") " pod="openstack/dnsmasq-dns-5c9776ccc5-tnfmz" Jan 26 17:09:44 crc kubenswrapper[4754]: I0126 17:09:44.483330 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/deac50d1-deea-4092-b2a3-b3a7ada97a42-ovndb-tls-certs\") pod \"neutron-55c9b8478b-6brjv\" (UID: \"deac50d1-deea-4092-b2a3-b3a7ada97a42\") " pod="openstack/neutron-55c9b8478b-6brjv" Jan 26 17:09:44 crc kubenswrapper[4754]: I0126 17:09:44.483360 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kr2nr\" (UniqueName: \"kubernetes.io/projected/952095c1-837b-4ff5-b61c-ce1e42102b56-kube-api-access-kr2nr\") pod \"dnsmasq-dns-5c9776ccc5-tnfmz\" (UID: \"952095c1-837b-4ff5-b61c-ce1e42102b56\") " pod="openstack/dnsmasq-dns-5c9776ccc5-tnfmz" Jan 26 17:09:44 crc kubenswrapper[4754]: I0126 17:09:44.483391 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/952095c1-837b-4ff5-b61c-ce1e42102b56-ovsdbserver-sb\") pod \"dnsmasq-dns-5c9776ccc5-tnfmz\" (UID: \"952095c1-837b-4ff5-b61c-ce1e42102b56\") " pod="openstack/dnsmasq-dns-5c9776ccc5-tnfmz" Jan 26 17:09:44 crc kubenswrapper[4754]: I0126 17:09:44.483429 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-28mh4\" (UniqueName: \"kubernetes.io/projected/deac50d1-deea-4092-b2a3-b3a7ada97a42-kube-api-access-28mh4\") pod \"neutron-55c9b8478b-6brjv\" (UID: \"deac50d1-deea-4092-b2a3-b3a7ada97a42\") " pod="openstack/neutron-55c9b8478b-6brjv" Jan 26 17:09:44 crc kubenswrapper[4754]: I0126 17:09:44.483471 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/952095c1-837b-4ff5-b61c-ce1e42102b56-ovsdbserver-nb\") pod \"dnsmasq-dns-5c9776ccc5-tnfmz\" (UID: \"952095c1-837b-4ff5-b61c-ce1e42102b56\") " pod="openstack/dnsmasq-dns-5c9776ccc5-tnfmz" Jan 26 17:09:44 crc kubenswrapper[4754]: I0126 17:09:44.483489 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/952095c1-837b-4ff5-b61c-ce1e42102b56-config\") pod \"dnsmasq-dns-5c9776ccc5-tnfmz\" (UID: \"952095c1-837b-4ff5-b61c-ce1e42102b56\") " pod="openstack/dnsmasq-dns-5c9776ccc5-tnfmz" Jan 26 17:09:44 crc kubenswrapper[4754]: I0126 17:09:44.483532 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/deac50d1-deea-4092-b2a3-b3a7ada97a42-combined-ca-bundle\") pod \"neutron-55c9b8478b-6brjv\" (UID: \"deac50d1-deea-4092-b2a3-b3a7ada97a42\") " pod="openstack/neutron-55c9b8478b-6brjv" Jan 26 17:09:44 crc kubenswrapper[4754]: I0126 17:09:44.483548 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/deac50d1-deea-4092-b2a3-b3a7ada97a42-config\") pod \"neutron-55c9b8478b-6brjv\" (UID: \"deac50d1-deea-4092-b2a3-b3a7ada97a42\") " pod="openstack/neutron-55c9b8478b-6brjv" Jan 26 17:09:44 crc kubenswrapper[4754]: I0126 17:09:44.483580 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/deac50d1-deea-4092-b2a3-b3a7ada97a42-httpd-config\") pod \"neutron-55c9b8478b-6brjv\" (UID: \"deac50d1-deea-4092-b2a3-b3a7ada97a42\") " pod="openstack/neutron-55c9b8478b-6brjv" Jan 26 17:09:44 crc kubenswrapper[4754]: I0126 17:09:44.483604 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/952095c1-837b-4ff5-b61c-ce1e42102b56-dns-svc\") pod \"dnsmasq-dns-5c9776ccc5-tnfmz\" (UID: \"952095c1-837b-4ff5-b61c-ce1e42102b56\") " pod="openstack/dnsmasq-dns-5c9776ccc5-tnfmz" Jan 26 17:09:44 crc kubenswrapper[4754]: I0126 17:09:44.487328 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/952095c1-837b-4ff5-b61c-ce1e42102b56-dns-swift-storage-0\") pod \"dnsmasq-dns-5c9776ccc5-tnfmz\" (UID: \"952095c1-837b-4ff5-b61c-ce1e42102b56\") " pod="openstack/dnsmasq-dns-5c9776ccc5-tnfmz" Jan 26 17:09:44 crc kubenswrapper[4754]: I0126 17:09:44.488285 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/952095c1-837b-4ff5-b61c-ce1e42102b56-ovsdbserver-sb\") pod \"dnsmasq-dns-5c9776ccc5-tnfmz\" (UID: \"952095c1-837b-4ff5-b61c-ce1e42102b56\") " pod="openstack/dnsmasq-dns-5c9776ccc5-tnfmz" Jan 26 17:09:44 crc kubenswrapper[4754]: I0126 17:09:44.489005 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/952095c1-837b-4ff5-b61c-ce1e42102b56-ovsdbserver-nb\") pod \"dnsmasq-dns-5c9776ccc5-tnfmz\" (UID: \"952095c1-837b-4ff5-b61c-ce1e42102b56\") " pod="openstack/dnsmasq-dns-5c9776ccc5-tnfmz" Jan 26 17:09:44 crc kubenswrapper[4754]: I0126 17:09:44.489392 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/952095c1-837b-4ff5-b61c-ce1e42102b56-dns-svc\") pod \"dnsmasq-dns-5c9776ccc5-tnfmz\" (UID: \"952095c1-837b-4ff5-b61c-ce1e42102b56\") " pod="openstack/dnsmasq-dns-5c9776ccc5-tnfmz" Jan 26 17:09:44 crc kubenswrapper[4754]: I0126 17:09:44.489579 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/952095c1-837b-4ff5-b61c-ce1e42102b56-config\") pod \"dnsmasq-dns-5c9776ccc5-tnfmz\" (UID: \"952095c1-837b-4ff5-b61c-ce1e42102b56\") " pod="openstack/dnsmasq-dns-5c9776ccc5-tnfmz" Jan 26 17:09:44 crc kubenswrapper[4754]: I0126 17:09:44.565510 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kr2nr\" (UniqueName: \"kubernetes.io/projected/952095c1-837b-4ff5-b61c-ce1e42102b56-kube-api-access-kr2nr\") pod \"dnsmasq-dns-5c9776ccc5-tnfmz\" (UID: \"952095c1-837b-4ff5-b61c-ce1e42102b56\") " pod="openstack/dnsmasq-dns-5c9776ccc5-tnfmz" Jan 26 17:09:44 crc kubenswrapper[4754]: I0126 17:09:44.585326 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/deac50d1-deea-4092-b2a3-b3a7ada97a42-ovndb-tls-certs\") pod \"neutron-55c9b8478b-6brjv\" (UID: \"deac50d1-deea-4092-b2a3-b3a7ada97a42\") " pod="openstack/neutron-55c9b8478b-6brjv" Jan 26 17:09:44 crc kubenswrapper[4754]: I0126 17:09:44.585398 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-28mh4\" (UniqueName: \"kubernetes.io/projected/deac50d1-deea-4092-b2a3-b3a7ada97a42-kube-api-access-28mh4\") pod \"neutron-55c9b8478b-6brjv\" (UID: \"deac50d1-deea-4092-b2a3-b3a7ada97a42\") " pod="openstack/neutron-55c9b8478b-6brjv" Jan 26 17:09:44 crc kubenswrapper[4754]: I0126 17:09:44.585459 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/deac50d1-deea-4092-b2a3-b3a7ada97a42-combined-ca-bundle\") pod \"neutron-55c9b8478b-6brjv\" (UID: \"deac50d1-deea-4092-b2a3-b3a7ada97a42\") " pod="openstack/neutron-55c9b8478b-6brjv" Jan 26 17:09:44 crc kubenswrapper[4754]: I0126 17:09:44.585476 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/deac50d1-deea-4092-b2a3-b3a7ada97a42-config\") pod \"neutron-55c9b8478b-6brjv\" (UID: \"deac50d1-deea-4092-b2a3-b3a7ada97a42\") " pod="openstack/neutron-55c9b8478b-6brjv" Jan 26 17:09:44 crc kubenswrapper[4754]: I0126 17:09:44.585502 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/deac50d1-deea-4092-b2a3-b3a7ada97a42-httpd-config\") pod \"neutron-55c9b8478b-6brjv\" (UID: \"deac50d1-deea-4092-b2a3-b3a7ada97a42\") " pod="openstack/neutron-55c9b8478b-6brjv" Jan 26 17:09:44 crc kubenswrapper[4754]: I0126 17:09:44.589494 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/deac50d1-deea-4092-b2a3-b3a7ada97a42-httpd-config\") pod \"neutron-55c9b8478b-6brjv\" (UID: \"deac50d1-deea-4092-b2a3-b3a7ada97a42\") " pod="openstack/neutron-55c9b8478b-6brjv" Jan 26 17:09:44 crc kubenswrapper[4754]: I0126 17:09:44.592349 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/deac50d1-deea-4092-b2a3-b3a7ada97a42-ovndb-tls-certs\") pod \"neutron-55c9b8478b-6brjv\" (UID: \"deac50d1-deea-4092-b2a3-b3a7ada97a42\") " pod="openstack/neutron-55c9b8478b-6brjv" Jan 26 17:09:44 crc kubenswrapper[4754]: I0126 17:09:44.594448 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/deac50d1-deea-4092-b2a3-b3a7ada97a42-combined-ca-bundle\") pod \"neutron-55c9b8478b-6brjv\" (UID: \"deac50d1-deea-4092-b2a3-b3a7ada97a42\") " pod="openstack/neutron-55c9b8478b-6brjv" Jan 26 17:09:44 crc kubenswrapper[4754]: I0126 17:09:44.595137 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/deac50d1-deea-4092-b2a3-b3a7ada97a42-config\") pod \"neutron-55c9b8478b-6brjv\" (UID: \"deac50d1-deea-4092-b2a3-b3a7ada97a42\") " pod="openstack/neutron-55c9b8478b-6brjv" Jan 26 17:09:44 crc kubenswrapper[4754]: I0126 17:09:44.620374 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-28mh4\" (UniqueName: \"kubernetes.io/projected/deac50d1-deea-4092-b2a3-b3a7ada97a42-kube-api-access-28mh4\") pod \"neutron-55c9b8478b-6brjv\" (UID: \"deac50d1-deea-4092-b2a3-b3a7ada97a42\") " pod="openstack/neutron-55c9b8478b-6brjv" Jan 26 17:09:44 crc kubenswrapper[4754]: I0126 17:09:44.653497 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-tnfmz" Jan 26 17:09:44 crc kubenswrapper[4754]: I0126 17:09:44.744734 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-55c9b8478b-6brjv" Jan 26 17:09:44 crc kubenswrapper[4754]: I0126 17:09:44.794119 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"6147c61d-7878-41b8-8ce6-a165c9f03ede","Type":"ContainerStarted","Data":"2a4dcd98993eee4412dbdc2109d37576d780ccd35a1c511e8a8204c149116542"} Jan 26 17:09:45 crc kubenswrapper[4754]: I0126 17:09:45.279715 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-tnfmz"] Jan 26 17:09:45 crc kubenswrapper[4754]: I0126 17:09:45.558268 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-55c9b8478b-6brjv"] Jan 26 17:09:45 crc kubenswrapper[4754]: W0126 17:09:45.566522 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddeac50d1_deea_4092_b2a3_b3a7ada97a42.slice/crio-a78fa0f03066fa45e5ac9bd4ab5ab92bd6fda65a61a2b81b357d3be6c6bf4522 WatchSource:0}: Error finding container a78fa0f03066fa45e5ac9bd4ab5ab92bd6fda65a61a2b81b357d3be6c6bf4522: Status 404 returned error can't find the container with id a78fa0f03066fa45e5ac9bd4ab5ab92bd6fda65a61a2b81b357d3be6c6bf4522 Jan 26 17:09:45 crc kubenswrapper[4754]: I0126 17:09:45.817884 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-55c9b8478b-6brjv" event={"ID":"deac50d1-deea-4092-b2a3-b3a7ada97a42","Type":"ContainerStarted","Data":"a78fa0f03066fa45e5ac9bd4ab5ab92bd6fda65a61a2b81b357d3be6c6bf4522"} Jan 26 17:09:45 crc kubenswrapper[4754]: I0126 17:09:45.824957 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b63306d0-0c02-4691-81c7-60846f3c3fdc","Type":"ContainerStarted","Data":"0c99693ff01997432dd6d02f5cc330efc21203955c8cb00258ed8043aebcab9c"} Jan 26 17:09:45 crc kubenswrapper[4754]: I0126 17:09:45.825149 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Jan 26 17:09:45 crc kubenswrapper[4754]: I0126 17:09:45.825173 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b63306d0-0c02-4691-81c7-60846f3c3fdc" containerName="sg-core" containerID="cri-o://8a78ded38db9ecaec0f9d2ab7c0ab888e5569d1c3ae81d30bda16a0bc370ff37" gracePeriod=30 Jan 26 17:09:45 crc kubenswrapper[4754]: I0126 17:09:45.825134 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b63306d0-0c02-4691-81c7-60846f3c3fdc" containerName="ceilometer-central-agent" containerID="cri-o://6dbb6881c0cdc0f402b9772f369d7222270b21d8425d63e6c42a80db5944dcde" gracePeriod=30 Jan 26 17:09:45 crc kubenswrapper[4754]: I0126 17:09:45.825224 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b63306d0-0c02-4691-81c7-60846f3c3fdc" containerName="ceilometer-notification-agent" containerID="cri-o://2670cc06722132104a1eed159fd13b54f7d18af3a482dc2d56fc11d1182a52d6" gracePeriod=30 Jan 26 17:09:45 crc kubenswrapper[4754]: I0126 17:09:45.825341 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b63306d0-0c02-4691-81c7-60846f3c3fdc" containerName="proxy-httpd" containerID="cri-o://0c99693ff01997432dd6d02f5cc330efc21203955c8cb00258ed8043aebcab9c" gracePeriod=30 Jan 26 17:09:45 crc kubenswrapper[4754]: I0126 17:09:45.827918 4754 generic.go:334] "Generic (PLEG): container finished" podID="952095c1-837b-4ff5-b61c-ce1e42102b56" containerID="86bc9c5562701990e9174a443eecc059bb82afee194d1bfe69d2a52c8e8dc8ed" exitCode=0 Jan 26 17:09:45 crc kubenswrapper[4754]: I0126 17:09:45.828228 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-tnfmz" event={"ID":"952095c1-837b-4ff5-b61c-ce1e42102b56","Type":"ContainerDied","Data":"86bc9c5562701990e9174a443eecc059bb82afee194d1bfe69d2a52c8e8dc8ed"} Jan 26 17:09:45 crc kubenswrapper[4754]: I0126 17:09:45.828257 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-tnfmz" event={"ID":"952095c1-837b-4ff5-b61c-ce1e42102b56","Type":"ContainerStarted","Data":"373bbf9cbb4c96b3d33d178926090aed6931500656ad7c69799b9b4486f29eb0"} Jan 26 17:09:45 crc kubenswrapper[4754]: I0126 17:09:45.866362 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.273282088 podStartE2EDuration="13.866347809s" podCreationTimestamp="2026-01-26 17:09:32 +0000 UTC" firstStartedPulling="2026-01-26 17:09:33.462411058 +0000 UTC m=+1339.986591492" lastFinishedPulling="2026-01-26 17:09:45.055476779 +0000 UTC m=+1351.579657213" observedRunningTime="2026-01-26 17:09:45.864529512 +0000 UTC m=+1352.388709946" watchObservedRunningTime="2026-01-26 17:09:45.866347809 +0000 UTC m=+1352.390528243" Jan 26 17:09:46 crc kubenswrapper[4754]: I0126 17:09:46.684652 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-767b9f4849-cfqbz"] Jan 26 17:09:46 crc kubenswrapper[4754]: I0126 17:09:46.686470 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-767b9f4849-cfqbz" Jan 26 17:09:46 crc kubenswrapper[4754]: I0126 17:09:46.690810 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Jan 26 17:09:46 crc kubenswrapper[4754]: I0126 17:09:46.691006 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Jan 26 17:09:46 crc kubenswrapper[4754]: I0126 17:09:46.701006 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-767b9f4849-cfqbz"] Jan 26 17:09:46 crc kubenswrapper[4754]: I0126 17:09:46.835453 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9l5w9\" (UniqueName: \"kubernetes.io/projected/36eccc33-604a-49ac-a1f2-c0bd41fd053d-kube-api-access-9l5w9\") pod \"neutron-767b9f4849-cfqbz\" (UID: \"36eccc33-604a-49ac-a1f2-c0bd41fd053d\") " pod="openstack/neutron-767b9f4849-cfqbz" Jan 26 17:09:46 crc kubenswrapper[4754]: I0126 17:09:46.835626 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/36eccc33-604a-49ac-a1f2-c0bd41fd053d-internal-tls-certs\") pod \"neutron-767b9f4849-cfqbz\" (UID: \"36eccc33-604a-49ac-a1f2-c0bd41fd053d\") " pod="openstack/neutron-767b9f4849-cfqbz" Jan 26 17:09:46 crc kubenswrapper[4754]: I0126 17:09:46.836012 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/36eccc33-604a-49ac-a1f2-c0bd41fd053d-ovndb-tls-certs\") pod \"neutron-767b9f4849-cfqbz\" (UID: \"36eccc33-604a-49ac-a1f2-c0bd41fd053d\") " pod="openstack/neutron-767b9f4849-cfqbz" Jan 26 17:09:46 crc kubenswrapper[4754]: I0126 17:09:46.836053 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/36eccc33-604a-49ac-a1f2-c0bd41fd053d-httpd-config\") pod \"neutron-767b9f4849-cfqbz\" (UID: \"36eccc33-604a-49ac-a1f2-c0bd41fd053d\") " pod="openstack/neutron-767b9f4849-cfqbz" Jan 26 17:09:46 crc kubenswrapper[4754]: I0126 17:09:46.836108 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/36eccc33-604a-49ac-a1f2-c0bd41fd053d-public-tls-certs\") pod \"neutron-767b9f4849-cfqbz\" (UID: \"36eccc33-604a-49ac-a1f2-c0bd41fd053d\") " pod="openstack/neutron-767b9f4849-cfqbz" Jan 26 17:09:46 crc kubenswrapper[4754]: I0126 17:09:46.836194 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36eccc33-604a-49ac-a1f2-c0bd41fd053d-combined-ca-bundle\") pod \"neutron-767b9f4849-cfqbz\" (UID: \"36eccc33-604a-49ac-a1f2-c0bd41fd053d\") " pod="openstack/neutron-767b9f4849-cfqbz" Jan 26 17:09:46 crc kubenswrapper[4754]: I0126 17:09:46.836257 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/36eccc33-604a-49ac-a1f2-c0bd41fd053d-config\") pod \"neutron-767b9f4849-cfqbz\" (UID: \"36eccc33-604a-49ac-a1f2-c0bd41fd053d\") " pod="openstack/neutron-767b9f4849-cfqbz" Jan 26 17:09:46 crc kubenswrapper[4754]: I0126 17:09:46.866793 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-tnfmz" event={"ID":"952095c1-837b-4ff5-b61c-ce1e42102b56","Type":"ContainerStarted","Data":"a9080148a080f6c8a5352be74755999d3eb0d22a9a34cdd78cd8072a9a4d3f4f"} Jan 26 17:09:46 crc kubenswrapper[4754]: I0126 17:09:46.866994 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5c9776ccc5-tnfmz" Jan 26 17:09:46 crc kubenswrapper[4754]: I0126 17:09:46.894283 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5c9776ccc5-tnfmz" podStartSLOduration=2.894264508 podStartE2EDuration="2.894264508s" podCreationTimestamp="2026-01-26 17:09:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 17:09:46.887016129 +0000 UTC m=+1353.411196563" watchObservedRunningTime="2026-01-26 17:09:46.894264508 +0000 UTC m=+1353.418444942" Jan 26 17:09:46 crc kubenswrapper[4754]: I0126 17:09:46.896380 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"6147c61d-7878-41b8-8ce6-a165c9f03ede","Type":"ContainerStarted","Data":"90c1ad6f06bc49a5342b2f8cd01b485081c48c78ac84edb7560b390019d8031f"} Jan 26 17:09:46 crc kubenswrapper[4754]: I0126 17:09:46.922015 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-55c9b8478b-6brjv" event={"ID":"deac50d1-deea-4092-b2a3-b3a7ada97a42","Type":"ContainerStarted","Data":"236452eb8f2e8caab9a57c5cf6bb43e5b06a51eb5120a207826a17214da8fbb2"} Jan 26 17:09:46 crc kubenswrapper[4754]: I0126 17:09:46.922064 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-55c9b8478b-6brjv" event={"ID":"deac50d1-deea-4092-b2a3-b3a7ada97a42","Type":"ContainerStarted","Data":"330d0f3072f01682231689eb69fa008dec465696c9b685a3f2ea9108f318cb1f"} Jan 26 17:09:46 crc kubenswrapper[4754]: I0126 17:09:46.923209 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-55c9b8478b-6brjv" Jan 26 17:09:46 crc kubenswrapper[4754]: I0126 17:09:46.927183 4754 generic.go:334] "Generic (PLEG): container finished" podID="b63306d0-0c02-4691-81c7-60846f3c3fdc" containerID="0c99693ff01997432dd6d02f5cc330efc21203955c8cb00258ed8043aebcab9c" exitCode=0 Jan 26 17:09:46 crc kubenswrapper[4754]: I0126 17:09:46.927216 4754 generic.go:334] "Generic (PLEG): container finished" podID="b63306d0-0c02-4691-81c7-60846f3c3fdc" containerID="8a78ded38db9ecaec0f9d2ab7c0ab888e5569d1c3ae81d30bda16a0bc370ff37" exitCode=2 Jan 26 17:09:46 crc kubenswrapper[4754]: I0126 17:09:46.927226 4754 generic.go:334] "Generic (PLEG): container finished" podID="b63306d0-0c02-4691-81c7-60846f3c3fdc" containerID="2670cc06722132104a1eed159fd13b54f7d18af3a482dc2d56fc11d1182a52d6" exitCode=0 Jan 26 17:09:46 crc kubenswrapper[4754]: I0126 17:09:46.927236 4754 generic.go:334] "Generic (PLEG): container finished" podID="b63306d0-0c02-4691-81c7-60846f3c3fdc" containerID="6dbb6881c0cdc0f402b9772f369d7222270b21d8425d63e6c42a80db5944dcde" exitCode=0 Jan 26 17:09:46 crc kubenswrapper[4754]: I0126 17:09:46.927256 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b63306d0-0c02-4691-81c7-60846f3c3fdc","Type":"ContainerDied","Data":"0c99693ff01997432dd6d02f5cc330efc21203955c8cb00258ed8043aebcab9c"} Jan 26 17:09:46 crc kubenswrapper[4754]: I0126 17:09:46.927276 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b63306d0-0c02-4691-81c7-60846f3c3fdc","Type":"ContainerDied","Data":"8a78ded38db9ecaec0f9d2ab7c0ab888e5569d1c3ae81d30bda16a0bc370ff37"} Jan 26 17:09:46 crc kubenswrapper[4754]: I0126 17:09:46.927288 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b63306d0-0c02-4691-81c7-60846f3c3fdc","Type":"ContainerDied","Data":"2670cc06722132104a1eed159fd13b54f7d18af3a482dc2d56fc11d1182a52d6"} Jan 26 17:09:46 crc kubenswrapper[4754]: I0126 17:09:46.927300 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b63306d0-0c02-4691-81c7-60846f3c3fdc","Type":"ContainerDied","Data":"6dbb6881c0cdc0f402b9772f369d7222270b21d8425d63e6c42a80db5944dcde"} Jan 26 17:09:46 crc kubenswrapper[4754]: I0126 17:09:46.928542 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=7.928529335 podStartE2EDuration="7.928529335s" podCreationTimestamp="2026-01-26 17:09:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 17:09:46.913123382 +0000 UTC m=+1353.437303836" watchObservedRunningTime="2026-01-26 17:09:46.928529335 +0000 UTC m=+1353.452709769" Jan 26 17:09:46 crc kubenswrapper[4754]: I0126 17:09:46.938272 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/36eccc33-604a-49ac-a1f2-c0bd41fd053d-ovndb-tls-certs\") pod \"neutron-767b9f4849-cfqbz\" (UID: \"36eccc33-604a-49ac-a1f2-c0bd41fd053d\") " pod="openstack/neutron-767b9f4849-cfqbz" Jan 26 17:09:46 crc kubenswrapper[4754]: I0126 17:09:46.938314 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/36eccc33-604a-49ac-a1f2-c0bd41fd053d-httpd-config\") pod \"neutron-767b9f4849-cfqbz\" (UID: \"36eccc33-604a-49ac-a1f2-c0bd41fd053d\") " pod="openstack/neutron-767b9f4849-cfqbz" Jan 26 17:09:46 crc kubenswrapper[4754]: I0126 17:09:46.938333 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/36eccc33-604a-49ac-a1f2-c0bd41fd053d-public-tls-certs\") pod \"neutron-767b9f4849-cfqbz\" (UID: \"36eccc33-604a-49ac-a1f2-c0bd41fd053d\") " pod="openstack/neutron-767b9f4849-cfqbz" Jan 26 17:09:46 crc kubenswrapper[4754]: I0126 17:09:46.938378 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36eccc33-604a-49ac-a1f2-c0bd41fd053d-combined-ca-bundle\") pod \"neutron-767b9f4849-cfqbz\" (UID: \"36eccc33-604a-49ac-a1f2-c0bd41fd053d\") " pod="openstack/neutron-767b9f4849-cfqbz" Jan 26 17:09:46 crc kubenswrapper[4754]: I0126 17:09:46.938421 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/36eccc33-604a-49ac-a1f2-c0bd41fd053d-config\") pod \"neutron-767b9f4849-cfqbz\" (UID: \"36eccc33-604a-49ac-a1f2-c0bd41fd053d\") " pod="openstack/neutron-767b9f4849-cfqbz" Jan 26 17:09:46 crc kubenswrapper[4754]: I0126 17:09:46.938449 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9l5w9\" (UniqueName: \"kubernetes.io/projected/36eccc33-604a-49ac-a1f2-c0bd41fd053d-kube-api-access-9l5w9\") pod \"neutron-767b9f4849-cfqbz\" (UID: \"36eccc33-604a-49ac-a1f2-c0bd41fd053d\") " pod="openstack/neutron-767b9f4849-cfqbz" Jan 26 17:09:46 crc kubenswrapper[4754]: I0126 17:09:46.938535 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/36eccc33-604a-49ac-a1f2-c0bd41fd053d-internal-tls-certs\") pod \"neutron-767b9f4849-cfqbz\" (UID: \"36eccc33-604a-49ac-a1f2-c0bd41fd053d\") " pod="openstack/neutron-767b9f4849-cfqbz" Jan 26 17:09:46 crc kubenswrapper[4754]: I0126 17:09:46.950003 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/36eccc33-604a-49ac-a1f2-c0bd41fd053d-ovndb-tls-certs\") pod \"neutron-767b9f4849-cfqbz\" (UID: \"36eccc33-604a-49ac-a1f2-c0bd41fd053d\") " pod="openstack/neutron-767b9f4849-cfqbz" Jan 26 17:09:46 crc kubenswrapper[4754]: I0126 17:09:46.955422 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-55c9b8478b-6brjv" podStartSLOduration=2.955401998 podStartE2EDuration="2.955401998s" podCreationTimestamp="2026-01-26 17:09:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 17:09:46.955040289 +0000 UTC m=+1353.479220713" watchObservedRunningTime="2026-01-26 17:09:46.955401998 +0000 UTC m=+1353.479582442" Jan 26 17:09:46 crc kubenswrapper[4754]: I0126 17:09:46.956275 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/36eccc33-604a-49ac-a1f2-c0bd41fd053d-httpd-config\") pod \"neutron-767b9f4849-cfqbz\" (UID: \"36eccc33-604a-49ac-a1f2-c0bd41fd053d\") " pod="openstack/neutron-767b9f4849-cfqbz" Jan 26 17:09:46 crc kubenswrapper[4754]: I0126 17:09:46.959204 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/36eccc33-604a-49ac-a1f2-c0bd41fd053d-public-tls-certs\") pod \"neutron-767b9f4849-cfqbz\" (UID: \"36eccc33-604a-49ac-a1f2-c0bd41fd053d\") " pod="openstack/neutron-767b9f4849-cfqbz" Jan 26 17:09:46 crc kubenswrapper[4754]: I0126 17:09:46.959962 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9l5w9\" (UniqueName: \"kubernetes.io/projected/36eccc33-604a-49ac-a1f2-c0bd41fd053d-kube-api-access-9l5w9\") pod \"neutron-767b9f4849-cfqbz\" (UID: \"36eccc33-604a-49ac-a1f2-c0bd41fd053d\") " pod="openstack/neutron-767b9f4849-cfqbz" Jan 26 17:09:46 crc kubenswrapper[4754]: I0126 17:09:46.974033 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/36eccc33-604a-49ac-a1f2-c0bd41fd053d-internal-tls-certs\") pod \"neutron-767b9f4849-cfqbz\" (UID: \"36eccc33-604a-49ac-a1f2-c0bd41fd053d\") " pod="openstack/neutron-767b9f4849-cfqbz" Jan 26 17:09:46 crc kubenswrapper[4754]: I0126 17:09:46.981791 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36eccc33-604a-49ac-a1f2-c0bd41fd053d-combined-ca-bundle\") pod \"neutron-767b9f4849-cfqbz\" (UID: \"36eccc33-604a-49ac-a1f2-c0bd41fd053d\") " pod="openstack/neutron-767b9f4849-cfqbz" Jan 26 17:09:46 crc kubenswrapper[4754]: I0126 17:09:46.982430 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/36eccc33-604a-49ac-a1f2-c0bd41fd053d-config\") pod \"neutron-767b9f4849-cfqbz\" (UID: \"36eccc33-604a-49ac-a1f2-c0bd41fd053d\") " pod="openstack/neutron-767b9f4849-cfqbz" Jan 26 17:09:47 crc kubenswrapper[4754]: I0126 17:09:47.039520 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-767b9f4849-cfqbz" Jan 26 17:09:47 crc kubenswrapper[4754]: I0126 17:09:47.184434 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 26 17:09:47 crc kubenswrapper[4754]: I0126 17:09:47.286322 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Jan 26 17:09:47 crc kubenswrapper[4754]: I0126 17:09:47.286592 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Jan 26 17:09:47 crc kubenswrapper[4754]: I0126 17:09:47.328153 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Jan 26 17:09:47 crc kubenswrapper[4754]: I0126 17:09:47.348233 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b63306d0-0c02-4691-81c7-60846f3c3fdc-sg-core-conf-yaml\") pod \"b63306d0-0c02-4691-81c7-60846f3c3fdc\" (UID: \"b63306d0-0c02-4691-81c7-60846f3c3fdc\") " Jan 26 17:09:47 crc kubenswrapper[4754]: I0126 17:09:47.348314 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b63306d0-0c02-4691-81c7-60846f3c3fdc-combined-ca-bundle\") pod \"b63306d0-0c02-4691-81c7-60846f3c3fdc\" (UID: \"b63306d0-0c02-4691-81c7-60846f3c3fdc\") " Jan 26 17:09:47 crc kubenswrapper[4754]: I0126 17:09:47.348341 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jpqgs\" (UniqueName: \"kubernetes.io/projected/b63306d0-0c02-4691-81c7-60846f3c3fdc-kube-api-access-jpqgs\") pod \"b63306d0-0c02-4691-81c7-60846f3c3fdc\" (UID: \"b63306d0-0c02-4691-81c7-60846f3c3fdc\") " Jan 26 17:09:47 crc kubenswrapper[4754]: I0126 17:09:47.348379 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b63306d0-0c02-4691-81c7-60846f3c3fdc-log-httpd\") pod \"b63306d0-0c02-4691-81c7-60846f3c3fdc\" (UID: \"b63306d0-0c02-4691-81c7-60846f3c3fdc\") " Jan 26 17:09:47 crc kubenswrapper[4754]: I0126 17:09:47.348417 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b63306d0-0c02-4691-81c7-60846f3c3fdc-config-data\") pod \"b63306d0-0c02-4691-81c7-60846f3c3fdc\" (UID: \"b63306d0-0c02-4691-81c7-60846f3c3fdc\") " Jan 26 17:09:47 crc kubenswrapper[4754]: I0126 17:09:47.348479 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b63306d0-0c02-4691-81c7-60846f3c3fdc-run-httpd\") pod \"b63306d0-0c02-4691-81c7-60846f3c3fdc\" (UID: \"b63306d0-0c02-4691-81c7-60846f3c3fdc\") " Jan 26 17:09:47 crc kubenswrapper[4754]: I0126 17:09:47.348603 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b63306d0-0c02-4691-81c7-60846f3c3fdc-scripts\") pod \"b63306d0-0c02-4691-81c7-60846f3c3fdc\" (UID: \"b63306d0-0c02-4691-81c7-60846f3c3fdc\") " Jan 26 17:09:47 crc kubenswrapper[4754]: I0126 17:09:47.350970 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b63306d0-0c02-4691-81c7-60846f3c3fdc-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "b63306d0-0c02-4691-81c7-60846f3c3fdc" (UID: "b63306d0-0c02-4691-81c7-60846f3c3fdc"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:09:47 crc kubenswrapper[4754]: I0126 17:09:47.351117 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b63306d0-0c02-4691-81c7-60846f3c3fdc-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "b63306d0-0c02-4691-81c7-60846f3c3fdc" (UID: "b63306d0-0c02-4691-81c7-60846f3c3fdc"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:09:47 crc kubenswrapper[4754]: I0126 17:09:47.355916 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Jan 26 17:09:47 crc kubenswrapper[4754]: I0126 17:09:47.381967 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b63306d0-0c02-4691-81c7-60846f3c3fdc-scripts" (OuterVolumeSpecName: "scripts") pod "b63306d0-0c02-4691-81c7-60846f3c3fdc" (UID: "b63306d0-0c02-4691-81c7-60846f3c3fdc"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:09:47 crc kubenswrapper[4754]: I0126 17:09:47.403915 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b63306d0-0c02-4691-81c7-60846f3c3fdc-kube-api-access-jpqgs" (OuterVolumeSpecName: "kube-api-access-jpqgs") pod "b63306d0-0c02-4691-81c7-60846f3c3fdc" (UID: "b63306d0-0c02-4691-81c7-60846f3c3fdc"). InnerVolumeSpecName "kube-api-access-jpqgs". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:09:47 crc kubenswrapper[4754]: I0126 17:09:47.421005 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b63306d0-0c02-4691-81c7-60846f3c3fdc-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "b63306d0-0c02-4691-81c7-60846f3c3fdc" (UID: "b63306d0-0c02-4691-81c7-60846f3c3fdc"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:09:47 crc kubenswrapper[4754]: I0126 17:09:47.456869 4754 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b63306d0-0c02-4691-81c7-60846f3c3fdc-run-httpd\") on node \"crc\" DevicePath \"\"" Jan 26 17:09:47 crc kubenswrapper[4754]: I0126 17:09:47.456909 4754 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b63306d0-0c02-4691-81c7-60846f3c3fdc-scripts\") on node \"crc\" DevicePath \"\"" Jan 26 17:09:47 crc kubenswrapper[4754]: I0126 17:09:47.456956 4754 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b63306d0-0c02-4691-81c7-60846f3c3fdc-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Jan 26 17:09:47 crc kubenswrapper[4754]: I0126 17:09:47.456973 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jpqgs\" (UniqueName: \"kubernetes.io/projected/b63306d0-0c02-4691-81c7-60846f3c3fdc-kube-api-access-jpqgs\") on node \"crc\" DevicePath \"\"" Jan 26 17:09:47 crc kubenswrapper[4754]: I0126 17:09:47.456986 4754 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b63306d0-0c02-4691-81c7-60846f3c3fdc-log-httpd\") on node \"crc\" DevicePath \"\"" Jan 26 17:09:47 crc kubenswrapper[4754]: I0126 17:09:47.508691 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b63306d0-0c02-4691-81c7-60846f3c3fdc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b63306d0-0c02-4691-81c7-60846f3c3fdc" (UID: "b63306d0-0c02-4691-81c7-60846f3c3fdc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:09:47 crc kubenswrapper[4754]: I0126 17:09:47.558757 4754 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b63306d0-0c02-4691-81c7-60846f3c3fdc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 26 17:09:47 crc kubenswrapper[4754]: I0126 17:09:47.579305 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b63306d0-0c02-4691-81c7-60846f3c3fdc-config-data" (OuterVolumeSpecName: "config-data") pod "b63306d0-0c02-4691-81c7-60846f3c3fdc" (UID: "b63306d0-0c02-4691-81c7-60846f3c3fdc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:09:47 crc kubenswrapper[4754]: I0126 17:09:47.661123 4754 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b63306d0-0c02-4691-81c7-60846f3c3fdc-config-data\") on node \"crc\" DevicePath \"\"" Jan 26 17:09:47 crc kubenswrapper[4754]: W0126 17:09:47.840117 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod36eccc33_604a_49ac_a1f2_c0bd41fd053d.slice/crio-23287c322db887ab69e349a2a7bb8ecf9d63773aeee232dcd1b157b77758877e WatchSource:0}: Error finding container 23287c322db887ab69e349a2a7bb8ecf9d63773aeee232dcd1b157b77758877e: Status 404 returned error can't find the container with id 23287c322db887ab69e349a2a7bb8ecf9d63773aeee232dcd1b157b77758877e Jan 26 17:09:47 crc kubenswrapper[4754]: I0126 17:09:47.843687 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-767b9f4849-cfqbz"] Jan 26 17:09:47 crc kubenswrapper[4754]: I0126 17:09:47.949370 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b63306d0-0c02-4691-81c7-60846f3c3fdc","Type":"ContainerDied","Data":"cbf2ae3def1abc7c686fb44f45c9e78d9557a69bc9f2ab121c6b1e8b279357ac"} Jan 26 17:09:47 crc kubenswrapper[4754]: I0126 17:09:47.949440 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 26 17:09:47 crc kubenswrapper[4754]: I0126 17:09:47.949725 4754 scope.go:117] "RemoveContainer" containerID="0c99693ff01997432dd6d02f5cc330efc21203955c8cb00258ed8043aebcab9c" Jan 26 17:09:47 crc kubenswrapper[4754]: I0126 17:09:47.958505 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-767b9f4849-cfqbz" event={"ID":"36eccc33-604a-49ac-a1f2-c0bd41fd053d","Type":"ContainerStarted","Data":"23287c322db887ab69e349a2a7bb8ecf9d63773aeee232dcd1b157b77758877e"} Jan 26 17:09:47 crc kubenswrapper[4754]: I0126 17:09:47.959472 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Jan 26 17:09:47 crc kubenswrapper[4754]: I0126 17:09:47.959511 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Jan 26 17:09:48 crc kubenswrapper[4754]: I0126 17:09:48.029168 4754 scope.go:117] "RemoveContainer" containerID="8a78ded38db9ecaec0f9d2ab7c0ab888e5569d1c3ae81d30bda16a0bc370ff37" Jan 26 17:09:48 crc kubenswrapper[4754]: I0126 17:09:48.047206 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Jan 26 17:09:48 crc kubenswrapper[4754]: I0126 17:09:48.060680 4754 scope.go:117] "RemoveContainer" containerID="2670cc06722132104a1eed159fd13b54f7d18af3a482dc2d56fc11d1182a52d6" Jan 26 17:09:48 crc kubenswrapper[4754]: I0126 17:09:48.068056 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Jan 26 17:09:48 crc kubenswrapper[4754]: I0126 17:09:48.082486 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Jan 26 17:09:48 crc kubenswrapper[4754]: E0126 17:09:48.083426 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b63306d0-0c02-4691-81c7-60846f3c3fdc" containerName="ceilometer-notification-agent" Jan 26 17:09:48 crc kubenswrapper[4754]: I0126 17:09:48.083483 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="b63306d0-0c02-4691-81c7-60846f3c3fdc" containerName="ceilometer-notification-agent" Jan 26 17:09:48 crc kubenswrapper[4754]: E0126 17:09:48.083497 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b63306d0-0c02-4691-81c7-60846f3c3fdc" containerName="proxy-httpd" Jan 26 17:09:48 crc kubenswrapper[4754]: I0126 17:09:48.083504 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="b63306d0-0c02-4691-81c7-60846f3c3fdc" containerName="proxy-httpd" Jan 26 17:09:48 crc kubenswrapper[4754]: E0126 17:09:48.083519 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b63306d0-0c02-4691-81c7-60846f3c3fdc" containerName="ceilometer-central-agent" Jan 26 17:09:48 crc kubenswrapper[4754]: I0126 17:09:48.083524 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="b63306d0-0c02-4691-81c7-60846f3c3fdc" containerName="ceilometer-central-agent" Jan 26 17:09:48 crc kubenswrapper[4754]: E0126 17:09:48.083538 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b63306d0-0c02-4691-81c7-60846f3c3fdc" containerName="sg-core" Jan 26 17:09:48 crc kubenswrapper[4754]: I0126 17:09:48.083544 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="b63306d0-0c02-4691-81c7-60846f3c3fdc" containerName="sg-core" Jan 26 17:09:48 crc kubenswrapper[4754]: I0126 17:09:48.083960 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="b63306d0-0c02-4691-81c7-60846f3c3fdc" containerName="proxy-httpd" Jan 26 17:09:48 crc kubenswrapper[4754]: I0126 17:09:48.083993 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="b63306d0-0c02-4691-81c7-60846f3c3fdc" containerName="ceilometer-notification-agent" Jan 26 17:09:48 crc kubenswrapper[4754]: I0126 17:09:48.084035 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="b63306d0-0c02-4691-81c7-60846f3c3fdc" containerName="sg-core" Jan 26 17:09:48 crc kubenswrapper[4754]: I0126 17:09:48.084048 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="b63306d0-0c02-4691-81c7-60846f3c3fdc" containerName="ceilometer-central-agent" Jan 26 17:09:48 crc kubenswrapper[4754]: I0126 17:09:48.085951 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 26 17:09:48 crc kubenswrapper[4754]: I0126 17:09:48.089852 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Jan 26 17:09:48 crc kubenswrapper[4754]: I0126 17:09:48.090115 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Jan 26 17:09:48 crc kubenswrapper[4754]: I0126 17:09:48.097892 4754 scope.go:117] "RemoveContainer" containerID="6dbb6881c0cdc0f402b9772f369d7222270b21d8425d63e6c42a80db5944dcde" Jan 26 17:09:48 crc kubenswrapper[4754]: I0126 17:09:48.099840 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Jan 26 17:09:48 crc kubenswrapper[4754]: I0126 17:09:48.170384 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ee49bc9b-edbd-4b7b-b977-98ec0b6b21c4-scripts\") pod \"ceilometer-0\" (UID: \"ee49bc9b-edbd-4b7b-b977-98ec0b6b21c4\") " pod="openstack/ceilometer-0" Jan 26 17:09:48 crc kubenswrapper[4754]: I0126 17:09:48.170523 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee49bc9b-edbd-4b7b-b977-98ec0b6b21c4-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ee49bc9b-edbd-4b7b-b977-98ec0b6b21c4\") " pod="openstack/ceilometer-0" Jan 26 17:09:48 crc kubenswrapper[4754]: I0126 17:09:48.170547 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ee49bc9b-edbd-4b7b-b977-98ec0b6b21c4-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ee49bc9b-edbd-4b7b-b977-98ec0b6b21c4\") " pod="openstack/ceilometer-0" Jan 26 17:09:48 crc kubenswrapper[4754]: I0126 17:09:48.170570 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xl995\" (UniqueName: \"kubernetes.io/projected/ee49bc9b-edbd-4b7b-b977-98ec0b6b21c4-kube-api-access-xl995\") pod \"ceilometer-0\" (UID: \"ee49bc9b-edbd-4b7b-b977-98ec0b6b21c4\") " pod="openstack/ceilometer-0" Jan 26 17:09:48 crc kubenswrapper[4754]: I0126 17:09:48.170590 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ee49bc9b-edbd-4b7b-b977-98ec0b6b21c4-config-data\") pod \"ceilometer-0\" (UID: \"ee49bc9b-edbd-4b7b-b977-98ec0b6b21c4\") " pod="openstack/ceilometer-0" Jan 26 17:09:48 crc kubenswrapper[4754]: I0126 17:09:48.170829 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ee49bc9b-edbd-4b7b-b977-98ec0b6b21c4-log-httpd\") pod \"ceilometer-0\" (UID: \"ee49bc9b-edbd-4b7b-b977-98ec0b6b21c4\") " pod="openstack/ceilometer-0" Jan 26 17:09:48 crc kubenswrapper[4754]: I0126 17:09:48.170988 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ee49bc9b-edbd-4b7b-b977-98ec0b6b21c4-run-httpd\") pod \"ceilometer-0\" (UID: \"ee49bc9b-edbd-4b7b-b977-98ec0b6b21c4\") " pod="openstack/ceilometer-0" Jan 26 17:09:48 crc kubenswrapper[4754]: I0126 17:09:48.272830 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ee49bc9b-edbd-4b7b-b977-98ec0b6b21c4-run-httpd\") pod \"ceilometer-0\" (UID: \"ee49bc9b-edbd-4b7b-b977-98ec0b6b21c4\") " pod="openstack/ceilometer-0" Jan 26 17:09:48 crc kubenswrapper[4754]: I0126 17:09:48.273206 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ee49bc9b-edbd-4b7b-b977-98ec0b6b21c4-scripts\") pod \"ceilometer-0\" (UID: \"ee49bc9b-edbd-4b7b-b977-98ec0b6b21c4\") " pod="openstack/ceilometer-0" Jan 26 17:09:48 crc kubenswrapper[4754]: I0126 17:09:48.273274 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ee49bc9b-edbd-4b7b-b977-98ec0b6b21c4-run-httpd\") pod \"ceilometer-0\" (UID: \"ee49bc9b-edbd-4b7b-b977-98ec0b6b21c4\") " pod="openstack/ceilometer-0" Jan 26 17:09:48 crc kubenswrapper[4754]: I0126 17:09:48.273287 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee49bc9b-edbd-4b7b-b977-98ec0b6b21c4-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ee49bc9b-edbd-4b7b-b977-98ec0b6b21c4\") " pod="openstack/ceilometer-0" Jan 26 17:09:48 crc kubenswrapper[4754]: I0126 17:09:48.273311 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ee49bc9b-edbd-4b7b-b977-98ec0b6b21c4-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ee49bc9b-edbd-4b7b-b977-98ec0b6b21c4\") " pod="openstack/ceilometer-0" Jan 26 17:09:48 crc kubenswrapper[4754]: I0126 17:09:48.273335 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xl995\" (UniqueName: \"kubernetes.io/projected/ee49bc9b-edbd-4b7b-b977-98ec0b6b21c4-kube-api-access-xl995\") pod \"ceilometer-0\" (UID: \"ee49bc9b-edbd-4b7b-b977-98ec0b6b21c4\") " pod="openstack/ceilometer-0" Jan 26 17:09:48 crc kubenswrapper[4754]: I0126 17:09:48.273364 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ee49bc9b-edbd-4b7b-b977-98ec0b6b21c4-config-data\") pod \"ceilometer-0\" (UID: \"ee49bc9b-edbd-4b7b-b977-98ec0b6b21c4\") " pod="openstack/ceilometer-0" Jan 26 17:09:48 crc kubenswrapper[4754]: I0126 17:09:48.273420 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ee49bc9b-edbd-4b7b-b977-98ec0b6b21c4-log-httpd\") pod \"ceilometer-0\" (UID: \"ee49bc9b-edbd-4b7b-b977-98ec0b6b21c4\") " pod="openstack/ceilometer-0" Jan 26 17:09:48 crc kubenswrapper[4754]: I0126 17:09:48.273932 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ee49bc9b-edbd-4b7b-b977-98ec0b6b21c4-log-httpd\") pod \"ceilometer-0\" (UID: \"ee49bc9b-edbd-4b7b-b977-98ec0b6b21c4\") " pod="openstack/ceilometer-0" Jan 26 17:09:48 crc kubenswrapper[4754]: I0126 17:09:48.285189 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ee49bc9b-edbd-4b7b-b977-98ec0b6b21c4-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ee49bc9b-edbd-4b7b-b977-98ec0b6b21c4\") " pod="openstack/ceilometer-0" Jan 26 17:09:48 crc kubenswrapper[4754]: I0126 17:09:48.285507 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee49bc9b-edbd-4b7b-b977-98ec0b6b21c4-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ee49bc9b-edbd-4b7b-b977-98ec0b6b21c4\") " pod="openstack/ceilometer-0" Jan 26 17:09:48 crc kubenswrapper[4754]: I0126 17:09:48.286085 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ee49bc9b-edbd-4b7b-b977-98ec0b6b21c4-config-data\") pod \"ceilometer-0\" (UID: \"ee49bc9b-edbd-4b7b-b977-98ec0b6b21c4\") " pod="openstack/ceilometer-0" Jan 26 17:09:48 crc kubenswrapper[4754]: I0126 17:09:48.286103 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ee49bc9b-edbd-4b7b-b977-98ec0b6b21c4-scripts\") pod \"ceilometer-0\" (UID: \"ee49bc9b-edbd-4b7b-b977-98ec0b6b21c4\") " pod="openstack/ceilometer-0" Jan 26 17:09:48 crc kubenswrapper[4754]: I0126 17:09:48.288858 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xl995\" (UniqueName: \"kubernetes.io/projected/ee49bc9b-edbd-4b7b-b977-98ec0b6b21c4-kube-api-access-xl995\") pod \"ceilometer-0\" (UID: \"ee49bc9b-edbd-4b7b-b977-98ec0b6b21c4\") " pod="openstack/ceilometer-0" Jan 26 17:09:48 crc kubenswrapper[4754]: I0126 17:09:48.418962 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 26 17:09:48 crc kubenswrapper[4754]: W0126 17:09:48.945926 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podee49bc9b_edbd_4b7b_b977_98ec0b6b21c4.slice/crio-a1e64314e99fdb6278c7c0f9380f49df9560203140d09ed1fa445b0413789260 WatchSource:0}: Error finding container a1e64314e99fdb6278c7c0f9380f49df9560203140d09ed1fa445b0413789260: Status 404 returned error can't find the container with id a1e64314e99fdb6278c7c0f9380f49df9560203140d09ed1fa445b0413789260 Jan 26 17:09:48 crc kubenswrapper[4754]: I0126 17:09:48.946074 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Jan 26 17:09:48 crc kubenswrapper[4754]: I0126 17:09:48.979343 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ee49bc9b-edbd-4b7b-b977-98ec0b6b21c4","Type":"ContainerStarted","Data":"a1e64314e99fdb6278c7c0f9380f49df9560203140d09ed1fa445b0413789260"} Jan 26 17:09:48 crc kubenswrapper[4754]: I0126 17:09:48.983754 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-767b9f4849-cfqbz" event={"ID":"36eccc33-604a-49ac-a1f2-c0bd41fd053d","Type":"ContainerStarted","Data":"cbebed1fcb325093ea9238bbec116e41e1a9dd7b7e7af28565ec2ec43aeb8ac9"} Jan 26 17:09:49 crc kubenswrapper[4754]: I0126 17:09:49.778986 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b63306d0-0c02-4691-81c7-60846f3c3fdc" path="/var/lib/kubelet/pods/b63306d0-0c02-4691-81c7-60846f3c3fdc/volumes" Jan 26 17:09:49 crc kubenswrapper[4754]: I0126 17:09:49.990223 4754 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Jan 26 17:09:49 crc kubenswrapper[4754]: I0126 17:09:49.990246 4754 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Jan 26 17:09:50 crc kubenswrapper[4754]: I0126 17:09:50.148156 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Jan 26 17:09:50 crc kubenswrapper[4754]: I0126 17:09:50.182774 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Jan 26 17:09:50 crc kubenswrapper[4754]: I0126 17:09:50.183699 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Jan 26 17:09:50 crc kubenswrapper[4754]: I0126 17:09:50.366867 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-rpnm5"] Jan 26 17:09:50 crc kubenswrapper[4754]: I0126 17:09:50.369117 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rpnm5" Jan 26 17:09:50 crc kubenswrapper[4754]: I0126 17:09:50.385716 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rpnm5"] Jan 26 17:09:50 crc kubenswrapper[4754]: I0126 17:09:50.419161 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Jan 26 17:09:50 crc kubenswrapper[4754]: I0126 17:09:50.531296 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7dc29a9d-f578-49ec-97a8-08a80bd3a2c0-catalog-content\") pod \"certified-operators-rpnm5\" (UID: \"7dc29a9d-f578-49ec-97a8-08a80bd3a2c0\") " pod="openshift-marketplace/certified-operators-rpnm5" Jan 26 17:09:50 crc kubenswrapper[4754]: I0126 17:09:50.531697 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7dc29a9d-f578-49ec-97a8-08a80bd3a2c0-utilities\") pod \"certified-operators-rpnm5\" (UID: \"7dc29a9d-f578-49ec-97a8-08a80bd3a2c0\") " pod="openshift-marketplace/certified-operators-rpnm5" Jan 26 17:09:50 crc kubenswrapper[4754]: I0126 17:09:50.531758 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pqvjk\" (UniqueName: \"kubernetes.io/projected/7dc29a9d-f578-49ec-97a8-08a80bd3a2c0-kube-api-access-pqvjk\") pod \"certified-operators-rpnm5\" (UID: \"7dc29a9d-f578-49ec-97a8-08a80bd3a2c0\") " pod="openshift-marketplace/certified-operators-rpnm5" Jan 26 17:09:50 crc kubenswrapper[4754]: I0126 17:09:50.633335 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7dc29a9d-f578-49ec-97a8-08a80bd3a2c0-utilities\") pod \"certified-operators-rpnm5\" (UID: \"7dc29a9d-f578-49ec-97a8-08a80bd3a2c0\") " pod="openshift-marketplace/certified-operators-rpnm5" Jan 26 17:09:50 crc kubenswrapper[4754]: I0126 17:09:50.633427 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pqvjk\" (UniqueName: \"kubernetes.io/projected/7dc29a9d-f578-49ec-97a8-08a80bd3a2c0-kube-api-access-pqvjk\") pod \"certified-operators-rpnm5\" (UID: \"7dc29a9d-f578-49ec-97a8-08a80bd3a2c0\") " pod="openshift-marketplace/certified-operators-rpnm5" Jan 26 17:09:50 crc kubenswrapper[4754]: I0126 17:09:50.633505 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7dc29a9d-f578-49ec-97a8-08a80bd3a2c0-catalog-content\") pod \"certified-operators-rpnm5\" (UID: \"7dc29a9d-f578-49ec-97a8-08a80bd3a2c0\") " pod="openshift-marketplace/certified-operators-rpnm5" Jan 26 17:09:50 crc kubenswrapper[4754]: I0126 17:09:50.633943 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7dc29a9d-f578-49ec-97a8-08a80bd3a2c0-utilities\") pod \"certified-operators-rpnm5\" (UID: \"7dc29a9d-f578-49ec-97a8-08a80bd3a2c0\") " pod="openshift-marketplace/certified-operators-rpnm5" Jan 26 17:09:50 crc kubenswrapper[4754]: I0126 17:09:50.633990 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7dc29a9d-f578-49ec-97a8-08a80bd3a2c0-catalog-content\") pod \"certified-operators-rpnm5\" (UID: \"7dc29a9d-f578-49ec-97a8-08a80bd3a2c0\") " pod="openshift-marketplace/certified-operators-rpnm5" Jan 26 17:09:50 crc kubenswrapper[4754]: I0126 17:09:50.657837 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pqvjk\" (UniqueName: \"kubernetes.io/projected/7dc29a9d-f578-49ec-97a8-08a80bd3a2c0-kube-api-access-pqvjk\") pod \"certified-operators-rpnm5\" (UID: \"7dc29a9d-f578-49ec-97a8-08a80bd3a2c0\") " pod="openshift-marketplace/certified-operators-rpnm5" Jan 26 17:09:50 crc kubenswrapper[4754]: I0126 17:09:50.698075 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rpnm5" Jan 26 17:09:51 crc kubenswrapper[4754]: I0126 17:09:51.001046 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ee49bc9b-edbd-4b7b-b977-98ec0b6b21c4","Type":"ContainerStarted","Data":"e350d7d898f163921729bb47f1b2fd06ba49d5a07efcd6a4f1fa7505a6a47da7"} Jan 26 17:09:51 crc kubenswrapper[4754]: I0126 17:09:51.018873 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-767b9f4849-cfqbz" event={"ID":"36eccc33-604a-49ac-a1f2-c0bd41fd053d","Type":"ContainerStarted","Data":"6db296d1f5e065022637d66edb555e565173bd34029dc8257cac73bb20b8a3ba"} Jan 26 17:09:51 crc kubenswrapper[4754]: I0126 17:09:51.018910 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-767b9f4849-cfqbz" Jan 26 17:09:51 crc kubenswrapper[4754]: I0126 17:09:51.043221 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-767b9f4849-cfqbz" podStartSLOduration=5.043200542 podStartE2EDuration="5.043200542s" podCreationTimestamp="2026-01-26 17:09:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 17:09:51.039194738 +0000 UTC m=+1357.563375172" watchObservedRunningTime="2026-01-26 17:09:51.043200542 +0000 UTC m=+1357.567380996" Jan 26 17:09:51 crc kubenswrapper[4754]: I0126 17:09:51.257550 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rpnm5"] Jan 26 17:09:51 crc kubenswrapper[4754]: W0126 17:09:51.261631 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7dc29a9d_f578_49ec_97a8_08a80bd3a2c0.slice/crio-591c15aa136778dda5707b6d79f17694f62bd03a891f83026afe4b5113edd8f1 WatchSource:0}: Error finding container 591c15aa136778dda5707b6d79f17694f62bd03a891f83026afe4b5113edd8f1: Status 404 returned error can't find the container with id 591c15aa136778dda5707b6d79f17694f62bd03a891f83026afe4b5113edd8f1 Jan 26 17:09:52 crc kubenswrapper[4754]: I0126 17:09:52.026845 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rpnm5" event={"ID":"7dc29a9d-f578-49ec-97a8-08a80bd3a2c0","Type":"ContainerStarted","Data":"591c15aa136778dda5707b6d79f17694f62bd03a891f83026afe4b5113edd8f1"} Jan 26 17:09:53 crc kubenswrapper[4754]: I0126 17:09:53.035763 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ee49bc9b-edbd-4b7b-b977-98ec0b6b21c4","Type":"ContainerStarted","Data":"09cad47c4a2f836bde7a8e847c73a4882a25b0a8bab6d861ac184fbd95f15505"} Jan 26 17:09:53 crc kubenswrapper[4754]: I0126 17:09:53.037484 4754 generic.go:334] "Generic (PLEG): container finished" podID="7dc29a9d-f578-49ec-97a8-08a80bd3a2c0" containerID="269c5e8cc3d5b27f5d220eb92d7616cc38a520f9c0c5a6c0d4f859c1a6c8a16d" exitCode=0 Jan 26 17:09:53 crc kubenswrapper[4754]: I0126 17:09:53.037563 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rpnm5" event={"ID":"7dc29a9d-f578-49ec-97a8-08a80bd3a2c0","Type":"ContainerDied","Data":"269c5e8cc3d5b27f5d220eb92d7616cc38a520f9c0c5a6c0d4f859c1a6c8a16d"} Jan 26 17:09:53 crc kubenswrapper[4754]: I0126 17:09:53.217173 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Jan 26 17:09:54 crc kubenswrapper[4754]: I0126 17:09:54.049854 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rpnm5" event={"ID":"7dc29a9d-f578-49ec-97a8-08a80bd3a2c0","Type":"ContainerStarted","Data":"e7d747ff339856686f9c794f4314c3a1d4356cd1668664a52a35543107d47c3e"} Jan 26 17:09:54 crc kubenswrapper[4754]: I0126 17:09:54.053295 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ee49bc9b-edbd-4b7b-b977-98ec0b6b21c4","Type":"ContainerStarted","Data":"d87184c817a1540d791103f965b1f40e2b84afbbfe054dc48e127f244e9911be"} Jan 26 17:09:54 crc kubenswrapper[4754]: I0126 17:09:54.655925 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5c9776ccc5-tnfmz" Jan 26 17:09:54 crc kubenswrapper[4754]: I0126 17:09:54.731708 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-795f4db4bc-h7mcp"] Jan 26 17:09:54 crc kubenswrapper[4754]: I0126 17:09:54.732136 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-795f4db4bc-h7mcp" podUID="977ff39d-2c55-438b-b957-62609c4c2ece" containerName="dnsmasq-dns" containerID="cri-o://9a6add83bff5d8df998c3ea59d3a4c3be4deffcc6f345db9dfe249b6f0c866f3" gracePeriod=10 Jan 26 17:09:55 crc kubenswrapper[4754]: I0126 17:09:55.069093 4754 generic.go:334] "Generic (PLEG): container finished" podID="977ff39d-2c55-438b-b957-62609c4c2ece" containerID="9a6add83bff5d8df998c3ea59d3a4c3be4deffcc6f345db9dfe249b6f0c866f3" exitCode=0 Jan 26 17:09:55 crc kubenswrapper[4754]: I0126 17:09:55.069302 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-795f4db4bc-h7mcp" event={"ID":"977ff39d-2c55-438b-b957-62609c4c2ece","Type":"ContainerDied","Data":"9a6add83bff5d8df998c3ea59d3a4c3be4deffcc6f345db9dfe249b6f0c866f3"} Jan 26 17:09:55 crc kubenswrapper[4754]: I0126 17:09:55.073456 4754 generic.go:334] "Generic (PLEG): container finished" podID="7dc29a9d-f578-49ec-97a8-08a80bd3a2c0" containerID="e7d747ff339856686f9c794f4314c3a1d4356cd1668664a52a35543107d47c3e" exitCode=0 Jan 26 17:09:55 crc kubenswrapper[4754]: I0126 17:09:55.073524 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rpnm5" event={"ID":"7dc29a9d-f578-49ec-97a8-08a80bd3a2c0","Type":"ContainerDied","Data":"e7d747ff339856686f9c794f4314c3a1d4356cd1668664a52a35543107d47c3e"} Jan 26 17:09:55 crc kubenswrapper[4754]: I0126 17:09:55.076960 4754 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 26 17:09:55 crc kubenswrapper[4754]: I0126 17:09:55.270532 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-795f4db4bc-h7mcp" Jan 26 17:09:55 crc kubenswrapper[4754]: I0126 17:09:55.334278 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/977ff39d-2c55-438b-b957-62609c4c2ece-ovsdbserver-sb\") pod \"977ff39d-2c55-438b-b957-62609c4c2ece\" (UID: \"977ff39d-2c55-438b-b957-62609c4c2ece\") " Jan 26 17:09:55 crc kubenswrapper[4754]: I0126 17:09:55.334360 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/977ff39d-2c55-438b-b957-62609c4c2ece-config\") pod \"977ff39d-2c55-438b-b957-62609c4c2ece\" (UID: \"977ff39d-2c55-438b-b957-62609c4c2ece\") " Jan 26 17:09:55 crc kubenswrapper[4754]: I0126 17:09:55.334452 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wndl7\" (UniqueName: \"kubernetes.io/projected/977ff39d-2c55-438b-b957-62609c4c2ece-kube-api-access-wndl7\") pod \"977ff39d-2c55-438b-b957-62609c4c2ece\" (UID: \"977ff39d-2c55-438b-b957-62609c4c2ece\") " Jan 26 17:09:55 crc kubenswrapper[4754]: I0126 17:09:55.334491 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/977ff39d-2c55-438b-b957-62609c4c2ece-dns-svc\") pod \"977ff39d-2c55-438b-b957-62609c4c2ece\" (UID: \"977ff39d-2c55-438b-b957-62609c4c2ece\") " Jan 26 17:09:55 crc kubenswrapper[4754]: I0126 17:09:55.334527 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/977ff39d-2c55-438b-b957-62609c4c2ece-ovsdbserver-nb\") pod \"977ff39d-2c55-438b-b957-62609c4c2ece\" (UID: \"977ff39d-2c55-438b-b957-62609c4c2ece\") " Jan 26 17:09:55 crc kubenswrapper[4754]: I0126 17:09:55.334570 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/977ff39d-2c55-438b-b957-62609c4c2ece-dns-swift-storage-0\") pod \"977ff39d-2c55-438b-b957-62609c4c2ece\" (UID: \"977ff39d-2c55-438b-b957-62609c4c2ece\") " Jan 26 17:09:55 crc kubenswrapper[4754]: I0126 17:09:55.342273 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/977ff39d-2c55-438b-b957-62609c4c2ece-kube-api-access-wndl7" (OuterVolumeSpecName: "kube-api-access-wndl7") pod "977ff39d-2c55-438b-b957-62609c4c2ece" (UID: "977ff39d-2c55-438b-b957-62609c4c2ece"). InnerVolumeSpecName "kube-api-access-wndl7". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:09:55 crc kubenswrapper[4754]: I0126 17:09:55.399532 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/977ff39d-2c55-438b-b957-62609c4c2ece-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "977ff39d-2c55-438b-b957-62609c4c2ece" (UID: "977ff39d-2c55-438b-b957-62609c4c2ece"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:09:55 crc kubenswrapper[4754]: I0126 17:09:55.399875 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/977ff39d-2c55-438b-b957-62609c4c2ece-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "977ff39d-2c55-438b-b957-62609c4c2ece" (UID: "977ff39d-2c55-438b-b957-62609c4c2ece"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:09:55 crc kubenswrapper[4754]: I0126 17:09:55.405038 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/977ff39d-2c55-438b-b957-62609c4c2ece-config" (OuterVolumeSpecName: "config") pod "977ff39d-2c55-438b-b957-62609c4c2ece" (UID: "977ff39d-2c55-438b-b957-62609c4c2ece"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:09:55 crc kubenswrapper[4754]: I0126 17:09:55.416252 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/977ff39d-2c55-438b-b957-62609c4c2ece-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "977ff39d-2c55-438b-b957-62609c4c2ece" (UID: "977ff39d-2c55-438b-b957-62609c4c2ece"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:09:55 crc kubenswrapper[4754]: I0126 17:09:55.420206 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/977ff39d-2c55-438b-b957-62609c4c2ece-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "977ff39d-2c55-438b-b957-62609c4c2ece" (UID: "977ff39d-2c55-438b-b957-62609c4c2ece"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:09:55 crc kubenswrapper[4754]: I0126 17:09:55.449109 4754 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/977ff39d-2c55-438b-b957-62609c4c2ece-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Jan 26 17:09:55 crc kubenswrapper[4754]: I0126 17:09:55.449175 4754 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/977ff39d-2c55-438b-b957-62609c4c2ece-config\") on node \"crc\" DevicePath \"\"" Jan 26 17:09:55 crc kubenswrapper[4754]: I0126 17:09:55.449190 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wndl7\" (UniqueName: \"kubernetes.io/projected/977ff39d-2c55-438b-b957-62609c4c2ece-kube-api-access-wndl7\") on node \"crc\" DevicePath \"\"" Jan 26 17:09:55 crc kubenswrapper[4754]: I0126 17:09:55.449203 4754 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/977ff39d-2c55-438b-b957-62609c4c2ece-dns-svc\") on node \"crc\" DevicePath \"\"" Jan 26 17:09:55 crc kubenswrapper[4754]: I0126 17:09:55.449214 4754 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/977ff39d-2c55-438b-b957-62609c4c2ece-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Jan 26 17:09:55 crc kubenswrapper[4754]: I0126 17:09:55.449228 4754 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/977ff39d-2c55-438b-b957-62609c4c2ece-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Jan 26 17:09:56 crc kubenswrapper[4754]: I0126 17:09:56.084121 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ee49bc9b-edbd-4b7b-b977-98ec0b6b21c4","Type":"ContainerStarted","Data":"dd10cb6e3eb2254b305e23513953ba8c1131ba1298169e69cdab441ae8a2ecc8"} Jan 26 17:09:56 crc kubenswrapper[4754]: I0126 17:09:56.084614 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ee49bc9b-edbd-4b7b-b977-98ec0b6b21c4" containerName="ceilometer-central-agent" containerID="cri-o://e350d7d898f163921729bb47f1b2fd06ba49d5a07efcd6a4f1fa7505a6a47da7" gracePeriod=30 Jan 26 17:09:56 crc kubenswrapper[4754]: I0126 17:09:56.084687 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Jan 26 17:09:56 crc kubenswrapper[4754]: I0126 17:09:56.084729 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ee49bc9b-edbd-4b7b-b977-98ec0b6b21c4" containerName="proxy-httpd" containerID="cri-o://dd10cb6e3eb2254b305e23513953ba8c1131ba1298169e69cdab441ae8a2ecc8" gracePeriod=30 Jan 26 17:09:56 crc kubenswrapper[4754]: I0126 17:09:56.084779 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ee49bc9b-edbd-4b7b-b977-98ec0b6b21c4" containerName="sg-core" containerID="cri-o://d87184c817a1540d791103f965b1f40e2b84afbbfe054dc48e127f244e9911be" gracePeriod=30 Jan 26 17:09:56 crc kubenswrapper[4754]: I0126 17:09:56.084822 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ee49bc9b-edbd-4b7b-b977-98ec0b6b21c4" containerName="ceilometer-notification-agent" containerID="cri-o://09cad47c4a2f836bde7a8e847c73a4882a25b0a8bab6d861ac184fbd95f15505" gracePeriod=30 Jan 26 17:09:56 crc kubenswrapper[4754]: I0126 17:09:56.088697 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-795f4db4bc-h7mcp" event={"ID":"977ff39d-2c55-438b-b957-62609c4c2ece","Type":"ContainerDied","Data":"bf3e959a5f40da261aa7cf7b87b3564bafac97e7808c4387dbbb133f67fc522f"} Jan 26 17:09:56 crc kubenswrapper[4754]: I0126 17:09:56.088748 4754 scope.go:117] "RemoveContainer" containerID="9a6add83bff5d8df998c3ea59d3a4c3be4deffcc6f345db9dfe249b6f0c866f3" Jan 26 17:09:56 crc kubenswrapper[4754]: I0126 17:09:56.088904 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-795f4db4bc-h7mcp" Jan 26 17:09:56 crc kubenswrapper[4754]: I0126 17:09:56.122393 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.220738759 podStartE2EDuration="8.122373535s" podCreationTimestamp="2026-01-26 17:09:48 +0000 UTC" firstStartedPulling="2026-01-26 17:09:48.947598702 +0000 UTC m=+1355.471779136" lastFinishedPulling="2026-01-26 17:09:55.849233478 +0000 UTC m=+1362.373413912" observedRunningTime="2026-01-26 17:09:56.114773416 +0000 UTC m=+1362.638953870" watchObservedRunningTime="2026-01-26 17:09:56.122373535 +0000 UTC m=+1362.646553969" Jan 26 17:09:56 crc kubenswrapper[4754]: I0126 17:09:56.134103 4754 scope.go:117] "RemoveContainer" containerID="38b3ccb1c0bbed7e3bde1802c479d558b8daa522cae6af0fb1c2ed14fdc80d28" Jan 26 17:09:56 crc kubenswrapper[4754]: I0126 17:09:56.147793 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-795f4db4bc-h7mcp"] Jan 26 17:09:56 crc kubenswrapper[4754]: I0126 17:09:56.170062 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-795f4db4bc-h7mcp"] Jan 26 17:09:57 crc kubenswrapper[4754]: I0126 17:09:57.098904 4754 generic.go:334] "Generic (PLEG): container finished" podID="ee49bc9b-edbd-4b7b-b977-98ec0b6b21c4" containerID="d87184c817a1540d791103f965b1f40e2b84afbbfe054dc48e127f244e9911be" exitCode=2 Jan 26 17:09:57 crc kubenswrapper[4754]: I0126 17:09:57.099240 4754 generic.go:334] "Generic (PLEG): container finished" podID="ee49bc9b-edbd-4b7b-b977-98ec0b6b21c4" containerID="09cad47c4a2f836bde7a8e847c73a4882a25b0a8bab6d861ac184fbd95f15505" exitCode=0 Jan 26 17:09:57 crc kubenswrapper[4754]: I0126 17:09:57.098974 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ee49bc9b-edbd-4b7b-b977-98ec0b6b21c4","Type":"ContainerDied","Data":"d87184c817a1540d791103f965b1f40e2b84afbbfe054dc48e127f244e9911be"} Jan 26 17:09:57 crc kubenswrapper[4754]: I0126 17:09:57.099304 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ee49bc9b-edbd-4b7b-b977-98ec0b6b21c4","Type":"ContainerDied","Data":"09cad47c4a2f836bde7a8e847c73a4882a25b0a8bab6d861ac184fbd95f15505"} Jan 26 17:09:57 crc kubenswrapper[4754]: I0126 17:09:57.102658 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rpnm5" event={"ID":"7dc29a9d-f578-49ec-97a8-08a80bd3a2c0","Type":"ContainerStarted","Data":"510e54a25e15cd76ffea612e300eea6a77041185907b8c15443c0adbf7ee8b78"} Jan 26 17:09:57 crc kubenswrapper[4754]: I0126 17:09:57.125271 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-rpnm5" podStartSLOduration=4.199083851 podStartE2EDuration="7.125255133s" podCreationTimestamp="2026-01-26 17:09:50 +0000 UTC" firstStartedPulling="2026-01-26 17:09:53.039296578 +0000 UTC m=+1359.563477012" lastFinishedPulling="2026-01-26 17:09:55.96546786 +0000 UTC m=+1362.489648294" observedRunningTime="2026-01-26 17:09:57.121977593 +0000 UTC m=+1363.646158047" watchObservedRunningTime="2026-01-26 17:09:57.125255133 +0000 UTC m=+1363.649435567" Jan 26 17:09:57 crc kubenswrapper[4754]: I0126 17:09:57.370063 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-j7p2d"] Jan 26 17:09:57 crc kubenswrapper[4754]: E0126 17:09:57.379069 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="977ff39d-2c55-438b-b957-62609c4c2ece" containerName="init" Jan 26 17:09:57 crc kubenswrapper[4754]: I0126 17:09:57.379115 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="977ff39d-2c55-438b-b957-62609c4c2ece" containerName="init" Jan 26 17:09:57 crc kubenswrapper[4754]: E0126 17:09:57.379175 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="977ff39d-2c55-438b-b957-62609c4c2ece" containerName="dnsmasq-dns" Jan 26 17:09:57 crc kubenswrapper[4754]: I0126 17:09:57.379185 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="977ff39d-2c55-438b-b957-62609c4c2ece" containerName="dnsmasq-dns" Jan 26 17:09:57 crc kubenswrapper[4754]: I0126 17:09:57.379472 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="977ff39d-2c55-438b-b957-62609c4c2ece" containerName="dnsmasq-dns" Jan 26 17:09:57 crc kubenswrapper[4754]: I0126 17:09:57.380286 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-j7p2d" Jan 26 17:09:57 crc kubenswrapper[4754]: I0126 17:09:57.394599 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-j7p2d"] Jan 26 17:09:57 crc kubenswrapper[4754]: I0126 17:09:57.461591 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-6qdbg"] Jan 26 17:09:57 crc kubenswrapper[4754]: I0126 17:09:57.462789 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-6qdbg" Jan 26 17:09:57 crc kubenswrapper[4754]: I0126 17:09:57.472385 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-6qdbg"] Jan 26 17:09:57 crc kubenswrapper[4754]: I0126 17:09:57.484733 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-e6d7-account-create-update-hs5pg"] Jan 26 17:09:57 crc kubenswrapper[4754]: I0126 17:09:57.485116 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cf071f6e-289f-4f86-9a39-e4dc62335b14-operator-scripts\") pod \"nova-api-db-create-j7p2d\" (UID: \"cf071f6e-289f-4f86-9a39-e4dc62335b14\") " pod="openstack/nova-api-db-create-j7p2d" Jan 26 17:09:57 crc kubenswrapper[4754]: I0126 17:09:57.485195 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hmdrj\" (UniqueName: \"kubernetes.io/projected/cf071f6e-289f-4f86-9a39-e4dc62335b14-kube-api-access-hmdrj\") pod \"nova-api-db-create-j7p2d\" (UID: \"cf071f6e-289f-4f86-9a39-e4dc62335b14\") " pod="openstack/nova-api-db-create-j7p2d" Jan 26 17:09:57 crc kubenswrapper[4754]: I0126 17:09:57.485874 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-e6d7-account-create-update-hs5pg" Jan 26 17:09:57 crc kubenswrapper[4754]: I0126 17:09:57.488125 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Jan 26 17:09:57 crc kubenswrapper[4754]: I0126 17:09:57.510590 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-e6d7-account-create-update-hs5pg"] Jan 26 17:09:57 crc kubenswrapper[4754]: I0126 17:09:57.571302 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-bjmmr"] Jan 26 17:09:57 crc kubenswrapper[4754]: I0126 17:09:57.572495 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-bjmmr" Jan 26 17:09:57 crc kubenswrapper[4754]: I0126 17:09:57.587332 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wmx9f\" (UniqueName: \"kubernetes.io/projected/f5c367c8-74fd-4073-bbc6-2f6d4373f5ef-kube-api-access-wmx9f\") pod \"nova-cell0-db-create-6qdbg\" (UID: \"f5c367c8-74fd-4073-bbc6-2f6d4373f5ef\") " pod="openstack/nova-cell0-db-create-6qdbg" Jan 26 17:09:57 crc kubenswrapper[4754]: I0126 17:09:57.587705 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1597209c-abf0-4bd4-91b4-a8661e7e496b-operator-scripts\") pod \"nova-api-e6d7-account-create-update-hs5pg\" (UID: \"1597209c-abf0-4bd4-91b4-a8661e7e496b\") " pod="openstack/nova-api-e6d7-account-create-update-hs5pg" Jan 26 17:09:57 crc kubenswrapper[4754]: I0126 17:09:57.587861 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cf071f6e-289f-4f86-9a39-e4dc62335b14-operator-scripts\") pod \"nova-api-db-create-j7p2d\" (UID: \"cf071f6e-289f-4f86-9a39-e4dc62335b14\") " pod="openstack/nova-api-db-create-j7p2d" Jan 26 17:09:57 crc kubenswrapper[4754]: I0126 17:09:57.587960 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f5c367c8-74fd-4073-bbc6-2f6d4373f5ef-operator-scripts\") pod \"nova-cell0-db-create-6qdbg\" (UID: \"f5c367c8-74fd-4073-bbc6-2f6d4373f5ef\") " pod="openstack/nova-cell0-db-create-6qdbg" Jan 26 17:09:57 crc kubenswrapper[4754]: I0126 17:09:57.588014 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hmdrj\" (UniqueName: \"kubernetes.io/projected/cf071f6e-289f-4f86-9a39-e4dc62335b14-kube-api-access-hmdrj\") pod \"nova-api-db-create-j7p2d\" (UID: \"cf071f6e-289f-4f86-9a39-e4dc62335b14\") " pod="openstack/nova-api-db-create-j7p2d" Jan 26 17:09:57 crc kubenswrapper[4754]: I0126 17:09:57.588045 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j5hr4\" (UniqueName: \"kubernetes.io/projected/1597209c-abf0-4bd4-91b4-a8661e7e496b-kube-api-access-j5hr4\") pod \"nova-api-e6d7-account-create-update-hs5pg\" (UID: \"1597209c-abf0-4bd4-91b4-a8661e7e496b\") " pod="openstack/nova-api-e6d7-account-create-update-hs5pg" Jan 26 17:09:57 crc kubenswrapper[4754]: I0126 17:09:57.589464 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cf071f6e-289f-4f86-9a39-e4dc62335b14-operator-scripts\") pod \"nova-api-db-create-j7p2d\" (UID: \"cf071f6e-289f-4f86-9a39-e4dc62335b14\") " pod="openstack/nova-api-db-create-j7p2d" Jan 26 17:09:57 crc kubenswrapper[4754]: I0126 17:09:57.593798 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-bjmmr"] Jan 26 17:09:57 crc kubenswrapper[4754]: I0126 17:09:57.627186 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hmdrj\" (UniqueName: \"kubernetes.io/projected/cf071f6e-289f-4f86-9a39-e4dc62335b14-kube-api-access-hmdrj\") pod \"nova-api-db-create-j7p2d\" (UID: \"cf071f6e-289f-4f86-9a39-e4dc62335b14\") " pod="openstack/nova-api-db-create-j7p2d" Jan 26 17:09:57 crc kubenswrapper[4754]: I0126 17:09:57.678177 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-813d-account-create-update-wlvxd"] Jan 26 17:09:57 crc kubenswrapper[4754]: I0126 17:09:57.679507 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-813d-account-create-update-wlvxd" Jan 26 17:09:57 crc kubenswrapper[4754]: I0126 17:09:57.682685 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Jan 26 17:09:57 crc kubenswrapper[4754]: I0126 17:09:57.690621 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1597209c-abf0-4bd4-91b4-a8661e7e496b-operator-scripts\") pod \"nova-api-e6d7-account-create-update-hs5pg\" (UID: \"1597209c-abf0-4bd4-91b4-a8661e7e496b\") " pod="openstack/nova-api-e6d7-account-create-update-hs5pg" Jan 26 17:09:57 crc kubenswrapper[4754]: I0126 17:09:57.691004 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-96snk\" (UniqueName: \"kubernetes.io/projected/755ea18f-9963-409b-a3a2-56d110120dc8-kube-api-access-96snk\") pod \"nova-cell1-db-create-bjmmr\" (UID: \"755ea18f-9963-409b-a3a2-56d110120dc8\") " pod="openstack/nova-cell1-db-create-bjmmr" Jan 26 17:09:57 crc kubenswrapper[4754]: I0126 17:09:57.691039 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f5c367c8-74fd-4073-bbc6-2f6d4373f5ef-operator-scripts\") pod \"nova-cell0-db-create-6qdbg\" (UID: \"f5c367c8-74fd-4073-bbc6-2f6d4373f5ef\") " pod="openstack/nova-cell0-db-create-6qdbg" Jan 26 17:09:57 crc kubenswrapper[4754]: I0126 17:09:57.691201 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/755ea18f-9963-409b-a3a2-56d110120dc8-operator-scripts\") pod \"nova-cell1-db-create-bjmmr\" (UID: \"755ea18f-9963-409b-a3a2-56d110120dc8\") " pod="openstack/nova-cell1-db-create-bjmmr" Jan 26 17:09:57 crc kubenswrapper[4754]: I0126 17:09:57.691281 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j5hr4\" (UniqueName: \"kubernetes.io/projected/1597209c-abf0-4bd4-91b4-a8661e7e496b-kube-api-access-j5hr4\") pod \"nova-api-e6d7-account-create-update-hs5pg\" (UID: \"1597209c-abf0-4bd4-91b4-a8661e7e496b\") " pod="openstack/nova-api-e6d7-account-create-update-hs5pg" Jan 26 17:09:57 crc kubenswrapper[4754]: I0126 17:09:57.691727 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1597209c-abf0-4bd4-91b4-a8661e7e496b-operator-scripts\") pod \"nova-api-e6d7-account-create-update-hs5pg\" (UID: \"1597209c-abf0-4bd4-91b4-a8661e7e496b\") " pod="openstack/nova-api-e6d7-account-create-update-hs5pg" Jan 26 17:09:57 crc kubenswrapper[4754]: I0126 17:09:57.691743 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wmx9f\" (UniqueName: \"kubernetes.io/projected/f5c367c8-74fd-4073-bbc6-2f6d4373f5ef-kube-api-access-wmx9f\") pod \"nova-cell0-db-create-6qdbg\" (UID: \"f5c367c8-74fd-4073-bbc6-2f6d4373f5ef\") " pod="openstack/nova-cell0-db-create-6qdbg" Jan 26 17:09:57 crc kubenswrapper[4754]: I0126 17:09:57.692958 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f5c367c8-74fd-4073-bbc6-2f6d4373f5ef-operator-scripts\") pod \"nova-cell0-db-create-6qdbg\" (UID: \"f5c367c8-74fd-4073-bbc6-2f6d4373f5ef\") " pod="openstack/nova-cell0-db-create-6qdbg" Jan 26 17:09:57 crc kubenswrapper[4754]: I0126 17:09:57.697844 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-813d-account-create-update-wlvxd"] Jan 26 17:09:57 crc kubenswrapper[4754]: I0126 17:09:57.705631 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-j7p2d" Jan 26 17:09:57 crc kubenswrapper[4754]: I0126 17:09:57.710372 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j5hr4\" (UniqueName: \"kubernetes.io/projected/1597209c-abf0-4bd4-91b4-a8661e7e496b-kube-api-access-j5hr4\") pod \"nova-api-e6d7-account-create-update-hs5pg\" (UID: \"1597209c-abf0-4bd4-91b4-a8661e7e496b\") " pod="openstack/nova-api-e6d7-account-create-update-hs5pg" Jan 26 17:09:57 crc kubenswrapper[4754]: I0126 17:09:57.713213 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wmx9f\" (UniqueName: \"kubernetes.io/projected/f5c367c8-74fd-4073-bbc6-2f6d4373f5ef-kube-api-access-wmx9f\") pod \"nova-cell0-db-create-6qdbg\" (UID: \"f5c367c8-74fd-4073-bbc6-2f6d4373f5ef\") " pod="openstack/nova-cell0-db-create-6qdbg" Jan 26 17:09:57 crc kubenswrapper[4754]: I0126 17:09:57.778361 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-6qdbg" Jan 26 17:09:57 crc kubenswrapper[4754]: I0126 17:09:57.804570 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="977ff39d-2c55-438b-b957-62609c4c2ece" path="/var/lib/kubelet/pods/977ff39d-2c55-438b-b957-62609c4c2ece/volumes" Jan 26 17:09:57 crc kubenswrapper[4754]: I0126 17:09:57.813050 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c4mzh\" (UniqueName: \"kubernetes.io/projected/98a8c733-ba31-46a3-97ea-409b69050b02-kube-api-access-c4mzh\") pod \"nova-cell0-813d-account-create-update-wlvxd\" (UID: \"98a8c733-ba31-46a3-97ea-409b69050b02\") " pod="openstack/nova-cell0-813d-account-create-update-wlvxd" Jan 26 17:09:57 crc kubenswrapper[4754]: I0126 17:09:57.813136 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-96snk\" (UniqueName: \"kubernetes.io/projected/755ea18f-9963-409b-a3a2-56d110120dc8-kube-api-access-96snk\") pod \"nova-cell1-db-create-bjmmr\" (UID: \"755ea18f-9963-409b-a3a2-56d110120dc8\") " pod="openstack/nova-cell1-db-create-bjmmr" Jan 26 17:09:57 crc kubenswrapper[4754]: I0126 17:09:57.813174 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/755ea18f-9963-409b-a3a2-56d110120dc8-operator-scripts\") pod \"nova-cell1-db-create-bjmmr\" (UID: \"755ea18f-9963-409b-a3a2-56d110120dc8\") " pod="openstack/nova-cell1-db-create-bjmmr" Jan 26 17:09:57 crc kubenswrapper[4754]: I0126 17:09:57.813212 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/98a8c733-ba31-46a3-97ea-409b69050b02-operator-scripts\") pod \"nova-cell0-813d-account-create-update-wlvxd\" (UID: \"98a8c733-ba31-46a3-97ea-409b69050b02\") " pod="openstack/nova-cell0-813d-account-create-update-wlvxd" Jan 26 17:09:57 crc kubenswrapper[4754]: I0126 17:09:57.814417 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/755ea18f-9963-409b-a3a2-56d110120dc8-operator-scripts\") pod \"nova-cell1-db-create-bjmmr\" (UID: \"755ea18f-9963-409b-a3a2-56d110120dc8\") " pod="openstack/nova-cell1-db-create-bjmmr" Jan 26 17:09:57 crc kubenswrapper[4754]: I0126 17:09:57.817970 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-e6d7-account-create-update-hs5pg" Jan 26 17:09:57 crc kubenswrapper[4754]: I0126 17:09:57.883008 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-96snk\" (UniqueName: \"kubernetes.io/projected/755ea18f-9963-409b-a3a2-56d110120dc8-kube-api-access-96snk\") pod \"nova-cell1-db-create-bjmmr\" (UID: \"755ea18f-9963-409b-a3a2-56d110120dc8\") " pod="openstack/nova-cell1-db-create-bjmmr" Jan 26 17:09:57 crc kubenswrapper[4754]: I0126 17:09:57.907888 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-bjmmr" Jan 26 17:09:57 crc kubenswrapper[4754]: I0126 17:09:57.910275 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-2d04-account-create-update-lnbkh"] Jan 26 17:09:57 crc kubenswrapper[4754]: I0126 17:09:57.912250 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-2d04-account-create-update-lnbkh" Jan 26 17:09:57 crc kubenswrapper[4754]: I0126 17:09:57.916795 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Jan 26 17:09:57 crc kubenswrapper[4754]: I0126 17:09:57.917360 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/98a8c733-ba31-46a3-97ea-409b69050b02-operator-scripts\") pod \"nova-cell0-813d-account-create-update-wlvxd\" (UID: \"98a8c733-ba31-46a3-97ea-409b69050b02\") " pod="openstack/nova-cell0-813d-account-create-update-wlvxd" Jan 26 17:09:57 crc kubenswrapper[4754]: I0126 17:09:57.917524 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c4mzh\" (UniqueName: \"kubernetes.io/projected/98a8c733-ba31-46a3-97ea-409b69050b02-kube-api-access-c4mzh\") pod \"nova-cell0-813d-account-create-update-wlvxd\" (UID: \"98a8c733-ba31-46a3-97ea-409b69050b02\") " pod="openstack/nova-cell0-813d-account-create-update-wlvxd" Jan 26 17:09:57 crc kubenswrapper[4754]: I0126 17:09:57.918084 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/98a8c733-ba31-46a3-97ea-409b69050b02-operator-scripts\") pod \"nova-cell0-813d-account-create-update-wlvxd\" (UID: \"98a8c733-ba31-46a3-97ea-409b69050b02\") " pod="openstack/nova-cell0-813d-account-create-update-wlvxd" Jan 26 17:09:57 crc kubenswrapper[4754]: I0126 17:09:57.922606 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-2d04-account-create-update-lnbkh"] Jan 26 17:09:57 crc kubenswrapper[4754]: I0126 17:09:57.952280 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c4mzh\" (UniqueName: \"kubernetes.io/projected/98a8c733-ba31-46a3-97ea-409b69050b02-kube-api-access-c4mzh\") pod \"nova-cell0-813d-account-create-update-wlvxd\" (UID: \"98a8c733-ba31-46a3-97ea-409b69050b02\") " pod="openstack/nova-cell0-813d-account-create-update-wlvxd" Jan 26 17:09:58 crc kubenswrapper[4754]: I0126 17:09:58.005766 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-813d-account-create-update-wlvxd" Jan 26 17:09:58 crc kubenswrapper[4754]: I0126 17:09:58.125917 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nc6ns\" (UniqueName: \"kubernetes.io/projected/bef625bb-6bee-4c89-a9b4-4f695ab4ffd5-kube-api-access-nc6ns\") pod \"nova-cell1-2d04-account-create-update-lnbkh\" (UID: \"bef625bb-6bee-4c89-a9b4-4f695ab4ffd5\") " pod="openstack/nova-cell1-2d04-account-create-update-lnbkh" Jan 26 17:09:58 crc kubenswrapper[4754]: I0126 17:09:58.125990 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bef625bb-6bee-4c89-a9b4-4f695ab4ffd5-operator-scripts\") pod \"nova-cell1-2d04-account-create-update-lnbkh\" (UID: \"bef625bb-6bee-4c89-a9b4-4f695ab4ffd5\") " pod="openstack/nova-cell1-2d04-account-create-update-lnbkh" Jan 26 17:09:58 crc kubenswrapper[4754]: I0126 17:09:58.228120 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nc6ns\" (UniqueName: \"kubernetes.io/projected/bef625bb-6bee-4c89-a9b4-4f695ab4ffd5-kube-api-access-nc6ns\") pod \"nova-cell1-2d04-account-create-update-lnbkh\" (UID: \"bef625bb-6bee-4c89-a9b4-4f695ab4ffd5\") " pod="openstack/nova-cell1-2d04-account-create-update-lnbkh" Jan 26 17:09:58 crc kubenswrapper[4754]: I0126 17:09:58.228413 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bef625bb-6bee-4c89-a9b4-4f695ab4ffd5-operator-scripts\") pod \"nova-cell1-2d04-account-create-update-lnbkh\" (UID: \"bef625bb-6bee-4c89-a9b4-4f695ab4ffd5\") " pod="openstack/nova-cell1-2d04-account-create-update-lnbkh" Jan 26 17:09:58 crc kubenswrapper[4754]: I0126 17:09:58.230151 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bef625bb-6bee-4c89-a9b4-4f695ab4ffd5-operator-scripts\") pod \"nova-cell1-2d04-account-create-update-lnbkh\" (UID: \"bef625bb-6bee-4c89-a9b4-4f695ab4ffd5\") " pod="openstack/nova-cell1-2d04-account-create-update-lnbkh" Jan 26 17:09:58 crc kubenswrapper[4754]: I0126 17:09:58.256387 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nc6ns\" (UniqueName: \"kubernetes.io/projected/bef625bb-6bee-4c89-a9b4-4f695ab4ffd5-kube-api-access-nc6ns\") pod \"nova-cell1-2d04-account-create-update-lnbkh\" (UID: \"bef625bb-6bee-4c89-a9b4-4f695ab4ffd5\") " pod="openstack/nova-cell1-2d04-account-create-update-lnbkh" Jan 26 17:09:58 crc kubenswrapper[4754]: I0126 17:09:58.499547 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-2d04-account-create-update-lnbkh" Jan 26 17:09:58 crc kubenswrapper[4754]: I0126 17:09:58.507103 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-6qdbg"] Jan 26 17:09:58 crc kubenswrapper[4754]: I0126 17:09:58.631607 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-e6d7-account-create-update-hs5pg"] Jan 26 17:09:58 crc kubenswrapper[4754]: I0126 17:09:58.707277 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-bjmmr"] Jan 26 17:09:58 crc kubenswrapper[4754]: I0126 17:09:58.715306 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-813d-account-create-update-wlvxd"] Jan 26 17:10:00 crc kubenswrapper[4754]: I0126 17:10:00.011788 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-j7p2d"] Jan 26 17:10:00 crc kubenswrapper[4754]: I0126 17:10:00.698810 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-rpnm5" Jan 26 17:10:00 crc kubenswrapper[4754]: I0126 17:10:00.698858 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-rpnm5" Jan 26 17:10:00 crc kubenswrapper[4754]: I0126 17:10:00.750336 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-rpnm5" Jan 26 17:10:01 crc kubenswrapper[4754]: I0126 17:10:01.237870 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-rpnm5" Jan 26 17:10:01 crc kubenswrapper[4754]: I0126 17:10:01.278870 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-rpnm5"] Jan 26 17:10:01 crc kubenswrapper[4754]: W0126 17:10:01.911825 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf5c367c8_74fd_4073_bbc6_2f6d4373f5ef.slice/crio-41754b8b3bd59d6a76cccbf5dd3492fcea85d0743c5cf8e94bda8d15c88322ee WatchSource:0}: Error finding container 41754b8b3bd59d6a76cccbf5dd3492fcea85d0743c5cf8e94bda8d15c88322ee: Status 404 returned error can't find the container with id 41754b8b3bd59d6a76cccbf5dd3492fcea85d0743c5cf8e94bda8d15c88322ee Jan 26 17:10:01 crc kubenswrapper[4754]: W0126 17:10:01.916050 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1597209c_abf0_4bd4_91b4_a8661e7e496b.slice/crio-5f4f915cc51fb2bb1ba18629b89d4736af75c21a808ea8e225d9cf5611b1bc7c WatchSource:0}: Error finding container 5f4f915cc51fb2bb1ba18629b89d4736af75c21a808ea8e225d9cf5611b1bc7c: Status 404 returned error can't find the container with id 5f4f915cc51fb2bb1ba18629b89d4736af75c21a808ea8e225d9cf5611b1bc7c Jan 26 17:10:02 crc kubenswrapper[4754]: I0126 17:10:02.214559 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-bjmmr" event={"ID":"755ea18f-9963-409b-a3a2-56d110120dc8","Type":"ContainerStarted","Data":"73c10d16d075256d7fcf476f01102132fdfc1e99dd2fc392b476299d71d9a526"} Jan 26 17:10:02 crc kubenswrapper[4754]: I0126 17:10:02.220867 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-813d-account-create-update-wlvxd" event={"ID":"98a8c733-ba31-46a3-97ea-409b69050b02","Type":"ContainerStarted","Data":"e9e5781b38ee757fcd7e87080808617d63a1378fa5c9d44bde00775cfd839f7f"} Jan 26 17:10:02 crc kubenswrapper[4754]: I0126 17:10:02.228145 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-j7p2d" event={"ID":"cf071f6e-289f-4f86-9a39-e4dc62335b14","Type":"ContainerStarted","Data":"111de648becfef30078789742f35e42f12d02d89975e6542e596768f244c95c0"} Jan 26 17:10:02 crc kubenswrapper[4754]: I0126 17:10:02.229443 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-e6d7-account-create-update-hs5pg" event={"ID":"1597209c-abf0-4bd4-91b4-a8661e7e496b","Type":"ContainerStarted","Data":"5f4f915cc51fb2bb1ba18629b89d4736af75c21a808ea8e225d9cf5611b1bc7c"} Jan 26 17:10:02 crc kubenswrapper[4754]: I0126 17:10:02.230921 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-6qdbg" event={"ID":"f5c367c8-74fd-4073-bbc6-2f6d4373f5ef","Type":"ContainerStarted","Data":"41754b8b3bd59d6a76cccbf5dd3492fcea85d0743c5cf8e94bda8d15c88322ee"} Jan 26 17:10:02 crc kubenswrapper[4754]: I0126 17:10:02.375683 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-2d04-account-create-update-lnbkh"] Jan 26 17:10:02 crc kubenswrapper[4754]: W0126 17:10:02.376222 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbef625bb_6bee_4c89_a9b4_4f695ab4ffd5.slice/crio-26aea507ec503e03e2e0eacab6cba4f0bf83432fca3ae41a661f742647a10c40 WatchSource:0}: Error finding container 26aea507ec503e03e2e0eacab6cba4f0bf83432fca3ae41a661f742647a10c40: Status 404 returned error can't find the container with id 26aea507ec503e03e2e0eacab6cba4f0bf83432fca3ae41a661f742647a10c40 Jan 26 17:10:03 crc kubenswrapper[4754]: I0126 17:10:03.241580 4754 generic.go:334] "Generic (PLEG): container finished" podID="98a8c733-ba31-46a3-97ea-409b69050b02" containerID="f52878497a7ebd51d5218180b3f05165fc0ac52700a5ed47fb29b6523f6597cb" exitCode=0 Jan 26 17:10:03 crc kubenswrapper[4754]: I0126 17:10:03.241695 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-813d-account-create-update-wlvxd" event={"ID":"98a8c733-ba31-46a3-97ea-409b69050b02","Type":"ContainerDied","Data":"f52878497a7ebd51d5218180b3f05165fc0ac52700a5ed47fb29b6523f6597cb"} Jan 26 17:10:03 crc kubenswrapper[4754]: I0126 17:10:03.243243 4754 generic.go:334] "Generic (PLEG): container finished" podID="cf071f6e-289f-4f86-9a39-e4dc62335b14" containerID="2ac2874e0106d3797b0f67a75cd93ccea19413c4e35875f30d77331553081397" exitCode=0 Jan 26 17:10:03 crc kubenswrapper[4754]: I0126 17:10:03.243271 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-j7p2d" event={"ID":"cf071f6e-289f-4f86-9a39-e4dc62335b14","Type":"ContainerDied","Data":"2ac2874e0106d3797b0f67a75cd93ccea19413c4e35875f30d77331553081397"} Jan 26 17:10:03 crc kubenswrapper[4754]: I0126 17:10:03.245606 4754 generic.go:334] "Generic (PLEG): container finished" podID="1597209c-abf0-4bd4-91b4-a8661e7e496b" containerID="7520af6d81b045bab0b0552880d2a9323b1410e4bdbc087b69f3024f2836cebf" exitCode=0 Jan 26 17:10:03 crc kubenswrapper[4754]: I0126 17:10:03.245655 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-e6d7-account-create-update-hs5pg" event={"ID":"1597209c-abf0-4bd4-91b4-a8661e7e496b","Type":"ContainerDied","Data":"7520af6d81b045bab0b0552880d2a9323b1410e4bdbc087b69f3024f2836cebf"} Jan 26 17:10:03 crc kubenswrapper[4754]: I0126 17:10:03.246948 4754 generic.go:334] "Generic (PLEG): container finished" podID="f5c367c8-74fd-4073-bbc6-2f6d4373f5ef" containerID="e3184d6ce076367c5bd58f3200fab9ce3aff97ec4c504666561622fa2dc4f696" exitCode=0 Jan 26 17:10:03 crc kubenswrapper[4754]: I0126 17:10:03.247012 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-6qdbg" event={"ID":"f5c367c8-74fd-4073-bbc6-2f6d4373f5ef","Type":"ContainerDied","Data":"e3184d6ce076367c5bd58f3200fab9ce3aff97ec4c504666561622fa2dc4f696"} Jan 26 17:10:03 crc kubenswrapper[4754]: I0126 17:10:03.248531 4754 generic.go:334] "Generic (PLEG): container finished" podID="bef625bb-6bee-4c89-a9b4-4f695ab4ffd5" containerID="c502c69ad12ddc1b2b3981e832270d6541d21c9abdf23ec4e61ebfe2fd929ac9" exitCode=0 Jan 26 17:10:03 crc kubenswrapper[4754]: I0126 17:10:03.248567 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-2d04-account-create-update-lnbkh" event={"ID":"bef625bb-6bee-4c89-a9b4-4f695ab4ffd5","Type":"ContainerDied","Data":"c502c69ad12ddc1b2b3981e832270d6541d21c9abdf23ec4e61ebfe2fd929ac9"} Jan 26 17:10:03 crc kubenswrapper[4754]: I0126 17:10:03.248597 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-2d04-account-create-update-lnbkh" event={"ID":"bef625bb-6bee-4c89-a9b4-4f695ab4ffd5","Type":"ContainerStarted","Data":"26aea507ec503e03e2e0eacab6cba4f0bf83432fca3ae41a661f742647a10c40"} Jan 26 17:10:03 crc kubenswrapper[4754]: I0126 17:10:03.249916 4754 generic.go:334] "Generic (PLEG): container finished" podID="755ea18f-9963-409b-a3a2-56d110120dc8" containerID="156feceeb22f23d1a2ff8accdb5d5abdf7fe3efad41579a732554fecce8bf645" exitCode=0 Jan 26 17:10:03 crc kubenswrapper[4754]: I0126 17:10:03.250007 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-bjmmr" event={"ID":"755ea18f-9963-409b-a3a2-56d110120dc8","Type":"ContainerDied","Data":"156feceeb22f23d1a2ff8accdb5d5abdf7fe3efad41579a732554fecce8bf645"} Jan 26 17:10:03 crc kubenswrapper[4754]: I0126 17:10:03.250097 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-rpnm5" podUID="7dc29a9d-f578-49ec-97a8-08a80bd3a2c0" containerName="registry-server" containerID="cri-o://510e54a25e15cd76ffea612e300eea6a77041185907b8c15443c0adbf7ee8b78" gracePeriod=2 Jan 26 17:10:03 crc kubenswrapper[4754]: I0126 17:10:03.801751 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rpnm5" Jan 26 17:10:03 crc kubenswrapper[4754]: I0126 17:10:03.940844 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7dc29a9d-f578-49ec-97a8-08a80bd3a2c0-catalog-content\") pod \"7dc29a9d-f578-49ec-97a8-08a80bd3a2c0\" (UID: \"7dc29a9d-f578-49ec-97a8-08a80bd3a2c0\") " Jan 26 17:10:03 crc kubenswrapper[4754]: I0126 17:10:03.941056 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pqvjk\" (UniqueName: \"kubernetes.io/projected/7dc29a9d-f578-49ec-97a8-08a80bd3a2c0-kube-api-access-pqvjk\") pod \"7dc29a9d-f578-49ec-97a8-08a80bd3a2c0\" (UID: \"7dc29a9d-f578-49ec-97a8-08a80bd3a2c0\") " Jan 26 17:10:03 crc kubenswrapper[4754]: I0126 17:10:03.941160 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7dc29a9d-f578-49ec-97a8-08a80bd3a2c0-utilities\") pod \"7dc29a9d-f578-49ec-97a8-08a80bd3a2c0\" (UID: \"7dc29a9d-f578-49ec-97a8-08a80bd3a2c0\") " Jan 26 17:10:03 crc kubenswrapper[4754]: I0126 17:10:03.941832 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7dc29a9d-f578-49ec-97a8-08a80bd3a2c0-utilities" (OuterVolumeSpecName: "utilities") pod "7dc29a9d-f578-49ec-97a8-08a80bd3a2c0" (UID: "7dc29a9d-f578-49ec-97a8-08a80bd3a2c0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:10:03 crc kubenswrapper[4754]: I0126 17:10:03.948625 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7dc29a9d-f578-49ec-97a8-08a80bd3a2c0-kube-api-access-pqvjk" (OuterVolumeSpecName: "kube-api-access-pqvjk") pod "7dc29a9d-f578-49ec-97a8-08a80bd3a2c0" (UID: "7dc29a9d-f578-49ec-97a8-08a80bd3a2c0"). InnerVolumeSpecName "kube-api-access-pqvjk". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:10:03 crc kubenswrapper[4754]: I0126 17:10:03.980961 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7dc29a9d-f578-49ec-97a8-08a80bd3a2c0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7dc29a9d-f578-49ec-97a8-08a80bd3a2c0" (UID: "7dc29a9d-f578-49ec-97a8-08a80bd3a2c0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:10:04 crc kubenswrapper[4754]: I0126 17:10:04.043146 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pqvjk\" (UniqueName: \"kubernetes.io/projected/7dc29a9d-f578-49ec-97a8-08a80bd3a2c0-kube-api-access-pqvjk\") on node \"crc\" DevicePath \"\"" Jan 26 17:10:04 crc kubenswrapper[4754]: I0126 17:10:04.043177 4754 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7dc29a9d-f578-49ec-97a8-08a80bd3a2c0-utilities\") on node \"crc\" DevicePath \"\"" Jan 26 17:10:04 crc kubenswrapper[4754]: I0126 17:10:04.043193 4754 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7dc29a9d-f578-49ec-97a8-08a80bd3a2c0-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 26 17:10:04 crc kubenswrapper[4754]: I0126 17:10:04.273541 4754 generic.go:334] "Generic (PLEG): container finished" podID="ee49bc9b-edbd-4b7b-b977-98ec0b6b21c4" containerID="e350d7d898f163921729bb47f1b2fd06ba49d5a07efcd6a4f1fa7505a6a47da7" exitCode=0 Jan 26 17:10:04 crc kubenswrapper[4754]: I0126 17:10:04.273911 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ee49bc9b-edbd-4b7b-b977-98ec0b6b21c4","Type":"ContainerDied","Data":"e350d7d898f163921729bb47f1b2fd06ba49d5a07efcd6a4f1fa7505a6a47da7"} Jan 26 17:10:04 crc kubenswrapper[4754]: I0126 17:10:04.277218 4754 generic.go:334] "Generic (PLEG): container finished" podID="7dc29a9d-f578-49ec-97a8-08a80bd3a2c0" containerID="510e54a25e15cd76ffea612e300eea6a77041185907b8c15443c0adbf7ee8b78" exitCode=0 Jan 26 17:10:04 crc kubenswrapper[4754]: I0126 17:10:04.277313 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rpnm5" event={"ID":"7dc29a9d-f578-49ec-97a8-08a80bd3a2c0","Type":"ContainerDied","Data":"510e54a25e15cd76ffea612e300eea6a77041185907b8c15443c0adbf7ee8b78"} Jan 26 17:10:04 crc kubenswrapper[4754]: I0126 17:10:04.277326 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rpnm5" Jan 26 17:10:04 crc kubenswrapper[4754]: I0126 17:10:04.277375 4754 scope.go:117] "RemoveContainer" containerID="510e54a25e15cd76ffea612e300eea6a77041185907b8c15443c0adbf7ee8b78" Jan 26 17:10:04 crc kubenswrapper[4754]: I0126 17:10:04.277359 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rpnm5" event={"ID":"7dc29a9d-f578-49ec-97a8-08a80bd3a2c0","Type":"ContainerDied","Data":"591c15aa136778dda5707b6d79f17694f62bd03a891f83026afe4b5113edd8f1"} Jan 26 17:10:04 crc kubenswrapper[4754]: I0126 17:10:04.313337 4754 scope.go:117] "RemoveContainer" containerID="e7d747ff339856686f9c794f4314c3a1d4356cd1668664a52a35543107d47c3e" Jan 26 17:10:04 crc kubenswrapper[4754]: I0126 17:10:04.345655 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-rpnm5"] Jan 26 17:10:04 crc kubenswrapper[4754]: I0126 17:10:04.357191 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-rpnm5"] Jan 26 17:10:04 crc kubenswrapper[4754]: I0126 17:10:04.360191 4754 scope.go:117] "RemoveContainer" containerID="269c5e8cc3d5b27f5d220eb92d7616cc38a520f9c0c5a6c0d4f859c1a6c8a16d" Jan 26 17:10:04 crc kubenswrapper[4754]: I0126 17:10:04.423930 4754 scope.go:117] "RemoveContainer" containerID="510e54a25e15cd76ffea612e300eea6a77041185907b8c15443c0adbf7ee8b78" Jan 26 17:10:04 crc kubenswrapper[4754]: E0126 17:10:04.428556 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"510e54a25e15cd76ffea612e300eea6a77041185907b8c15443c0adbf7ee8b78\": container with ID starting with 510e54a25e15cd76ffea612e300eea6a77041185907b8c15443c0adbf7ee8b78 not found: ID does not exist" containerID="510e54a25e15cd76ffea612e300eea6a77041185907b8c15443c0adbf7ee8b78" Jan 26 17:10:04 crc kubenswrapper[4754]: I0126 17:10:04.428615 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"510e54a25e15cd76ffea612e300eea6a77041185907b8c15443c0adbf7ee8b78"} err="failed to get container status \"510e54a25e15cd76ffea612e300eea6a77041185907b8c15443c0adbf7ee8b78\": rpc error: code = NotFound desc = could not find container \"510e54a25e15cd76ffea612e300eea6a77041185907b8c15443c0adbf7ee8b78\": container with ID starting with 510e54a25e15cd76ffea612e300eea6a77041185907b8c15443c0adbf7ee8b78 not found: ID does not exist" Jan 26 17:10:04 crc kubenswrapper[4754]: I0126 17:10:04.428647 4754 scope.go:117] "RemoveContainer" containerID="e7d747ff339856686f9c794f4314c3a1d4356cd1668664a52a35543107d47c3e" Jan 26 17:10:04 crc kubenswrapper[4754]: E0126 17:10:04.429156 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e7d747ff339856686f9c794f4314c3a1d4356cd1668664a52a35543107d47c3e\": container with ID starting with e7d747ff339856686f9c794f4314c3a1d4356cd1668664a52a35543107d47c3e not found: ID does not exist" containerID="e7d747ff339856686f9c794f4314c3a1d4356cd1668664a52a35543107d47c3e" Jan 26 17:10:04 crc kubenswrapper[4754]: I0126 17:10:04.429201 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e7d747ff339856686f9c794f4314c3a1d4356cd1668664a52a35543107d47c3e"} err="failed to get container status \"e7d747ff339856686f9c794f4314c3a1d4356cd1668664a52a35543107d47c3e\": rpc error: code = NotFound desc = could not find container \"e7d747ff339856686f9c794f4314c3a1d4356cd1668664a52a35543107d47c3e\": container with ID starting with e7d747ff339856686f9c794f4314c3a1d4356cd1668664a52a35543107d47c3e not found: ID does not exist" Jan 26 17:10:04 crc kubenswrapper[4754]: I0126 17:10:04.429219 4754 scope.go:117] "RemoveContainer" containerID="269c5e8cc3d5b27f5d220eb92d7616cc38a520f9c0c5a6c0d4f859c1a6c8a16d" Jan 26 17:10:04 crc kubenswrapper[4754]: E0126 17:10:04.429454 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"269c5e8cc3d5b27f5d220eb92d7616cc38a520f9c0c5a6c0d4f859c1a6c8a16d\": container with ID starting with 269c5e8cc3d5b27f5d220eb92d7616cc38a520f9c0c5a6c0d4f859c1a6c8a16d not found: ID does not exist" containerID="269c5e8cc3d5b27f5d220eb92d7616cc38a520f9c0c5a6c0d4f859c1a6c8a16d" Jan 26 17:10:04 crc kubenswrapper[4754]: I0126 17:10:04.429481 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"269c5e8cc3d5b27f5d220eb92d7616cc38a520f9c0c5a6c0d4f859c1a6c8a16d"} err="failed to get container status \"269c5e8cc3d5b27f5d220eb92d7616cc38a520f9c0c5a6c0d4f859c1a6c8a16d\": rpc error: code = NotFound desc = could not find container \"269c5e8cc3d5b27f5d220eb92d7616cc38a520f9c0c5a6c0d4f859c1a6c8a16d\": container with ID starting with 269c5e8cc3d5b27f5d220eb92d7616cc38a520f9c0c5a6c0d4f859c1a6c8a16d not found: ID does not exist" Jan 26 17:10:04 crc kubenswrapper[4754]: I0126 17:10:04.675761 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-bjmmr" Jan 26 17:10:04 crc kubenswrapper[4754]: I0126 17:10:04.757899 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-96snk\" (UniqueName: \"kubernetes.io/projected/755ea18f-9963-409b-a3a2-56d110120dc8-kube-api-access-96snk\") pod \"755ea18f-9963-409b-a3a2-56d110120dc8\" (UID: \"755ea18f-9963-409b-a3a2-56d110120dc8\") " Jan 26 17:10:04 crc kubenswrapper[4754]: I0126 17:10:04.758310 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/755ea18f-9963-409b-a3a2-56d110120dc8-operator-scripts\") pod \"755ea18f-9963-409b-a3a2-56d110120dc8\" (UID: \"755ea18f-9963-409b-a3a2-56d110120dc8\") " Jan 26 17:10:04 crc kubenswrapper[4754]: I0126 17:10:04.758888 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/755ea18f-9963-409b-a3a2-56d110120dc8-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "755ea18f-9963-409b-a3a2-56d110120dc8" (UID: "755ea18f-9963-409b-a3a2-56d110120dc8"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:10:04 crc kubenswrapper[4754]: I0126 17:10:04.759144 4754 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/755ea18f-9963-409b-a3a2-56d110120dc8-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 26 17:10:04 crc kubenswrapper[4754]: I0126 17:10:04.762656 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/755ea18f-9963-409b-a3a2-56d110120dc8-kube-api-access-96snk" (OuterVolumeSpecName: "kube-api-access-96snk") pod "755ea18f-9963-409b-a3a2-56d110120dc8" (UID: "755ea18f-9963-409b-a3a2-56d110120dc8"). InnerVolumeSpecName "kube-api-access-96snk". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:10:04 crc kubenswrapper[4754]: I0126 17:10:04.850161 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-6qdbg" Jan 26 17:10:04 crc kubenswrapper[4754]: I0126 17:10:04.857877 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-e6d7-account-create-update-hs5pg" Jan 26 17:10:04 crc kubenswrapper[4754]: I0126 17:10:04.860002 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1597209c-abf0-4bd4-91b4-a8661e7e496b-operator-scripts\") pod \"1597209c-abf0-4bd4-91b4-a8661e7e496b\" (UID: \"1597209c-abf0-4bd4-91b4-a8661e7e496b\") " Jan 26 17:10:04 crc kubenswrapper[4754]: I0126 17:10:04.860073 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f5c367c8-74fd-4073-bbc6-2f6d4373f5ef-operator-scripts\") pod \"f5c367c8-74fd-4073-bbc6-2f6d4373f5ef\" (UID: \"f5c367c8-74fd-4073-bbc6-2f6d4373f5ef\") " Jan 26 17:10:04 crc kubenswrapper[4754]: I0126 17:10:04.860112 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j5hr4\" (UniqueName: \"kubernetes.io/projected/1597209c-abf0-4bd4-91b4-a8661e7e496b-kube-api-access-j5hr4\") pod \"1597209c-abf0-4bd4-91b4-a8661e7e496b\" (UID: \"1597209c-abf0-4bd4-91b4-a8661e7e496b\") " Jan 26 17:10:04 crc kubenswrapper[4754]: I0126 17:10:04.860213 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wmx9f\" (UniqueName: \"kubernetes.io/projected/f5c367c8-74fd-4073-bbc6-2f6d4373f5ef-kube-api-access-wmx9f\") pod \"f5c367c8-74fd-4073-bbc6-2f6d4373f5ef\" (UID: \"f5c367c8-74fd-4073-bbc6-2f6d4373f5ef\") " Jan 26 17:10:04 crc kubenswrapper[4754]: I0126 17:10:04.860541 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1597209c-abf0-4bd4-91b4-a8661e7e496b-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "1597209c-abf0-4bd4-91b4-a8661e7e496b" (UID: "1597209c-abf0-4bd4-91b4-a8661e7e496b"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:10:04 crc kubenswrapper[4754]: I0126 17:10:04.860711 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f5c367c8-74fd-4073-bbc6-2f6d4373f5ef-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "f5c367c8-74fd-4073-bbc6-2f6d4373f5ef" (UID: "f5c367c8-74fd-4073-bbc6-2f6d4373f5ef"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:10:04 crc kubenswrapper[4754]: I0126 17:10:04.860999 4754 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1597209c-abf0-4bd4-91b4-a8661e7e496b-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 26 17:10:04 crc kubenswrapper[4754]: I0126 17:10:04.861026 4754 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f5c367c8-74fd-4073-bbc6-2f6d4373f5ef-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 26 17:10:04 crc kubenswrapper[4754]: I0126 17:10:04.861039 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-96snk\" (UniqueName: \"kubernetes.io/projected/755ea18f-9963-409b-a3a2-56d110120dc8-kube-api-access-96snk\") on node \"crc\" DevicePath \"\"" Jan 26 17:10:04 crc kubenswrapper[4754]: I0126 17:10:04.864910 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f5c367c8-74fd-4073-bbc6-2f6d4373f5ef-kube-api-access-wmx9f" (OuterVolumeSpecName: "kube-api-access-wmx9f") pod "f5c367c8-74fd-4073-bbc6-2f6d4373f5ef" (UID: "f5c367c8-74fd-4073-bbc6-2f6d4373f5ef"). InnerVolumeSpecName "kube-api-access-wmx9f". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:10:04 crc kubenswrapper[4754]: I0126 17:10:04.866742 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1597209c-abf0-4bd4-91b4-a8661e7e496b-kube-api-access-j5hr4" (OuterVolumeSpecName: "kube-api-access-j5hr4") pod "1597209c-abf0-4bd4-91b4-a8661e7e496b" (UID: "1597209c-abf0-4bd4-91b4-a8661e7e496b"). InnerVolumeSpecName "kube-api-access-j5hr4". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:10:04 crc kubenswrapper[4754]: I0126 17:10:04.870900 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-813d-account-create-update-wlvxd" Jan 26 17:10:04 crc kubenswrapper[4754]: I0126 17:10:04.878984 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-j7p2d" Jan 26 17:10:04 crc kubenswrapper[4754]: I0126 17:10:04.888950 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-2d04-account-create-update-lnbkh" Jan 26 17:10:04 crc kubenswrapper[4754]: I0126 17:10:04.961965 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c4mzh\" (UniqueName: \"kubernetes.io/projected/98a8c733-ba31-46a3-97ea-409b69050b02-kube-api-access-c4mzh\") pod \"98a8c733-ba31-46a3-97ea-409b69050b02\" (UID: \"98a8c733-ba31-46a3-97ea-409b69050b02\") " Jan 26 17:10:04 crc kubenswrapper[4754]: I0126 17:10:04.962410 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bef625bb-6bee-4c89-a9b4-4f695ab4ffd5-operator-scripts\") pod \"bef625bb-6bee-4c89-a9b4-4f695ab4ffd5\" (UID: \"bef625bb-6bee-4c89-a9b4-4f695ab4ffd5\") " Jan 26 17:10:04 crc kubenswrapper[4754]: I0126 17:10:04.962447 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nc6ns\" (UniqueName: \"kubernetes.io/projected/bef625bb-6bee-4c89-a9b4-4f695ab4ffd5-kube-api-access-nc6ns\") pod \"bef625bb-6bee-4c89-a9b4-4f695ab4ffd5\" (UID: \"bef625bb-6bee-4c89-a9b4-4f695ab4ffd5\") " Jan 26 17:10:04 crc kubenswrapper[4754]: I0126 17:10:04.962482 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/98a8c733-ba31-46a3-97ea-409b69050b02-operator-scripts\") pod \"98a8c733-ba31-46a3-97ea-409b69050b02\" (UID: \"98a8c733-ba31-46a3-97ea-409b69050b02\") " Jan 26 17:10:04 crc kubenswrapper[4754]: I0126 17:10:04.962565 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hmdrj\" (UniqueName: \"kubernetes.io/projected/cf071f6e-289f-4f86-9a39-e4dc62335b14-kube-api-access-hmdrj\") pod \"cf071f6e-289f-4f86-9a39-e4dc62335b14\" (UID: \"cf071f6e-289f-4f86-9a39-e4dc62335b14\") " Jan 26 17:10:04 crc kubenswrapper[4754]: I0126 17:10:04.962688 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cf071f6e-289f-4f86-9a39-e4dc62335b14-operator-scripts\") pod \"cf071f6e-289f-4f86-9a39-e4dc62335b14\" (UID: \"cf071f6e-289f-4f86-9a39-e4dc62335b14\") " Jan 26 17:10:04 crc kubenswrapper[4754]: I0126 17:10:04.962844 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bef625bb-6bee-4c89-a9b4-4f695ab4ffd5-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "bef625bb-6bee-4c89-a9b4-4f695ab4ffd5" (UID: "bef625bb-6bee-4c89-a9b4-4f695ab4ffd5"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:10:04 crc kubenswrapper[4754]: I0126 17:10:04.963063 4754 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bef625bb-6bee-4c89-a9b4-4f695ab4ffd5-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 26 17:10:04 crc kubenswrapper[4754]: I0126 17:10:04.963075 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j5hr4\" (UniqueName: \"kubernetes.io/projected/1597209c-abf0-4bd4-91b4-a8661e7e496b-kube-api-access-j5hr4\") on node \"crc\" DevicePath \"\"" Jan 26 17:10:04 crc kubenswrapper[4754]: I0126 17:10:04.963085 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wmx9f\" (UniqueName: \"kubernetes.io/projected/f5c367c8-74fd-4073-bbc6-2f6d4373f5ef-kube-api-access-wmx9f\") on node \"crc\" DevicePath \"\"" Jan 26 17:10:04 crc kubenswrapper[4754]: I0126 17:10:04.963156 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/98a8c733-ba31-46a3-97ea-409b69050b02-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "98a8c733-ba31-46a3-97ea-409b69050b02" (UID: "98a8c733-ba31-46a3-97ea-409b69050b02"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:10:04 crc kubenswrapper[4754]: I0126 17:10:04.963339 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cf071f6e-289f-4f86-9a39-e4dc62335b14-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "cf071f6e-289f-4f86-9a39-e4dc62335b14" (UID: "cf071f6e-289f-4f86-9a39-e4dc62335b14"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:10:04 crc kubenswrapper[4754]: I0126 17:10:04.965222 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/98a8c733-ba31-46a3-97ea-409b69050b02-kube-api-access-c4mzh" (OuterVolumeSpecName: "kube-api-access-c4mzh") pod "98a8c733-ba31-46a3-97ea-409b69050b02" (UID: "98a8c733-ba31-46a3-97ea-409b69050b02"). InnerVolumeSpecName "kube-api-access-c4mzh". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:10:04 crc kubenswrapper[4754]: I0126 17:10:04.966874 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bef625bb-6bee-4c89-a9b4-4f695ab4ffd5-kube-api-access-nc6ns" (OuterVolumeSpecName: "kube-api-access-nc6ns") pod "bef625bb-6bee-4c89-a9b4-4f695ab4ffd5" (UID: "bef625bb-6bee-4c89-a9b4-4f695ab4ffd5"). InnerVolumeSpecName "kube-api-access-nc6ns". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:10:04 crc kubenswrapper[4754]: I0126 17:10:04.979580 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cf071f6e-289f-4f86-9a39-e4dc62335b14-kube-api-access-hmdrj" (OuterVolumeSpecName: "kube-api-access-hmdrj") pod "cf071f6e-289f-4f86-9a39-e4dc62335b14" (UID: "cf071f6e-289f-4f86-9a39-e4dc62335b14"). InnerVolumeSpecName "kube-api-access-hmdrj". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:10:05 crc kubenswrapper[4754]: I0126 17:10:05.064622 4754 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cf071f6e-289f-4f86-9a39-e4dc62335b14-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 26 17:10:05 crc kubenswrapper[4754]: I0126 17:10:05.064651 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c4mzh\" (UniqueName: \"kubernetes.io/projected/98a8c733-ba31-46a3-97ea-409b69050b02-kube-api-access-c4mzh\") on node \"crc\" DevicePath \"\"" Jan 26 17:10:05 crc kubenswrapper[4754]: I0126 17:10:05.064661 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nc6ns\" (UniqueName: \"kubernetes.io/projected/bef625bb-6bee-4c89-a9b4-4f695ab4ffd5-kube-api-access-nc6ns\") on node \"crc\" DevicePath \"\"" Jan 26 17:10:05 crc kubenswrapper[4754]: I0126 17:10:05.064720 4754 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/98a8c733-ba31-46a3-97ea-409b69050b02-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 26 17:10:05 crc kubenswrapper[4754]: I0126 17:10:05.064729 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hmdrj\" (UniqueName: \"kubernetes.io/projected/cf071f6e-289f-4f86-9a39-e4dc62335b14-kube-api-access-hmdrj\") on node \"crc\" DevicePath \"\"" Jan 26 17:10:05 crc kubenswrapper[4754]: I0126 17:10:05.286130 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-6qdbg" event={"ID":"f5c367c8-74fd-4073-bbc6-2f6d4373f5ef","Type":"ContainerDied","Data":"41754b8b3bd59d6a76cccbf5dd3492fcea85d0743c5cf8e94bda8d15c88322ee"} Jan 26 17:10:05 crc kubenswrapper[4754]: I0126 17:10:05.286357 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="41754b8b3bd59d6a76cccbf5dd3492fcea85d0743c5cf8e94bda8d15c88322ee" Jan 26 17:10:05 crc kubenswrapper[4754]: I0126 17:10:05.286171 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-6qdbg" Jan 26 17:10:05 crc kubenswrapper[4754]: I0126 17:10:05.287488 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-2d04-account-create-update-lnbkh" event={"ID":"bef625bb-6bee-4c89-a9b4-4f695ab4ffd5","Type":"ContainerDied","Data":"26aea507ec503e03e2e0eacab6cba4f0bf83432fca3ae41a661f742647a10c40"} Jan 26 17:10:05 crc kubenswrapper[4754]: I0126 17:10:05.287513 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="26aea507ec503e03e2e0eacab6cba4f0bf83432fca3ae41a661f742647a10c40" Jan 26 17:10:05 crc kubenswrapper[4754]: I0126 17:10:05.287495 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-2d04-account-create-update-lnbkh" Jan 26 17:10:05 crc kubenswrapper[4754]: I0126 17:10:05.290918 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-bjmmr" Jan 26 17:10:05 crc kubenswrapper[4754]: I0126 17:10:05.290891 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-bjmmr" event={"ID":"755ea18f-9963-409b-a3a2-56d110120dc8","Type":"ContainerDied","Data":"73c10d16d075256d7fcf476f01102132fdfc1e99dd2fc392b476299d71d9a526"} Jan 26 17:10:05 crc kubenswrapper[4754]: I0126 17:10:05.291249 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="73c10d16d075256d7fcf476f01102132fdfc1e99dd2fc392b476299d71d9a526" Jan 26 17:10:05 crc kubenswrapper[4754]: I0126 17:10:05.292579 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-813d-account-create-update-wlvxd" event={"ID":"98a8c733-ba31-46a3-97ea-409b69050b02","Type":"ContainerDied","Data":"e9e5781b38ee757fcd7e87080808617d63a1378fa5c9d44bde00775cfd839f7f"} Jan 26 17:10:05 crc kubenswrapper[4754]: I0126 17:10:05.292621 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e9e5781b38ee757fcd7e87080808617d63a1378fa5c9d44bde00775cfd839f7f" Jan 26 17:10:05 crc kubenswrapper[4754]: I0126 17:10:05.292750 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-813d-account-create-update-wlvxd" Jan 26 17:10:05 crc kubenswrapper[4754]: I0126 17:10:05.298825 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-j7p2d" event={"ID":"cf071f6e-289f-4f86-9a39-e4dc62335b14","Type":"ContainerDied","Data":"111de648becfef30078789742f35e42f12d02d89975e6542e596768f244c95c0"} Jan 26 17:10:05 crc kubenswrapper[4754]: I0126 17:10:05.298868 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="111de648becfef30078789742f35e42f12d02d89975e6542e596768f244c95c0" Jan 26 17:10:05 crc kubenswrapper[4754]: I0126 17:10:05.298966 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-j7p2d" Jan 26 17:10:05 crc kubenswrapper[4754]: I0126 17:10:05.302593 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-e6d7-account-create-update-hs5pg" event={"ID":"1597209c-abf0-4bd4-91b4-a8661e7e496b","Type":"ContainerDied","Data":"5f4f915cc51fb2bb1ba18629b89d4736af75c21a808ea8e225d9cf5611b1bc7c"} Jan 26 17:10:05 crc kubenswrapper[4754]: I0126 17:10:05.302630 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5f4f915cc51fb2bb1ba18629b89d4736af75c21a808ea8e225d9cf5611b1bc7c" Jan 26 17:10:05 crc kubenswrapper[4754]: I0126 17:10:05.302703 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-e6d7-account-create-update-hs5pg" Jan 26 17:10:05 crc kubenswrapper[4754]: I0126 17:10:05.779789 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7dc29a9d-f578-49ec-97a8-08a80bd3a2c0" path="/var/lib/kubelet/pods/7dc29a9d-f578-49ec-97a8-08a80bd3a2c0/volumes" Jan 26 17:10:07 crc kubenswrapper[4754]: I0126 17:10:07.129396 4754 patch_prober.go:28] interesting pod/machine-config-daemon-x65wv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 26 17:10:07 crc kubenswrapper[4754]: I0126 17:10:07.129854 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 26 17:10:07 crc kubenswrapper[4754]: I0126 17:10:07.129932 4754 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" Jan 26 17:10:07 crc kubenswrapper[4754]: I0126 17:10:07.130839 4754 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"7bdda0a504a055c787cfe163a27b6598986e7e593ba6562f6669a36ea09cda24"} pod="openshift-machine-config-operator/machine-config-daemon-x65wv" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 26 17:10:07 crc kubenswrapper[4754]: I0126 17:10:07.130907 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" containerName="machine-config-daemon" containerID="cri-o://7bdda0a504a055c787cfe163a27b6598986e7e593ba6562f6669a36ea09cda24" gracePeriod=600 Jan 26 17:10:07 crc kubenswrapper[4754]: I0126 17:10:07.320602 4754 generic.go:334] "Generic (PLEG): container finished" podID="8c3718a4-f354-4284-92e0-fdfb45a692bd" containerID="7bdda0a504a055c787cfe163a27b6598986e7e593ba6562f6669a36ea09cda24" exitCode=0 Jan 26 17:10:07 crc kubenswrapper[4754]: I0126 17:10:07.320654 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" event={"ID":"8c3718a4-f354-4284-92e0-fdfb45a692bd","Type":"ContainerDied","Data":"7bdda0a504a055c787cfe163a27b6598986e7e593ba6562f6669a36ea09cda24"} Jan 26 17:10:07 crc kubenswrapper[4754]: I0126 17:10:07.320710 4754 scope.go:117] "RemoveContainer" containerID="a1662bcf9213c8129b38140c4208b765dd0d2dd18def4d53852919fa5dab4d3e" Jan 26 17:10:07 crc kubenswrapper[4754]: I0126 17:10:07.894786 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-8lrb8"] Jan 26 17:10:07 crc kubenswrapper[4754]: E0126 17:10:07.895546 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7dc29a9d-f578-49ec-97a8-08a80bd3a2c0" containerName="extract-content" Jan 26 17:10:07 crc kubenswrapper[4754]: I0126 17:10:07.895569 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="7dc29a9d-f578-49ec-97a8-08a80bd3a2c0" containerName="extract-content" Jan 26 17:10:07 crc kubenswrapper[4754]: E0126 17:10:07.895588 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cf071f6e-289f-4f86-9a39-e4dc62335b14" containerName="mariadb-database-create" Jan 26 17:10:07 crc kubenswrapper[4754]: I0126 17:10:07.895599 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="cf071f6e-289f-4f86-9a39-e4dc62335b14" containerName="mariadb-database-create" Jan 26 17:10:07 crc kubenswrapper[4754]: E0126 17:10:07.895626 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7dc29a9d-f578-49ec-97a8-08a80bd3a2c0" containerName="registry-server" Jan 26 17:10:07 crc kubenswrapper[4754]: I0126 17:10:07.895635 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="7dc29a9d-f578-49ec-97a8-08a80bd3a2c0" containerName="registry-server" Jan 26 17:10:07 crc kubenswrapper[4754]: E0126 17:10:07.895657 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="98a8c733-ba31-46a3-97ea-409b69050b02" containerName="mariadb-account-create-update" Jan 26 17:10:07 crc kubenswrapper[4754]: I0126 17:10:07.895682 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="98a8c733-ba31-46a3-97ea-409b69050b02" containerName="mariadb-account-create-update" Jan 26 17:10:07 crc kubenswrapper[4754]: E0126 17:10:07.895699 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1597209c-abf0-4bd4-91b4-a8661e7e496b" containerName="mariadb-account-create-update" Jan 26 17:10:07 crc kubenswrapper[4754]: I0126 17:10:07.895707 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="1597209c-abf0-4bd4-91b4-a8661e7e496b" containerName="mariadb-account-create-update" Jan 26 17:10:07 crc kubenswrapper[4754]: E0126 17:10:07.895725 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7dc29a9d-f578-49ec-97a8-08a80bd3a2c0" containerName="extract-utilities" Jan 26 17:10:07 crc kubenswrapper[4754]: I0126 17:10:07.895733 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="7dc29a9d-f578-49ec-97a8-08a80bd3a2c0" containerName="extract-utilities" Jan 26 17:10:07 crc kubenswrapper[4754]: E0126 17:10:07.895749 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bef625bb-6bee-4c89-a9b4-4f695ab4ffd5" containerName="mariadb-account-create-update" Jan 26 17:10:07 crc kubenswrapper[4754]: I0126 17:10:07.895756 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="bef625bb-6bee-4c89-a9b4-4f695ab4ffd5" containerName="mariadb-account-create-update" Jan 26 17:10:07 crc kubenswrapper[4754]: E0126 17:10:07.895807 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="755ea18f-9963-409b-a3a2-56d110120dc8" containerName="mariadb-database-create" Jan 26 17:10:07 crc kubenswrapper[4754]: I0126 17:10:07.895817 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="755ea18f-9963-409b-a3a2-56d110120dc8" containerName="mariadb-database-create" Jan 26 17:10:07 crc kubenswrapper[4754]: E0126 17:10:07.895834 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5c367c8-74fd-4073-bbc6-2f6d4373f5ef" containerName="mariadb-database-create" Jan 26 17:10:07 crc kubenswrapper[4754]: I0126 17:10:07.895843 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5c367c8-74fd-4073-bbc6-2f6d4373f5ef" containerName="mariadb-database-create" Jan 26 17:10:07 crc kubenswrapper[4754]: I0126 17:10:07.896065 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="7dc29a9d-f578-49ec-97a8-08a80bd3a2c0" containerName="registry-server" Jan 26 17:10:07 crc kubenswrapper[4754]: I0126 17:10:07.896090 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="bef625bb-6bee-4c89-a9b4-4f695ab4ffd5" containerName="mariadb-account-create-update" Jan 26 17:10:07 crc kubenswrapper[4754]: I0126 17:10:07.896110 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="98a8c733-ba31-46a3-97ea-409b69050b02" containerName="mariadb-account-create-update" Jan 26 17:10:07 crc kubenswrapper[4754]: I0126 17:10:07.896130 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="1597209c-abf0-4bd4-91b4-a8661e7e496b" containerName="mariadb-account-create-update" Jan 26 17:10:07 crc kubenswrapper[4754]: I0126 17:10:07.896144 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="755ea18f-9963-409b-a3a2-56d110120dc8" containerName="mariadb-database-create" Jan 26 17:10:07 crc kubenswrapper[4754]: I0126 17:10:07.896160 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="cf071f6e-289f-4f86-9a39-e4dc62335b14" containerName="mariadb-database-create" Jan 26 17:10:07 crc kubenswrapper[4754]: I0126 17:10:07.896174 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="f5c367c8-74fd-4073-bbc6-2f6d4373f5ef" containerName="mariadb-database-create" Jan 26 17:10:07 crc kubenswrapper[4754]: I0126 17:10:07.896875 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-8lrb8" Jan 26 17:10:07 crc kubenswrapper[4754]: I0126 17:10:07.898696 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Jan 26 17:10:07 crc kubenswrapper[4754]: I0126 17:10:07.899323 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Jan 26 17:10:07 crc kubenswrapper[4754]: I0126 17:10:07.899370 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-rvgdm" Jan 26 17:10:07 crc kubenswrapper[4754]: I0126 17:10:07.905951 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/db76cab2-a25b-4072-a7c3-fcc0d6e5401a-config-data\") pod \"nova-cell0-conductor-db-sync-8lrb8\" (UID: \"db76cab2-a25b-4072-a7c3-fcc0d6e5401a\") " pod="openstack/nova-cell0-conductor-db-sync-8lrb8" Jan 26 17:10:07 crc kubenswrapper[4754]: I0126 17:10:07.906065 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db76cab2-a25b-4072-a7c3-fcc0d6e5401a-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-8lrb8\" (UID: \"db76cab2-a25b-4072-a7c3-fcc0d6e5401a\") " pod="openstack/nova-cell0-conductor-db-sync-8lrb8" Jan 26 17:10:07 crc kubenswrapper[4754]: I0126 17:10:07.906262 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/db76cab2-a25b-4072-a7c3-fcc0d6e5401a-scripts\") pod \"nova-cell0-conductor-db-sync-8lrb8\" (UID: \"db76cab2-a25b-4072-a7c3-fcc0d6e5401a\") " pod="openstack/nova-cell0-conductor-db-sync-8lrb8" Jan 26 17:10:07 crc kubenswrapper[4754]: I0126 17:10:07.906321 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xdrx4\" (UniqueName: \"kubernetes.io/projected/db76cab2-a25b-4072-a7c3-fcc0d6e5401a-kube-api-access-xdrx4\") pod \"nova-cell0-conductor-db-sync-8lrb8\" (UID: \"db76cab2-a25b-4072-a7c3-fcc0d6e5401a\") " pod="openstack/nova-cell0-conductor-db-sync-8lrb8" Jan 26 17:10:07 crc kubenswrapper[4754]: I0126 17:10:07.910525 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-8lrb8"] Jan 26 17:10:08 crc kubenswrapper[4754]: I0126 17:10:08.008158 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db76cab2-a25b-4072-a7c3-fcc0d6e5401a-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-8lrb8\" (UID: \"db76cab2-a25b-4072-a7c3-fcc0d6e5401a\") " pod="openstack/nova-cell0-conductor-db-sync-8lrb8" Jan 26 17:10:08 crc kubenswrapper[4754]: I0126 17:10:08.008388 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/db76cab2-a25b-4072-a7c3-fcc0d6e5401a-scripts\") pod \"nova-cell0-conductor-db-sync-8lrb8\" (UID: \"db76cab2-a25b-4072-a7c3-fcc0d6e5401a\") " pod="openstack/nova-cell0-conductor-db-sync-8lrb8" Jan 26 17:10:08 crc kubenswrapper[4754]: I0126 17:10:08.008428 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xdrx4\" (UniqueName: \"kubernetes.io/projected/db76cab2-a25b-4072-a7c3-fcc0d6e5401a-kube-api-access-xdrx4\") pod \"nova-cell0-conductor-db-sync-8lrb8\" (UID: \"db76cab2-a25b-4072-a7c3-fcc0d6e5401a\") " pod="openstack/nova-cell0-conductor-db-sync-8lrb8" Jan 26 17:10:08 crc kubenswrapper[4754]: I0126 17:10:08.008584 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/db76cab2-a25b-4072-a7c3-fcc0d6e5401a-config-data\") pod \"nova-cell0-conductor-db-sync-8lrb8\" (UID: \"db76cab2-a25b-4072-a7c3-fcc0d6e5401a\") " pod="openstack/nova-cell0-conductor-db-sync-8lrb8" Jan 26 17:10:08 crc kubenswrapper[4754]: I0126 17:10:08.014578 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/db76cab2-a25b-4072-a7c3-fcc0d6e5401a-scripts\") pod \"nova-cell0-conductor-db-sync-8lrb8\" (UID: \"db76cab2-a25b-4072-a7c3-fcc0d6e5401a\") " pod="openstack/nova-cell0-conductor-db-sync-8lrb8" Jan 26 17:10:08 crc kubenswrapper[4754]: I0126 17:10:08.026828 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/db76cab2-a25b-4072-a7c3-fcc0d6e5401a-config-data\") pod \"nova-cell0-conductor-db-sync-8lrb8\" (UID: \"db76cab2-a25b-4072-a7c3-fcc0d6e5401a\") " pod="openstack/nova-cell0-conductor-db-sync-8lrb8" Jan 26 17:10:08 crc kubenswrapper[4754]: I0126 17:10:08.026827 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db76cab2-a25b-4072-a7c3-fcc0d6e5401a-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-8lrb8\" (UID: \"db76cab2-a25b-4072-a7c3-fcc0d6e5401a\") " pod="openstack/nova-cell0-conductor-db-sync-8lrb8" Jan 26 17:10:08 crc kubenswrapper[4754]: I0126 17:10:08.031133 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xdrx4\" (UniqueName: \"kubernetes.io/projected/db76cab2-a25b-4072-a7c3-fcc0d6e5401a-kube-api-access-xdrx4\") pod \"nova-cell0-conductor-db-sync-8lrb8\" (UID: \"db76cab2-a25b-4072-a7c3-fcc0d6e5401a\") " pod="openstack/nova-cell0-conductor-db-sync-8lrb8" Jan 26 17:10:08 crc kubenswrapper[4754]: I0126 17:10:08.214212 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-8lrb8" Jan 26 17:10:08 crc kubenswrapper[4754]: I0126 17:10:08.338152 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" event={"ID":"8c3718a4-f354-4284-92e0-fdfb45a692bd","Type":"ContainerStarted","Data":"19eac0e87afaeef93088a8556aa836389e2ebff93a96b482df571cb7b852f0b7"} Jan 26 17:10:08 crc kubenswrapper[4754]: I0126 17:10:08.691925 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-8lrb8"] Jan 26 17:10:08 crc kubenswrapper[4754]: W0126 17:10:08.693229 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddb76cab2_a25b_4072_a7c3_fcc0d6e5401a.slice/crio-d79a7852225bf0cd5e35681bf2e35471e95134c2f4b0194cbc7d0f3bc5b271c6 WatchSource:0}: Error finding container d79a7852225bf0cd5e35681bf2e35471e95134c2f4b0194cbc7d0f3bc5b271c6: Status 404 returned error can't find the container with id d79a7852225bf0cd5e35681bf2e35471e95134c2f4b0194cbc7d0f3bc5b271c6 Jan 26 17:10:09 crc kubenswrapper[4754]: I0126 17:10:09.346745 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-8lrb8" event={"ID":"db76cab2-a25b-4072-a7c3-fcc0d6e5401a","Type":"ContainerStarted","Data":"d79a7852225bf0cd5e35681bf2e35471e95134c2f4b0194cbc7d0f3bc5b271c6"} Jan 26 17:10:14 crc kubenswrapper[4754]: I0126 17:10:14.756833 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-55c9b8478b-6brjv" Jan 26 17:10:17 crc kubenswrapper[4754]: I0126 17:10:17.050678 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-767b9f4849-cfqbz" Jan 26 17:10:17 crc kubenswrapper[4754]: I0126 17:10:17.123169 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-55c9b8478b-6brjv"] Jan 26 17:10:17 crc kubenswrapper[4754]: I0126 17:10:17.123400 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-55c9b8478b-6brjv" podUID="deac50d1-deea-4092-b2a3-b3a7ada97a42" containerName="neutron-api" containerID="cri-o://330d0f3072f01682231689eb69fa008dec465696c9b685a3f2ea9108f318cb1f" gracePeriod=30 Jan 26 17:10:17 crc kubenswrapper[4754]: I0126 17:10:17.123457 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-55c9b8478b-6brjv" podUID="deac50d1-deea-4092-b2a3-b3a7ada97a42" containerName="neutron-httpd" containerID="cri-o://236452eb8f2e8caab9a57c5cf6bb43e5b06a51eb5120a207826a17214da8fbb2" gracePeriod=30 Jan 26 17:10:17 crc kubenswrapper[4754]: I0126 17:10:17.425203 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-8lrb8" event={"ID":"db76cab2-a25b-4072-a7c3-fcc0d6e5401a","Type":"ContainerStarted","Data":"62265a25a1e473fa628a59ad5be6830d0677bee3af7afe2a0e8f64ed52f241f3"} Jan 26 17:10:17 crc kubenswrapper[4754]: I0126 17:10:17.446883 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-8lrb8" podStartSLOduration=2.4366075609999998 podStartE2EDuration="10.446865205s" podCreationTimestamp="2026-01-26 17:10:07 +0000 UTC" firstStartedPulling="2026-01-26 17:10:08.6948504 +0000 UTC m=+1375.219030834" lastFinishedPulling="2026-01-26 17:10:16.705108044 +0000 UTC m=+1383.229288478" observedRunningTime="2026-01-26 17:10:17.438120664 +0000 UTC m=+1383.962301138" watchObservedRunningTime="2026-01-26 17:10:17.446865205 +0000 UTC m=+1383.971045639" Jan 26 17:10:18 crc kubenswrapper[4754]: I0126 17:10:18.425121 4754 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="ee49bc9b-edbd-4b7b-b977-98ec0b6b21c4" containerName="proxy-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 503" Jan 26 17:10:18 crc kubenswrapper[4754]: I0126 17:10:18.436279 4754 generic.go:334] "Generic (PLEG): container finished" podID="deac50d1-deea-4092-b2a3-b3a7ada97a42" containerID="236452eb8f2e8caab9a57c5cf6bb43e5b06a51eb5120a207826a17214da8fbb2" exitCode=0 Jan 26 17:10:18 crc kubenswrapper[4754]: I0126 17:10:18.436343 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-55c9b8478b-6brjv" event={"ID":"deac50d1-deea-4092-b2a3-b3a7ada97a42","Type":"ContainerDied","Data":"236452eb8f2e8caab9a57c5cf6bb43e5b06a51eb5120a207826a17214da8fbb2"} Jan 26 17:10:26 crc kubenswrapper[4754]: E0126 17:10:26.343606 4754 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podee49bc9b_edbd_4b7b_b977_98ec0b6b21c4.slice/crio-conmon-dd10cb6e3eb2254b305e23513953ba8c1131ba1298169e69cdab441ae8a2ecc8.scope\": RecentStats: unable to find data in memory cache]" Jan 26 17:10:26 crc kubenswrapper[4754]: I0126 17:10:26.507181 4754 generic.go:334] "Generic (PLEG): container finished" podID="ee49bc9b-edbd-4b7b-b977-98ec0b6b21c4" containerID="dd10cb6e3eb2254b305e23513953ba8c1131ba1298169e69cdab441ae8a2ecc8" exitCode=137 Jan 26 17:10:26 crc kubenswrapper[4754]: I0126 17:10:26.507233 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ee49bc9b-edbd-4b7b-b977-98ec0b6b21c4","Type":"ContainerDied","Data":"dd10cb6e3eb2254b305e23513953ba8c1131ba1298169e69cdab441ae8a2ecc8"} Jan 26 17:10:27 crc kubenswrapper[4754]: I0126 17:10:27.107099 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 26 17:10:27 crc kubenswrapper[4754]: I0126 17:10:27.212039 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ee49bc9b-edbd-4b7b-b977-98ec0b6b21c4-run-httpd\") pod \"ee49bc9b-edbd-4b7b-b977-98ec0b6b21c4\" (UID: \"ee49bc9b-edbd-4b7b-b977-98ec0b6b21c4\") " Jan 26 17:10:27 crc kubenswrapper[4754]: I0126 17:10:27.212081 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ee49bc9b-edbd-4b7b-b977-98ec0b6b21c4-scripts\") pod \"ee49bc9b-edbd-4b7b-b977-98ec0b6b21c4\" (UID: \"ee49bc9b-edbd-4b7b-b977-98ec0b6b21c4\") " Jan 26 17:10:27 crc kubenswrapper[4754]: I0126 17:10:27.212110 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xl995\" (UniqueName: \"kubernetes.io/projected/ee49bc9b-edbd-4b7b-b977-98ec0b6b21c4-kube-api-access-xl995\") pod \"ee49bc9b-edbd-4b7b-b977-98ec0b6b21c4\" (UID: \"ee49bc9b-edbd-4b7b-b977-98ec0b6b21c4\") " Jan 26 17:10:27 crc kubenswrapper[4754]: I0126 17:10:27.212145 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ee49bc9b-edbd-4b7b-b977-98ec0b6b21c4-log-httpd\") pod \"ee49bc9b-edbd-4b7b-b977-98ec0b6b21c4\" (UID: \"ee49bc9b-edbd-4b7b-b977-98ec0b6b21c4\") " Jan 26 17:10:27 crc kubenswrapper[4754]: I0126 17:10:27.212211 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee49bc9b-edbd-4b7b-b977-98ec0b6b21c4-combined-ca-bundle\") pod \"ee49bc9b-edbd-4b7b-b977-98ec0b6b21c4\" (UID: \"ee49bc9b-edbd-4b7b-b977-98ec0b6b21c4\") " Jan 26 17:10:27 crc kubenswrapper[4754]: I0126 17:10:27.212253 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ee49bc9b-edbd-4b7b-b977-98ec0b6b21c4-sg-core-conf-yaml\") pod \"ee49bc9b-edbd-4b7b-b977-98ec0b6b21c4\" (UID: \"ee49bc9b-edbd-4b7b-b977-98ec0b6b21c4\") " Jan 26 17:10:27 crc kubenswrapper[4754]: I0126 17:10:27.212376 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ee49bc9b-edbd-4b7b-b977-98ec0b6b21c4-config-data\") pod \"ee49bc9b-edbd-4b7b-b977-98ec0b6b21c4\" (UID: \"ee49bc9b-edbd-4b7b-b977-98ec0b6b21c4\") " Jan 26 17:10:27 crc kubenswrapper[4754]: I0126 17:10:27.212752 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ee49bc9b-edbd-4b7b-b977-98ec0b6b21c4-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "ee49bc9b-edbd-4b7b-b977-98ec0b6b21c4" (UID: "ee49bc9b-edbd-4b7b-b977-98ec0b6b21c4"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:10:27 crc kubenswrapper[4754]: I0126 17:10:27.213581 4754 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ee49bc9b-edbd-4b7b-b977-98ec0b6b21c4-run-httpd\") on node \"crc\" DevicePath \"\"" Jan 26 17:10:27 crc kubenswrapper[4754]: I0126 17:10:27.213589 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ee49bc9b-edbd-4b7b-b977-98ec0b6b21c4-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "ee49bc9b-edbd-4b7b-b977-98ec0b6b21c4" (UID: "ee49bc9b-edbd-4b7b-b977-98ec0b6b21c4"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:10:27 crc kubenswrapper[4754]: I0126 17:10:27.218182 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee49bc9b-edbd-4b7b-b977-98ec0b6b21c4-scripts" (OuterVolumeSpecName: "scripts") pod "ee49bc9b-edbd-4b7b-b977-98ec0b6b21c4" (UID: "ee49bc9b-edbd-4b7b-b977-98ec0b6b21c4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:10:27 crc kubenswrapper[4754]: I0126 17:10:27.233484 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ee49bc9b-edbd-4b7b-b977-98ec0b6b21c4-kube-api-access-xl995" (OuterVolumeSpecName: "kube-api-access-xl995") pod "ee49bc9b-edbd-4b7b-b977-98ec0b6b21c4" (UID: "ee49bc9b-edbd-4b7b-b977-98ec0b6b21c4"). InnerVolumeSpecName "kube-api-access-xl995". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:10:27 crc kubenswrapper[4754]: I0126 17:10:27.258529 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee49bc9b-edbd-4b7b-b977-98ec0b6b21c4-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "ee49bc9b-edbd-4b7b-b977-98ec0b6b21c4" (UID: "ee49bc9b-edbd-4b7b-b977-98ec0b6b21c4"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:10:27 crc kubenswrapper[4754]: I0126 17:10:27.309960 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee49bc9b-edbd-4b7b-b977-98ec0b6b21c4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ee49bc9b-edbd-4b7b-b977-98ec0b6b21c4" (UID: "ee49bc9b-edbd-4b7b-b977-98ec0b6b21c4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:10:27 crc kubenswrapper[4754]: I0126 17:10:27.315862 4754 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ee49bc9b-edbd-4b7b-b977-98ec0b6b21c4-scripts\") on node \"crc\" DevicePath \"\"" Jan 26 17:10:27 crc kubenswrapper[4754]: I0126 17:10:27.315899 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xl995\" (UniqueName: \"kubernetes.io/projected/ee49bc9b-edbd-4b7b-b977-98ec0b6b21c4-kube-api-access-xl995\") on node \"crc\" DevicePath \"\"" Jan 26 17:10:27 crc kubenswrapper[4754]: I0126 17:10:27.315911 4754 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ee49bc9b-edbd-4b7b-b977-98ec0b6b21c4-log-httpd\") on node \"crc\" DevicePath \"\"" Jan 26 17:10:27 crc kubenswrapper[4754]: I0126 17:10:27.315922 4754 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee49bc9b-edbd-4b7b-b977-98ec0b6b21c4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 26 17:10:27 crc kubenswrapper[4754]: I0126 17:10:27.315932 4754 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ee49bc9b-edbd-4b7b-b977-98ec0b6b21c4-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Jan 26 17:10:27 crc kubenswrapper[4754]: I0126 17:10:27.328821 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee49bc9b-edbd-4b7b-b977-98ec0b6b21c4-config-data" (OuterVolumeSpecName: "config-data") pod "ee49bc9b-edbd-4b7b-b977-98ec0b6b21c4" (UID: "ee49bc9b-edbd-4b7b-b977-98ec0b6b21c4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:10:27 crc kubenswrapper[4754]: I0126 17:10:27.417598 4754 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ee49bc9b-edbd-4b7b-b977-98ec0b6b21c4-config-data\") on node \"crc\" DevicePath \"\"" Jan 26 17:10:27 crc kubenswrapper[4754]: I0126 17:10:27.518881 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ee49bc9b-edbd-4b7b-b977-98ec0b6b21c4","Type":"ContainerDied","Data":"a1e64314e99fdb6278c7c0f9380f49df9560203140d09ed1fa445b0413789260"} Jan 26 17:10:27 crc kubenswrapper[4754]: I0126 17:10:27.518938 4754 scope.go:117] "RemoveContainer" containerID="dd10cb6e3eb2254b305e23513953ba8c1131ba1298169e69cdab441ae8a2ecc8" Jan 26 17:10:27 crc kubenswrapper[4754]: I0126 17:10:27.518960 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 26 17:10:27 crc kubenswrapper[4754]: I0126 17:10:27.521454 4754 generic.go:334] "Generic (PLEG): container finished" podID="deac50d1-deea-4092-b2a3-b3a7ada97a42" containerID="330d0f3072f01682231689eb69fa008dec465696c9b685a3f2ea9108f318cb1f" exitCode=0 Jan 26 17:10:27 crc kubenswrapper[4754]: I0126 17:10:27.521501 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-55c9b8478b-6brjv" event={"ID":"deac50d1-deea-4092-b2a3-b3a7ada97a42","Type":"ContainerDied","Data":"330d0f3072f01682231689eb69fa008dec465696c9b685a3f2ea9108f318cb1f"} Jan 26 17:10:27 crc kubenswrapper[4754]: I0126 17:10:27.574437 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Jan 26 17:10:27 crc kubenswrapper[4754]: I0126 17:10:27.576461 4754 scope.go:117] "RemoveContainer" containerID="d87184c817a1540d791103f965b1f40e2b84afbbfe054dc48e127f244e9911be" Jan 26 17:10:27 crc kubenswrapper[4754]: I0126 17:10:27.585778 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Jan 26 17:10:27 crc kubenswrapper[4754]: I0126 17:10:27.602102 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Jan 26 17:10:27 crc kubenswrapper[4754]: E0126 17:10:27.602611 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee49bc9b-edbd-4b7b-b977-98ec0b6b21c4" containerName="sg-core" Jan 26 17:10:27 crc kubenswrapper[4754]: I0126 17:10:27.602708 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee49bc9b-edbd-4b7b-b977-98ec0b6b21c4" containerName="sg-core" Jan 26 17:10:27 crc kubenswrapper[4754]: E0126 17:10:27.602792 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee49bc9b-edbd-4b7b-b977-98ec0b6b21c4" containerName="ceilometer-notification-agent" Jan 26 17:10:27 crc kubenswrapper[4754]: I0126 17:10:27.602848 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee49bc9b-edbd-4b7b-b977-98ec0b6b21c4" containerName="ceilometer-notification-agent" Jan 26 17:10:27 crc kubenswrapper[4754]: E0126 17:10:27.602906 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee49bc9b-edbd-4b7b-b977-98ec0b6b21c4" containerName="proxy-httpd" Jan 26 17:10:27 crc kubenswrapper[4754]: I0126 17:10:27.602955 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee49bc9b-edbd-4b7b-b977-98ec0b6b21c4" containerName="proxy-httpd" Jan 26 17:10:27 crc kubenswrapper[4754]: E0126 17:10:27.603016 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee49bc9b-edbd-4b7b-b977-98ec0b6b21c4" containerName="ceilometer-central-agent" Jan 26 17:10:27 crc kubenswrapper[4754]: I0126 17:10:27.603069 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee49bc9b-edbd-4b7b-b977-98ec0b6b21c4" containerName="ceilometer-central-agent" Jan 26 17:10:27 crc kubenswrapper[4754]: I0126 17:10:27.603319 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee49bc9b-edbd-4b7b-b977-98ec0b6b21c4" containerName="ceilometer-notification-agent" Jan 26 17:10:27 crc kubenswrapper[4754]: I0126 17:10:27.603400 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee49bc9b-edbd-4b7b-b977-98ec0b6b21c4" containerName="proxy-httpd" Jan 26 17:10:27 crc kubenswrapper[4754]: I0126 17:10:27.603456 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee49bc9b-edbd-4b7b-b977-98ec0b6b21c4" containerName="ceilometer-central-agent" Jan 26 17:10:27 crc kubenswrapper[4754]: I0126 17:10:27.603514 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee49bc9b-edbd-4b7b-b977-98ec0b6b21c4" containerName="sg-core" Jan 26 17:10:27 crc kubenswrapper[4754]: I0126 17:10:27.605137 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 26 17:10:27 crc kubenswrapper[4754]: I0126 17:10:27.607593 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Jan 26 17:10:27 crc kubenswrapper[4754]: I0126 17:10:27.608731 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Jan 26 17:10:27 crc kubenswrapper[4754]: I0126 17:10:27.610987 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Jan 26 17:10:27 crc kubenswrapper[4754]: I0126 17:10:27.619925 4754 scope.go:117] "RemoveContainer" containerID="09cad47c4a2f836bde7a8e847c73a4882a25b0a8bab6d861ac184fbd95f15505" Jan 26 17:10:27 crc kubenswrapper[4754]: I0126 17:10:27.642063 4754 scope.go:117] "RemoveContainer" containerID="e350d7d898f163921729bb47f1b2fd06ba49d5a07efcd6a4f1fa7505a6a47da7" Jan 26 17:10:27 crc kubenswrapper[4754]: I0126 17:10:27.723535 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6d0563ff-f0f4-45b7-aaba-014f4426f4b0-log-httpd\") pod \"ceilometer-0\" (UID: \"6d0563ff-f0f4-45b7-aaba-014f4426f4b0\") " pod="openstack/ceilometer-0" Jan 26 17:10:27 crc kubenswrapper[4754]: I0126 17:10:27.723588 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5g8db\" (UniqueName: \"kubernetes.io/projected/6d0563ff-f0f4-45b7-aaba-014f4426f4b0-kube-api-access-5g8db\") pod \"ceilometer-0\" (UID: \"6d0563ff-f0f4-45b7-aaba-014f4426f4b0\") " pod="openstack/ceilometer-0" Jan 26 17:10:27 crc kubenswrapper[4754]: I0126 17:10:27.723642 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6d0563ff-f0f4-45b7-aaba-014f4426f4b0-scripts\") pod \"ceilometer-0\" (UID: \"6d0563ff-f0f4-45b7-aaba-014f4426f4b0\") " pod="openstack/ceilometer-0" Jan 26 17:10:27 crc kubenswrapper[4754]: I0126 17:10:27.723937 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d0563ff-f0f4-45b7-aaba-014f4426f4b0-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"6d0563ff-f0f4-45b7-aaba-014f4426f4b0\") " pod="openstack/ceilometer-0" Jan 26 17:10:27 crc kubenswrapper[4754]: I0126 17:10:27.723974 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6d0563ff-f0f4-45b7-aaba-014f4426f4b0-run-httpd\") pod \"ceilometer-0\" (UID: \"6d0563ff-f0f4-45b7-aaba-014f4426f4b0\") " pod="openstack/ceilometer-0" Jan 26 17:10:27 crc kubenswrapper[4754]: I0126 17:10:27.724013 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6d0563ff-f0f4-45b7-aaba-014f4426f4b0-config-data\") pod \"ceilometer-0\" (UID: \"6d0563ff-f0f4-45b7-aaba-014f4426f4b0\") " pod="openstack/ceilometer-0" Jan 26 17:10:27 crc kubenswrapper[4754]: I0126 17:10:27.724049 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/6d0563ff-f0f4-45b7-aaba-014f4426f4b0-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"6d0563ff-f0f4-45b7-aaba-014f4426f4b0\") " pod="openstack/ceilometer-0" Jan 26 17:10:27 crc kubenswrapper[4754]: I0126 17:10:27.777531 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ee49bc9b-edbd-4b7b-b977-98ec0b6b21c4" path="/var/lib/kubelet/pods/ee49bc9b-edbd-4b7b-b977-98ec0b6b21c4/volumes" Jan 26 17:10:27 crc kubenswrapper[4754]: I0126 17:10:27.825204 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d0563ff-f0f4-45b7-aaba-014f4426f4b0-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"6d0563ff-f0f4-45b7-aaba-014f4426f4b0\") " pod="openstack/ceilometer-0" Jan 26 17:10:27 crc kubenswrapper[4754]: I0126 17:10:27.825256 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6d0563ff-f0f4-45b7-aaba-014f4426f4b0-run-httpd\") pod \"ceilometer-0\" (UID: \"6d0563ff-f0f4-45b7-aaba-014f4426f4b0\") " pod="openstack/ceilometer-0" Jan 26 17:10:27 crc kubenswrapper[4754]: I0126 17:10:27.825274 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6d0563ff-f0f4-45b7-aaba-014f4426f4b0-config-data\") pod \"ceilometer-0\" (UID: \"6d0563ff-f0f4-45b7-aaba-014f4426f4b0\") " pod="openstack/ceilometer-0" Jan 26 17:10:27 crc kubenswrapper[4754]: I0126 17:10:27.825298 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/6d0563ff-f0f4-45b7-aaba-014f4426f4b0-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"6d0563ff-f0f4-45b7-aaba-014f4426f4b0\") " pod="openstack/ceilometer-0" Jan 26 17:10:27 crc kubenswrapper[4754]: I0126 17:10:27.825350 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6d0563ff-f0f4-45b7-aaba-014f4426f4b0-log-httpd\") pod \"ceilometer-0\" (UID: \"6d0563ff-f0f4-45b7-aaba-014f4426f4b0\") " pod="openstack/ceilometer-0" Jan 26 17:10:27 crc kubenswrapper[4754]: I0126 17:10:27.825369 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5g8db\" (UniqueName: \"kubernetes.io/projected/6d0563ff-f0f4-45b7-aaba-014f4426f4b0-kube-api-access-5g8db\") pod \"ceilometer-0\" (UID: \"6d0563ff-f0f4-45b7-aaba-014f4426f4b0\") " pod="openstack/ceilometer-0" Jan 26 17:10:27 crc kubenswrapper[4754]: I0126 17:10:27.825392 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6d0563ff-f0f4-45b7-aaba-014f4426f4b0-scripts\") pod \"ceilometer-0\" (UID: \"6d0563ff-f0f4-45b7-aaba-014f4426f4b0\") " pod="openstack/ceilometer-0" Jan 26 17:10:27 crc kubenswrapper[4754]: I0126 17:10:27.826393 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6d0563ff-f0f4-45b7-aaba-014f4426f4b0-run-httpd\") pod \"ceilometer-0\" (UID: \"6d0563ff-f0f4-45b7-aaba-014f4426f4b0\") " pod="openstack/ceilometer-0" Jan 26 17:10:27 crc kubenswrapper[4754]: I0126 17:10:27.827293 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6d0563ff-f0f4-45b7-aaba-014f4426f4b0-log-httpd\") pod \"ceilometer-0\" (UID: \"6d0563ff-f0f4-45b7-aaba-014f4426f4b0\") " pod="openstack/ceilometer-0" Jan 26 17:10:27 crc kubenswrapper[4754]: I0126 17:10:27.829482 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d0563ff-f0f4-45b7-aaba-014f4426f4b0-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"6d0563ff-f0f4-45b7-aaba-014f4426f4b0\") " pod="openstack/ceilometer-0" Jan 26 17:10:27 crc kubenswrapper[4754]: I0126 17:10:27.831710 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6d0563ff-f0f4-45b7-aaba-014f4426f4b0-scripts\") pod \"ceilometer-0\" (UID: \"6d0563ff-f0f4-45b7-aaba-014f4426f4b0\") " pod="openstack/ceilometer-0" Jan 26 17:10:27 crc kubenswrapper[4754]: I0126 17:10:27.832186 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6d0563ff-f0f4-45b7-aaba-014f4426f4b0-config-data\") pod \"ceilometer-0\" (UID: \"6d0563ff-f0f4-45b7-aaba-014f4426f4b0\") " pod="openstack/ceilometer-0" Jan 26 17:10:27 crc kubenswrapper[4754]: I0126 17:10:27.840280 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/6d0563ff-f0f4-45b7-aaba-014f4426f4b0-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"6d0563ff-f0f4-45b7-aaba-014f4426f4b0\") " pod="openstack/ceilometer-0" Jan 26 17:10:27 crc kubenswrapper[4754]: I0126 17:10:27.849233 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5g8db\" (UniqueName: \"kubernetes.io/projected/6d0563ff-f0f4-45b7-aaba-014f4426f4b0-kube-api-access-5g8db\") pod \"ceilometer-0\" (UID: \"6d0563ff-f0f4-45b7-aaba-014f4426f4b0\") " pod="openstack/ceilometer-0" Jan 26 17:10:27 crc kubenswrapper[4754]: I0126 17:10:27.941054 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 26 17:10:27 crc kubenswrapper[4754]: I0126 17:10:27.970969 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-55c9b8478b-6brjv" Jan 26 17:10:28 crc kubenswrapper[4754]: I0126 17:10:28.028348 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/deac50d1-deea-4092-b2a3-b3a7ada97a42-combined-ca-bundle\") pod \"deac50d1-deea-4092-b2a3-b3a7ada97a42\" (UID: \"deac50d1-deea-4092-b2a3-b3a7ada97a42\") " Jan 26 17:10:28 crc kubenswrapper[4754]: I0126 17:10:28.028485 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-28mh4\" (UniqueName: \"kubernetes.io/projected/deac50d1-deea-4092-b2a3-b3a7ada97a42-kube-api-access-28mh4\") pod \"deac50d1-deea-4092-b2a3-b3a7ada97a42\" (UID: \"deac50d1-deea-4092-b2a3-b3a7ada97a42\") " Jan 26 17:10:28 crc kubenswrapper[4754]: I0126 17:10:28.028532 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/deac50d1-deea-4092-b2a3-b3a7ada97a42-httpd-config\") pod \"deac50d1-deea-4092-b2a3-b3a7ada97a42\" (UID: \"deac50d1-deea-4092-b2a3-b3a7ada97a42\") " Jan 26 17:10:28 crc kubenswrapper[4754]: I0126 17:10:28.028586 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/deac50d1-deea-4092-b2a3-b3a7ada97a42-ovndb-tls-certs\") pod \"deac50d1-deea-4092-b2a3-b3a7ada97a42\" (UID: \"deac50d1-deea-4092-b2a3-b3a7ada97a42\") " Jan 26 17:10:28 crc kubenswrapper[4754]: I0126 17:10:28.028612 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/deac50d1-deea-4092-b2a3-b3a7ada97a42-config\") pod \"deac50d1-deea-4092-b2a3-b3a7ada97a42\" (UID: \"deac50d1-deea-4092-b2a3-b3a7ada97a42\") " Jan 26 17:10:28 crc kubenswrapper[4754]: I0126 17:10:28.033837 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/deac50d1-deea-4092-b2a3-b3a7ada97a42-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "deac50d1-deea-4092-b2a3-b3a7ada97a42" (UID: "deac50d1-deea-4092-b2a3-b3a7ada97a42"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:10:28 crc kubenswrapper[4754]: I0126 17:10:28.033847 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/deac50d1-deea-4092-b2a3-b3a7ada97a42-kube-api-access-28mh4" (OuterVolumeSpecName: "kube-api-access-28mh4") pod "deac50d1-deea-4092-b2a3-b3a7ada97a42" (UID: "deac50d1-deea-4092-b2a3-b3a7ada97a42"). InnerVolumeSpecName "kube-api-access-28mh4". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:10:28 crc kubenswrapper[4754]: I0126 17:10:28.091857 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/deac50d1-deea-4092-b2a3-b3a7ada97a42-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "deac50d1-deea-4092-b2a3-b3a7ada97a42" (UID: "deac50d1-deea-4092-b2a3-b3a7ada97a42"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:10:28 crc kubenswrapper[4754]: I0126 17:10:28.095844 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/deac50d1-deea-4092-b2a3-b3a7ada97a42-config" (OuterVolumeSpecName: "config") pod "deac50d1-deea-4092-b2a3-b3a7ada97a42" (UID: "deac50d1-deea-4092-b2a3-b3a7ada97a42"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:10:28 crc kubenswrapper[4754]: I0126 17:10:28.105878 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/deac50d1-deea-4092-b2a3-b3a7ada97a42-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "deac50d1-deea-4092-b2a3-b3a7ada97a42" (UID: "deac50d1-deea-4092-b2a3-b3a7ada97a42"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:10:28 crc kubenswrapper[4754]: I0126 17:10:28.131090 4754 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/deac50d1-deea-4092-b2a3-b3a7ada97a42-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 26 17:10:28 crc kubenswrapper[4754]: I0126 17:10:28.131130 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-28mh4\" (UniqueName: \"kubernetes.io/projected/deac50d1-deea-4092-b2a3-b3a7ada97a42-kube-api-access-28mh4\") on node \"crc\" DevicePath \"\"" Jan 26 17:10:28 crc kubenswrapper[4754]: I0126 17:10:28.131144 4754 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/deac50d1-deea-4092-b2a3-b3a7ada97a42-httpd-config\") on node \"crc\" DevicePath \"\"" Jan 26 17:10:28 crc kubenswrapper[4754]: I0126 17:10:28.131155 4754 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/deac50d1-deea-4092-b2a3-b3a7ada97a42-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Jan 26 17:10:28 crc kubenswrapper[4754]: I0126 17:10:28.131166 4754 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/deac50d1-deea-4092-b2a3-b3a7ada97a42-config\") on node \"crc\" DevicePath \"\"" Jan 26 17:10:28 crc kubenswrapper[4754]: W0126 17:10:28.420270 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6d0563ff_f0f4_45b7_aaba_014f4426f4b0.slice/crio-01491f02574e1817229e56b88ccabe69e44826d913624b64f6546ef1af57aee7 WatchSource:0}: Error finding container 01491f02574e1817229e56b88ccabe69e44826d913624b64f6546ef1af57aee7: Status 404 returned error can't find the container with id 01491f02574e1817229e56b88ccabe69e44826d913624b64f6546ef1af57aee7 Jan 26 17:10:28 crc kubenswrapper[4754]: I0126 17:10:28.426103 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Jan 26 17:10:28 crc kubenswrapper[4754]: I0126 17:10:28.532565 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6d0563ff-f0f4-45b7-aaba-014f4426f4b0","Type":"ContainerStarted","Data":"01491f02574e1817229e56b88ccabe69e44826d913624b64f6546ef1af57aee7"} Jan 26 17:10:28 crc kubenswrapper[4754]: I0126 17:10:28.535031 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-55c9b8478b-6brjv" Jan 26 17:10:28 crc kubenswrapper[4754]: I0126 17:10:28.535032 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-55c9b8478b-6brjv" event={"ID":"deac50d1-deea-4092-b2a3-b3a7ada97a42","Type":"ContainerDied","Data":"a78fa0f03066fa45e5ac9bd4ab5ab92bd6fda65a61a2b81b357d3be6c6bf4522"} Jan 26 17:10:28 crc kubenswrapper[4754]: I0126 17:10:28.535113 4754 scope.go:117] "RemoveContainer" containerID="236452eb8f2e8caab9a57c5cf6bb43e5b06a51eb5120a207826a17214da8fbb2" Jan 26 17:10:28 crc kubenswrapper[4754]: I0126 17:10:28.559515 4754 scope.go:117] "RemoveContainer" containerID="330d0f3072f01682231689eb69fa008dec465696c9b685a3f2ea9108f318cb1f" Jan 26 17:10:28 crc kubenswrapper[4754]: I0126 17:10:28.571266 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-55c9b8478b-6brjv"] Jan 26 17:10:28 crc kubenswrapper[4754]: I0126 17:10:28.580934 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-55c9b8478b-6brjv"] Jan 26 17:10:29 crc kubenswrapper[4754]: I0126 17:10:29.781632 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="deac50d1-deea-4092-b2a3-b3a7ada97a42" path="/var/lib/kubelet/pods/deac50d1-deea-4092-b2a3-b3a7ada97a42/volumes" Jan 26 17:10:30 crc kubenswrapper[4754]: I0126 17:10:30.566496 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6d0563ff-f0f4-45b7-aaba-014f4426f4b0","Type":"ContainerStarted","Data":"eb80b00f37f12db37a6f667c832c82d481c8513f0c10b367ce2bb2cb977c988b"} Jan 26 17:10:31 crc kubenswrapper[4754]: I0126 17:10:31.578305 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6d0563ff-f0f4-45b7-aaba-014f4426f4b0","Type":"ContainerStarted","Data":"234f898b8e364e60561ed34cd7475fe072e0bd9490b1145b1909804af99b448d"} Jan 26 17:10:32 crc kubenswrapper[4754]: I0126 17:10:32.589155 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6d0563ff-f0f4-45b7-aaba-014f4426f4b0","Type":"ContainerStarted","Data":"2450afc5412f20e482cf129b8e70df60fec15191d846f728199beaa14ae587fe"} Jan 26 17:10:35 crc kubenswrapper[4754]: I0126 17:10:35.616596 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6d0563ff-f0f4-45b7-aaba-014f4426f4b0","Type":"ContainerStarted","Data":"ebd4a3041094da9f63be25f198192702e8ece75b22e3c5c564ca86dfb8a0148c"} Jan 26 17:10:35 crc kubenswrapper[4754]: I0126 17:10:35.617153 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Jan 26 17:10:35 crc kubenswrapper[4754]: I0126 17:10:35.664384 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.051744603 podStartE2EDuration="8.664354959s" podCreationTimestamp="2026-01-26 17:10:27 +0000 UTC" firstStartedPulling="2026-01-26 17:10:28.422437 +0000 UTC m=+1394.946617454" lastFinishedPulling="2026-01-26 17:10:35.035047376 +0000 UTC m=+1401.559227810" observedRunningTime="2026-01-26 17:10:35.644935955 +0000 UTC m=+1402.169116409" watchObservedRunningTime="2026-01-26 17:10:35.664354959 +0000 UTC m=+1402.188535393" Jan 26 17:10:40 crc kubenswrapper[4754]: I0126 17:10:40.654347 4754 generic.go:334] "Generic (PLEG): container finished" podID="db76cab2-a25b-4072-a7c3-fcc0d6e5401a" containerID="62265a25a1e473fa628a59ad5be6830d0677bee3af7afe2a0e8f64ed52f241f3" exitCode=0 Jan 26 17:10:40 crc kubenswrapper[4754]: I0126 17:10:40.654443 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-8lrb8" event={"ID":"db76cab2-a25b-4072-a7c3-fcc0d6e5401a","Type":"ContainerDied","Data":"62265a25a1e473fa628a59ad5be6830d0677bee3af7afe2a0e8f64ed52f241f3"} Jan 26 17:10:41 crc kubenswrapper[4754]: I0126 17:10:41.988287 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-8lrb8" Jan 26 17:10:42 crc kubenswrapper[4754]: I0126 17:10:42.105893 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/db76cab2-a25b-4072-a7c3-fcc0d6e5401a-config-data\") pod \"db76cab2-a25b-4072-a7c3-fcc0d6e5401a\" (UID: \"db76cab2-a25b-4072-a7c3-fcc0d6e5401a\") " Jan 26 17:10:42 crc kubenswrapper[4754]: I0126 17:10:42.106021 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/db76cab2-a25b-4072-a7c3-fcc0d6e5401a-scripts\") pod \"db76cab2-a25b-4072-a7c3-fcc0d6e5401a\" (UID: \"db76cab2-a25b-4072-a7c3-fcc0d6e5401a\") " Jan 26 17:10:42 crc kubenswrapper[4754]: I0126 17:10:42.106186 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db76cab2-a25b-4072-a7c3-fcc0d6e5401a-combined-ca-bundle\") pod \"db76cab2-a25b-4072-a7c3-fcc0d6e5401a\" (UID: \"db76cab2-a25b-4072-a7c3-fcc0d6e5401a\") " Jan 26 17:10:42 crc kubenswrapper[4754]: I0126 17:10:42.106219 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xdrx4\" (UniqueName: \"kubernetes.io/projected/db76cab2-a25b-4072-a7c3-fcc0d6e5401a-kube-api-access-xdrx4\") pod \"db76cab2-a25b-4072-a7c3-fcc0d6e5401a\" (UID: \"db76cab2-a25b-4072-a7c3-fcc0d6e5401a\") " Jan 26 17:10:42 crc kubenswrapper[4754]: I0126 17:10:42.111429 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/db76cab2-a25b-4072-a7c3-fcc0d6e5401a-kube-api-access-xdrx4" (OuterVolumeSpecName: "kube-api-access-xdrx4") pod "db76cab2-a25b-4072-a7c3-fcc0d6e5401a" (UID: "db76cab2-a25b-4072-a7c3-fcc0d6e5401a"). InnerVolumeSpecName "kube-api-access-xdrx4". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:10:42 crc kubenswrapper[4754]: I0126 17:10:42.111907 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/db76cab2-a25b-4072-a7c3-fcc0d6e5401a-scripts" (OuterVolumeSpecName: "scripts") pod "db76cab2-a25b-4072-a7c3-fcc0d6e5401a" (UID: "db76cab2-a25b-4072-a7c3-fcc0d6e5401a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:10:42 crc kubenswrapper[4754]: I0126 17:10:42.133068 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/db76cab2-a25b-4072-a7c3-fcc0d6e5401a-config-data" (OuterVolumeSpecName: "config-data") pod "db76cab2-a25b-4072-a7c3-fcc0d6e5401a" (UID: "db76cab2-a25b-4072-a7c3-fcc0d6e5401a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:10:42 crc kubenswrapper[4754]: I0126 17:10:42.135844 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/db76cab2-a25b-4072-a7c3-fcc0d6e5401a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "db76cab2-a25b-4072-a7c3-fcc0d6e5401a" (UID: "db76cab2-a25b-4072-a7c3-fcc0d6e5401a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:10:42 crc kubenswrapper[4754]: I0126 17:10:42.208826 4754 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db76cab2-a25b-4072-a7c3-fcc0d6e5401a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 26 17:10:42 crc kubenswrapper[4754]: I0126 17:10:42.208881 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xdrx4\" (UniqueName: \"kubernetes.io/projected/db76cab2-a25b-4072-a7c3-fcc0d6e5401a-kube-api-access-xdrx4\") on node \"crc\" DevicePath \"\"" Jan 26 17:10:42 crc kubenswrapper[4754]: I0126 17:10:42.208920 4754 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/db76cab2-a25b-4072-a7c3-fcc0d6e5401a-config-data\") on node \"crc\" DevicePath \"\"" Jan 26 17:10:42 crc kubenswrapper[4754]: I0126 17:10:42.208930 4754 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/db76cab2-a25b-4072-a7c3-fcc0d6e5401a-scripts\") on node \"crc\" DevicePath \"\"" Jan 26 17:10:42 crc kubenswrapper[4754]: I0126 17:10:42.676513 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-8lrb8" event={"ID":"db76cab2-a25b-4072-a7c3-fcc0d6e5401a","Type":"ContainerDied","Data":"d79a7852225bf0cd5e35681bf2e35471e95134c2f4b0194cbc7d0f3bc5b271c6"} Jan 26 17:10:42 crc kubenswrapper[4754]: I0126 17:10:42.677031 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d79a7852225bf0cd5e35681bf2e35471e95134c2f4b0194cbc7d0f3bc5b271c6" Jan 26 17:10:42 crc kubenswrapper[4754]: I0126 17:10:42.676734 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-8lrb8" Jan 26 17:10:42 crc kubenswrapper[4754]: I0126 17:10:42.777828 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Jan 26 17:10:42 crc kubenswrapper[4754]: E0126 17:10:42.778276 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="deac50d1-deea-4092-b2a3-b3a7ada97a42" containerName="neutron-api" Jan 26 17:10:42 crc kubenswrapper[4754]: I0126 17:10:42.778298 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="deac50d1-deea-4092-b2a3-b3a7ada97a42" containerName="neutron-api" Jan 26 17:10:42 crc kubenswrapper[4754]: E0126 17:10:42.778315 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db76cab2-a25b-4072-a7c3-fcc0d6e5401a" containerName="nova-cell0-conductor-db-sync" Jan 26 17:10:42 crc kubenswrapper[4754]: I0126 17:10:42.778323 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="db76cab2-a25b-4072-a7c3-fcc0d6e5401a" containerName="nova-cell0-conductor-db-sync" Jan 26 17:10:42 crc kubenswrapper[4754]: E0126 17:10:42.778350 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="deac50d1-deea-4092-b2a3-b3a7ada97a42" containerName="neutron-httpd" Jan 26 17:10:42 crc kubenswrapper[4754]: I0126 17:10:42.778358 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="deac50d1-deea-4092-b2a3-b3a7ada97a42" containerName="neutron-httpd" Jan 26 17:10:42 crc kubenswrapper[4754]: I0126 17:10:42.778579 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="deac50d1-deea-4092-b2a3-b3a7ada97a42" containerName="neutron-httpd" Jan 26 17:10:42 crc kubenswrapper[4754]: I0126 17:10:42.778604 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="deac50d1-deea-4092-b2a3-b3a7ada97a42" containerName="neutron-api" Jan 26 17:10:42 crc kubenswrapper[4754]: I0126 17:10:42.778626 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="db76cab2-a25b-4072-a7c3-fcc0d6e5401a" containerName="nova-cell0-conductor-db-sync" Jan 26 17:10:42 crc kubenswrapper[4754]: I0126 17:10:42.779398 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Jan 26 17:10:42 crc kubenswrapper[4754]: I0126 17:10:42.783066 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-rvgdm" Jan 26 17:10:42 crc kubenswrapper[4754]: I0126 17:10:42.783325 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Jan 26 17:10:42 crc kubenswrapper[4754]: I0126 17:10:42.788483 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Jan 26 17:10:42 crc kubenswrapper[4754]: I0126 17:10:42.922172 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k9hfn\" (UniqueName: \"kubernetes.io/projected/1d793a29-dacb-46ec-a0a8-ead07be4ce50-kube-api-access-k9hfn\") pod \"nova-cell0-conductor-0\" (UID: \"1d793a29-dacb-46ec-a0a8-ead07be4ce50\") " pod="openstack/nova-cell0-conductor-0" Jan 26 17:10:42 crc kubenswrapper[4754]: I0126 17:10:42.922274 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d793a29-dacb-46ec-a0a8-ead07be4ce50-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"1d793a29-dacb-46ec-a0a8-ead07be4ce50\") " pod="openstack/nova-cell0-conductor-0" Jan 26 17:10:42 crc kubenswrapper[4754]: I0126 17:10:42.922308 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d793a29-dacb-46ec-a0a8-ead07be4ce50-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"1d793a29-dacb-46ec-a0a8-ead07be4ce50\") " pod="openstack/nova-cell0-conductor-0" Jan 26 17:10:43 crc kubenswrapper[4754]: I0126 17:10:43.023684 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k9hfn\" (UniqueName: \"kubernetes.io/projected/1d793a29-dacb-46ec-a0a8-ead07be4ce50-kube-api-access-k9hfn\") pod \"nova-cell0-conductor-0\" (UID: \"1d793a29-dacb-46ec-a0a8-ead07be4ce50\") " pod="openstack/nova-cell0-conductor-0" Jan 26 17:10:43 crc kubenswrapper[4754]: I0126 17:10:43.023816 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d793a29-dacb-46ec-a0a8-ead07be4ce50-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"1d793a29-dacb-46ec-a0a8-ead07be4ce50\") " pod="openstack/nova-cell0-conductor-0" Jan 26 17:10:43 crc kubenswrapper[4754]: I0126 17:10:43.023849 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d793a29-dacb-46ec-a0a8-ead07be4ce50-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"1d793a29-dacb-46ec-a0a8-ead07be4ce50\") " pod="openstack/nova-cell0-conductor-0" Jan 26 17:10:43 crc kubenswrapper[4754]: I0126 17:10:43.028316 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d793a29-dacb-46ec-a0a8-ead07be4ce50-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"1d793a29-dacb-46ec-a0a8-ead07be4ce50\") " pod="openstack/nova-cell0-conductor-0" Jan 26 17:10:43 crc kubenswrapper[4754]: I0126 17:10:43.029211 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d793a29-dacb-46ec-a0a8-ead07be4ce50-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"1d793a29-dacb-46ec-a0a8-ead07be4ce50\") " pod="openstack/nova-cell0-conductor-0" Jan 26 17:10:43 crc kubenswrapper[4754]: I0126 17:10:43.041440 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k9hfn\" (UniqueName: \"kubernetes.io/projected/1d793a29-dacb-46ec-a0a8-ead07be4ce50-kube-api-access-k9hfn\") pod \"nova-cell0-conductor-0\" (UID: \"1d793a29-dacb-46ec-a0a8-ead07be4ce50\") " pod="openstack/nova-cell0-conductor-0" Jan 26 17:10:43 crc kubenswrapper[4754]: I0126 17:10:43.097124 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Jan 26 17:10:43 crc kubenswrapper[4754]: I0126 17:10:43.570210 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Jan 26 17:10:43 crc kubenswrapper[4754]: I0126 17:10:43.685110 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"1d793a29-dacb-46ec-a0a8-ead07be4ce50","Type":"ContainerStarted","Data":"540fec69880010b5635c6b33d2ef40247add163783f69c7eedd87ed67c84d9ae"} Jan 26 17:10:44 crc kubenswrapper[4754]: I0126 17:10:44.695361 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"1d793a29-dacb-46ec-a0a8-ead07be4ce50","Type":"ContainerStarted","Data":"61da7ab7823c8a0332619ca56402fca1aeebd5df27b4e4e588093e2c0053d1e7"} Jan 26 17:10:44 crc kubenswrapper[4754]: I0126 17:10:44.695867 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Jan 26 17:10:44 crc kubenswrapper[4754]: I0126 17:10:44.716660 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.716637631 podStartE2EDuration="2.716637631s" podCreationTimestamp="2026-01-26 17:10:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 17:10:44.712125826 +0000 UTC m=+1411.236306260" watchObservedRunningTime="2026-01-26 17:10:44.716637631 +0000 UTC m=+1411.240818065" Jan 26 17:10:48 crc kubenswrapper[4754]: I0126 17:10:48.127372 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Jan 26 17:10:50 crc kubenswrapper[4754]: I0126 17:10:50.230327 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-njgdn"] Jan 26 17:10:50 crc kubenswrapper[4754]: I0126 17:10:50.233234 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-njgdn" Jan 26 17:10:50 crc kubenswrapper[4754]: I0126 17:10:50.236048 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Jan 26 17:10:50 crc kubenswrapper[4754]: I0126 17:10:50.236130 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Jan 26 17:10:50 crc kubenswrapper[4754]: I0126 17:10:50.240019 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-njgdn"] Jan 26 17:10:50 crc kubenswrapper[4754]: I0126 17:10:50.368011 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/54617228-f6c0-4db0-a468-2b9283bcae21-scripts\") pod \"nova-cell0-cell-mapping-njgdn\" (UID: \"54617228-f6c0-4db0-a468-2b9283bcae21\") " pod="openstack/nova-cell0-cell-mapping-njgdn" Jan 26 17:10:50 crc kubenswrapper[4754]: I0126 17:10:50.368320 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9br5h\" (UniqueName: \"kubernetes.io/projected/54617228-f6c0-4db0-a468-2b9283bcae21-kube-api-access-9br5h\") pod \"nova-cell0-cell-mapping-njgdn\" (UID: \"54617228-f6c0-4db0-a468-2b9283bcae21\") " pod="openstack/nova-cell0-cell-mapping-njgdn" Jan 26 17:10:50 crc kubenswrapper[4754]: I0126 17:10:50.368355 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/54617228-f6c0-4db0-a468-2b9283bcae21-config-data\") pod \"nova-cell0-cell-mapping-njgdn\" (UID: \"54617228-f6c0-4db0-a468-2b9283bcae21\") " pod="openstack/nova-cell0-cell-mapping-njgdn" Jan 26 17:10:50 crc kubenswrapper[4754]: I0126 17:10:50.368377 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54617228-f6c0-4db0-a468-2b9283bcae21-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-njgdn\" (UID: \"54617228-f6c0-4db0-a468-2b9283bcae21\") " pod="openstack/nova-cell0-cell-mapping-njgdn" Jan 26 17:10:50 crc kubenswrapper[4754]: I0126 17:10:50.401449 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Jan 26 17:10:50 crc kubenswrapper[4754]: I0126 17:10:50.402516 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Jan 26 17:10:50 crc kubenswrapper[4754]: I0126 17:10:50.406895 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Jan 26 17:10:50 crc kubenswrapper[4754]: I0126 17:10:50.443359 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Jan 26 17:10:50 crc kubenswrapper[4754]: I0126 17:10:50.469600 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4q7bk\" (UniqueName: \"kubernetes.io/projected/7922c387-0282-4b8e-a907-af058a637575-kube-api-access-4q7bk\") pod \"nova-cell1-novncproxy-0\" (UID: \"7922c387-0282-4b8e-a907-af058a637575\") " pod="openstack/nova-cell1-novncproxy-0" Jan 26 17:10:50 crc kubenswrapper[4754]: I0126 17:10:50.469641 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7922c387-0282-4b8e-a907-af058a637575-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"7922c387-0282-4b8e-a907-af058a637575\") " pod="openstack/nova-cell1-novncproxy-0" Jan 26 17:10:50 crc kubenswrapper[4754]: I0126 17:10:50.469743 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/54617228-f6c0-4db0-a468-2b9283bcae21-scripts\") pod \"nova-cell0-cell-mapping-njgdn\" (UID: \"54617228-f6c0-4db0-a468-2b9283bcae21\") " pod="openstack/nova-cell0-cell-mapping-njgdn" Jan 26 17:10:50 crc kubenswrapper[4754]: I0126 17:10:50.469763 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9br5h\" (UniqueName: \"kubernetes.io/projected/54617228-f6c0-4db0-a468-2b9283bcae21-kube-api-access-9br5h\") pod \"nova-cell0-cell-mapping-njgdn\" (UID: \"54617228-f6c0-4db0-a468-2b9283bcae21\") " pod="openstack/nova-cell0-cell-mapping-njgdn" Jan 26 17:10:50 crc kubenswrapper[4754]: I0126 17:10:50.469784 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7922c387-0282-4b8e-a907-af058a637575-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"7922c387-0282-4b8e-a907-af058a637575\") " pod="openstack/nova-cell1-novncproxy-0" Jan 26 17:10:50 crc kubenswrapper[4754]: I0126 17:10:50.469814 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/54617228-f6c0-4db0-a468-2b9283bcae21-config-data\") pod \"nova-cell0-cell-mapping-njgdn\" (UID: \"54617228-f6c0-4db0-a468-2b9283bcae21\") " pod="openstack/nova-cell0-cell-mapping-njgdn" Jan 26 17:10:50 crc kubenswrapper[4754]: I0126 17:10:50.469832 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54617228-f6c0-4db0-a468-2b9283bcae21-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-njgdn\" (UID: \"54617228-f6c0-4db0-a468-2b9283bcae21\") " pod="openstack/nova-cell0-cell-mapping-njgdn" Jan 26 17:10:50 crc kubenswrapper[4754]: I0126 17:10:50.481160 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/54617228-f6c0-4db0-a468-2b9283bcae21-config-data\") pod \"nova-cell0-cell-mapping-njgdn\" (UID: \"54617228-f6c0-4db0-a468-2b9283bcae21\") " pod="openstack/nova-cell0-cell-mapping-njgdn" Jan 26 17:10:50 crc kubenswrapper[4754]: I0126 17:10:50.483108 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54617228-f6c0-4db0-a468-2b9283bcae21-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-njgdn\" (UID: \"54617228-f6c0-4db0-a468-2b9283bcae21\") " pod="openstack/nova-cell0-cell-mapping-njgdn" Jan 26 17:10:50 crc kubenswrapper[4754]: I0126 17:10:50.489713 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/54617228-f6c0-4db0-a468-2b9283bcae21-scripts\") pod \"nova-cell0-cell-mapping-njgdn\" (UID: \"54617228-f6c0-4db0-a468-2b9283bcae21\") " pod="openstack/nova-cell0-cell-mapping-njgdn" Jan 26 17:10:50 crc kubenswrapper[4754]: I0126 17:10:50.502236 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9br5h\" (UniqueName: \"kubernetes.io/projected/54617228-f6c0-4db0-a468-2b9283bcae21-kube-api-access-9br5h\") pod \"nova-cell0-cell-mapping-njgdn\" (UID: \"54617228-f6c0-4db0-a468-2b9283bcae21\") " pod="openstack/nova-cell0-cell-mapping-njgdn" Jan 26 17:10:50 crc kubenswrapper[4754]: I0126 17:10:50.534161 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Jan 26 17:10:50 crc kubenswrapper[4754]: I0126 17:10:50.536171 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Jan 26 17:10:50 crc kubenswrapper[4754]: I0126 17:10:50.546441 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Jan 26 17:10:50 crc kubenswrapper[4754]: I0126 17:10:50.551299 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Jan 26 17:10:50 crc kubenswrapper[4754]: I0126 17:10:50.573286 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-njgdn" Jan 26 17:10:50 crc kubenswrapper[4754]: I0126 17:10:50.574837 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4q7bk\" (UniqueName: \"kubernetes.io/projected/7922c387-0282-4b8e-a907-af058a637575-kube-api-access-4q7bk\") pod \"nova-cell1-novncproxy-0\" (UID: \"7922c387-0282-4b8e-a907-af058a637575\") " pod="openstack/nova-cell1-novncproxy-0" Jan 26 17:10:50 crc kubenswrapper[4754]: I0126 17:10:50.574890 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7922c387-0282-4b8e-a907-af058a637575-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"7922c387-0282-4b8e-a907-af058a637575\") " pod="openstack/nova-cell1-novncproxy-0" Jan 26 17:10:50 crc kubenswrapper[4754]: I0126 17:10:50.574979 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7922c387-0282-4b8e-a907-af058a637575-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"7922c387-0282-4b8e-a907-af058a637575\") " pod="openstack/nova-cell1-novncproxy-0" Jan 26 17:10:50 crc kubenswrapper[4754]: I0126 17:10:50.581972 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7922c387-0282-4b8e-a907-af058a637575-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"7922c387-0282-4b8e-a907-af058a637575\") " pod="openstack/nova-cell1-novncproxy-0" Jan 26 17:10:50 crc kubenswrapper[4754]: I0126 17:10:50.604834 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7922c387-0282-4b8e-a907-af058a637575-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"7922c387-0282-4b8e-a907-af058a637575\") " pod="openstack/nova-cell1-novncproxy-0" Jan 26 17:10:50 crc kubenswrapper[4754]: I0126 17:10:50.628456 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4q7bk\" (UniqueName: \"kubernetes.io/projected/7922c387-0282-4b8e-a907-af058a637575-kube-api-access-4q7bk\") pod \"nova-cell1-novncproxy-0\" (UID: \"7922c387-0282-4b8e-a907-af058a637575\") " pod="openstack/nova-cell1-novncproxy-0" Jan 26 17:10:50 crc kubenswrapper[4754]: I0126 17:10:50.680192 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9b50d057-68a1-44dc-8204-0366161e3cf4-config-data\") pod \"nova-api-0\" (UID: \"9b50d057-68a1-44dc-8204-0366161e3cf4\") " pod="openstack/nova-api-0" Jan 26 17:10:50 crc kubenswrapper[4754]: I0126 17:10:50.680583 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9b50d057-68a1-44dc-8204-0366161e3cf4-logs\") pod \"nova-api-0\" (UID: \"9b50d057-68a1-44dc-8204-0366161e3cf4\") " pod="openstack/nova-api-0" Jan 26 17:10:50 crc kubenswrapper[4754]: I0126 17:10:50.681002 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9b50d057-68a1-44dc-8204-0366161e3cf4-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"9b50d057-68a1-44dc-8204-0366161e3cf4\") " pod="openstack/nova-api-0" Jan 26 17:10:50 crc kubenswrapper[4754]: I0126 17:10:50.681053 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gfxp5\" (UniqueName: \"kubernetes.io/projected/9b50d057-68a1-44dc-8204-0366161e3cf4-kube-api-access-gfxp5\") pod \"nova-api-0\" (UID: \"9b50d057-68a1-44dc-8204-0366161e3cf4\") " pod="openstack/nova-api-0" Jan 26 17:10:50 crc kubenswrapper[4754]: I0126 17:10:50.690215 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Jan 26 17:10:50 crc kubenswrapper[4754]: I0126 17:10:50.691854 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Jan 26 17:10:50 crc kubenswrapper[4754]: I0126 17:10:50.695735 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Jan 26 17:10:50 crc kubenswrapper[4754]: I0126 17:10:50.718742 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Jan 26 17:10:50 crc kubenswrapper[4754]: I0126 17:10:50.719911 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Jan 26 17:10:50 crc kubenswrapper[4754]: I0126 17:10:50.721810 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Jan 26 17:10:50 crc kubenswrapper[4754]: I0126 17:10:50.742501 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Jan 26 17:10:50 crc kubenswrapper[4754]: I0126 17:10:50.743770 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Jan 26 17:10:50 crc kubenswrapper[4754]: I0126 17:10:50.791723 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eca0fcbc-dfb5-4bc6-9c26-621230282fb4-config-data\") pod \"nova-metadata-0\" (UID: \"eca0fcbc-dfb5-4bc6-9c26-621230282fb4\") " pod="openstack/nova-metadata-0" Jan 26 17:10:50 crc kubenswrapper[4754]: I0126 17:10:50.791805 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9b50d057-68a1-44dc-8204-0366161e3cf4-logs\") pod \"nova-api-0\" (UID: \"9b50d057-68a1-44dc-8204-0366161e3cf4\") " pod="openstack/nova-api-0" Jan 26 17:10:50 crc kubenswrapper[4754]: I0126 17:10:50.791833 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9b50d057-68a1-44dc-8204-0366161e3cf4-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"9b50d057-68a1-44dc-8204-0366161e3cf4\") " pod="openstack/nova-api-0" Jan 26 17:10:50 crc kubenswrapper[4754]: I0126 17:10:50.791866 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/270ffa9f-f033-488c-887e-565ca34620c6-config-data\") pod \"nova-scheduler-0\" (UID: \"270ffa9f-f033-488c-887e-565ca34620c6\") " pod="openstack/nova-scheduler-0" Jan 26 17:10:50 crc kubenswrapper[4754]: I0126 17:10:50.791887 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/eca0fcbc-dfb5-4bc6-9c26-621230282fb4-logs\") pod \"nova-metadata-0\" (UID: \"eca0fcbc-dfb5-4bc6-9c26-621230282fb4\") " pod="openstack/nova-metadata-0" Jan 26 17:10:50 crc kubenswrapper[4754]: I0126 17:10:50.791926 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gfxp5\" (UniqueName: \"kubernetes.io/projected/9b50d057-68a1-44dc-8204-0366161e3cf4-kube-api-access-gfxp5\") pod \"nova-api-0\" (UID: \"9b50d057-68a1-44dc-8204-0366161e3cf4\") " pod="openstack/nova-api-0" Jan 26 17:10:50 crc kubenswrapper[4754]: I0126 17:10:50.791963 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qh4h8\" (UniqueName: \"kubernetes.io/projected/270ffa9f-f033-488c-887e-565ca34620c6-kube-api-access-qh4h8\") pod \"nova-scheduler-0\" (UID: \"270ffa9f-f033-488c-887e-565ca34620c6\") " pod="openstack/nova-scheduler-0" Jan 26 17:10:50 crc kubenswrapper[4754]: I0126 17:10:50.791998 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eca0fcbc-dfb5-4bc6-9c26-621230282fb4-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"eca0fcbc-dfb5-4bc6-9c26-621230282fb4\") " pod="openstack/nova-metadata-0" Jan 26 17:10:50 crc kubenswrapper[4754]: I0126 17:10:50.792040 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9b50d057-68a1-44dc-8204-0366161e3cf4-config-data\") pod \"nova-api-0\" (UID: \"9b50d057-68a1-44dc-8204-0366161e3cf4\") " pod="openstack/nova-api-0" Jan 26 17:10:50 crc kubenswrapper[4754]: I0126 17:10:50.792071 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/270ffa9f-f033-488c-887e-565ca34620c6-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"270ffa9f-f033-488c-887e-565ca34620c6\") " pod="openstack/nova-scheduler-0" Jan 26 17:10:50 crc kubenswrapper[4754]: I0126 17:10:50.792113 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-njfcd\" (UniqueName: \"kubernetes.io/projected/eca0fcbc-dfb5-4bc6-9c26-621230282fb4-kube-api-access-njfcd\") pod \"nova-metadata-0\" (UID: \"eca0fcbc-dfb5-4bc6-9c26-621230282fb4\") " pod="openstack/nova-metadata-0" Jan 26 17:10:50 crc kubenswrapper[4754]: I0126 17:10:50.792585 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9b50d057-68a1-44dc-8204-0366161e3cf4-logs\") pod \"nova-api-0\" (UID: \"9b50d057-68a1-44dc-8204-0366161e3cf4\") " pod="openstack/nova-api-0" Jan 26 17:10:50 crc kubenswrapper[4754]: I0126 17:10:50.827946 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Jan 26 17:10:50 crc kubenswrapper[4754]: I0126 17:10:50.836152 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-gr868"] Jan 26 17:10:50 crc kubenswrapper[4754]: I0126 17:10:50.836398 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gfxp5\" (UniqueName: \"kubernetes.io/projected/9b50d057-68a1-44dc-8204-0366161e3cf4-kube-api-access-gfxp5\") pod \"nova-api-0\" (UID: \"9b50d057-68a1-44dc-8204-0366161e3cf4\") " pod="openstack/nova-api-0" Jan 26 17:10:50 crc kubenswrapper[4754]: I0126 17:10:50.841732 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-757b4f8459-gr868" Jan 26 17:10:50 crc kubenswrapper[4754]: I0126 17:10:50.849638 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9b50d057-68a1-44dc-8204-0366161e3cf4-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"9b50d057-68a1-44dc-8204-0366161e3cf4\") " pod="openstack/nova-api-0" Jan 26 17:10:50 crc kubenswrapper[4754]: I0126 17:10:50.871131 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9b50d057-68a1-44dc-8204-0366161e3cf4-config-data\") pod \"nova-api-0\" (UID: \"9b50d057-68a1-44dc-8204-0366161e3cf4\") " pod="openstack/nova-api-0" Jan 26 17:10:50 crc kubenswrapper[4754]: I0126 17:10:50.893778 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3dbb5ed9-e2d4-45c9-90c2-3d56a4b4a6ba-ovsdbserver-nb\") pod \"dnsmasq-dns-757b4f8459-gr868\" (UID: \"3dbb5ed9-e2d4-45c9-90c2-3d56a4b4a6ba\") " pod="openstack/dnsmasq-dns-757b4f8459-gr868" Jan 26 17:10:50 crc kubenswrapper[4754]: I0126 17:10:50.893856 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/270ffa9f-f033-488c-887e-565ca34620c6-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"270ffa9f-f033-488c-887e-565ca34620c6\") " pod="openstack/nova-scheduler-0" Jan 26 17:10:50 crc kubenswrapper[4754]: I0126 17:10:50.893972 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-njfcd\" (UniqueName: \"kubernetes.io/projected/eca0fcbc-dfb5-4bc6-9c26-621230282fb4-kube-api-access-njfcd\") pod \"nova-metadata-0\" (UID: \"eca0fcbc-dfb5-4bc6-9c26-621230282fb4\") " pod="openstack/nova-metadata-0" Jan 26 17:10:50 crc kubenswrapper[4754]: I0126 17:10:50.894055 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3dbb5ed9-e2d4-45c9-90c2-3d56a4b4a6ba-dns-swift-storage-0\") pod \"dnsmasq-dns-757b4f8459-gr868\" (UID: \"3dbb5ed9-e2d4-45c9-90c2-3d56a4b4a6ba\") " pod="openstack/dnsmasq-dns-757b4f8459-gr868" Jan 26 17:10:50 crc kubenswrapper[4754]: I0126 17:10:50.894114 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eca0fcbc-dfb5-4bc6-9c26-621230282fb4-config-data\") pod \"nova-metadata-0\" (UID: \"eca0fcbc-dfb5-4bc6-9c26-621230282fb4\") " pod="openstack/nova-metadata-0" Jan 26 17:10:50 crc kubenswrapper[4754]: I0126 17:10:50.894166 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3dbb5ed9-e2d4-45c9-90c2-3d56a4b4a6ba-ovsdbserver-sb\") pod \"dnsmasq-dns-757b4f8459-gr868\" (UID: \"3dbb5ed9-e2d4-45c9-90c2-3d56a4b4a6ba\") " pod="openstack/dnsmasq-dns-757b4f8459-gr868" Jan 26 17:10:50 crc kubenswrapper[4754]: I0126 17:10:50.894211 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h5psp\" (UniqueName: \"kubernetes.io/projected/3dbb5ed9-e2d4-45c9-90c2-3d56a4b4a6ba-kube-api-access-h5psp\") pod \"dnsmasq-dns-757b4f8459-gr868\" (UID: \"3dbb5ed9-e2d4-45c9-90c2-3d56a4b4a6ba\") " pod="openstack/dnsmasq-dns-757b4f8459-gr868" Jan 26 17:10:50 crc kubenswrapper[4754]: I0126 17:10:50.894251 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/270ffa9f-f033-488c-887e-565ca34620c6-config-data\") pod \"nova-scheduler-0\" (UID: \"270ffa9f-f033-488c-887e-565ca34620c6\") " pod="openstack/nova-scheduler-0" Jan 26 17:10:50 crc kubenswrapper[4754]: I0126 17:10:50.894272 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/eca0fcbc-dfb5-4bc6-9c26-621230282fb4-logs\") pod \"nova-metadata-0\" (UID: \"eca0fcbc-dfb5-4bc6-9c26-621230282fb4\") " pod="openstack/nova-metadata-0" Jan 26 17:10:50 crc kubenswrapper[4754]: I0126 17:10:50.894291 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3dbb5ed9-e2d4-45c9-90c2-3d56a4b4a6ba-config\") pod \"dnsmasq-dns-757b4f8459-gr868\" (UID: \"3dbb5ed9-e2d4-45c9-90c2-3d56a4b4a6ba\") " pod="openstack/dnsmasq-dns-757b4f8459-gr868" Jan 26 17:10:50 crc kubenswrapper[4754]: I0126 17:10:50.894338 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3dbb5ed9-e2d4-45c9-90c2-3d56a4b4a6ba-dns-svc\") pod \"dnsmasq-dns-757b4f8459-gr868\" (UID: \"3dbb5ed9-e2d4-45c9-90c2-3d56a4b4a6ba\") " pod="openstack/dnsmasq-dns-757b4f8459-gr868" Jan 26 17:10:50 crc kubenswrapper[4754]: I0126 17:10:50.894369 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qh4h8\" (UniqueName: \"kubernetes.io/projected/270ffa9f-f033-488c-887e-565ca34620c6-kube-api-access-qh4h8\") pod \"nova-scheduler-0\" (UID: \"270ffa9f-f033-488c-887e-565ca34620c6\") " pod="openstack/nova-scheduler-0" Jan 26 17:10:50 crc kubenswrapper[4754]: I0126 17:10:50.894409 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eca0fcbc-dfb5-4bc6-9c26-621230282fb4-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"eca0fcbc-dfb5-4bc6-9c26-621230282fb4\") " pod="openstack/nova-metadata-0" Jan 26 17:10:50 crc kubenswrapper[4754]: I0126 17:10:50.900793 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/270ffa9f-f033-488c-887e-565ca34620c6-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"270ffa9f-f033-488c-887e-565ca34620c6\") " pod="openstack/nova-scheduler-0" Jan 26 17:10:50 crc kubenswrapper[4754]: I0126 17:10:50.901295 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/eca0fcbc-dfb5-4bc6-9c26-621230282fb4-logs\") pod \"nova-metadata-0\" (UID: \"eca0fcbc-dfb5-4bc6-9c26-621230282fb4\") " pod="openstack/nova-metadata-0" Jan 26 17:10:50 crc kubenswrapper[4754]: I0126 17:10:50.901854 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-gr868"] Jan 26 17:10:50 crc kubenswrapper[4754]: I0126 17:10:50.902276 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eca0fcbc-dfb5-4bc6-9c26-621230282fb4-config-data\") pod \"nova-metadata-0\" (UID: \"eca0fcbc-dfb5-4bc6-9c26-621230282fb4\") " pod="openstack/nova-metadata-0" Jan 26 17:10:50 crc kubenswrapper[4754]: I0126 17:10:50.904957 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eca0fcbc-dfb5-4bc6-9c26-621230282fb4-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"eca0fcbc-dfb5-4bc6-9c26-621230282fb4\") " pod="openstack/nova-metadata-0" Jan 26 17:10:50 crc kubenswrapper[4754]: I0126 17:10:50.911583 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/270ffa9f-f033-488c-887e-565ca34620c6-config-data\") pod \"nova-scheduler-0\" (UID: \"270ffa9f-f033-488c-887e-565ca34620c6\") " pod="openstack/nova-scheduler-0" Jan 26 17:10:50 crc kubenswrapper[4754]: I0126 17:10:50.926397 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qh4h8\" (UniqueName: \"kubernetes.io/projected/270ffa9f-f033-488c-887e-565ca34620c6-kube-api-access-qh4h8\") pod \"nova-scheduler-0\" (UID: \"270ffa9f-f033-488c-887e-565ca34620c6\") " pod="openstack/nova-scheduler-0" Jan 26 17:10:50 crc kubenswrapper[4754]: I0126 17:10:50.926756 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-njfcd\" (UniqueName: \"kubernetes.io/projected/eca0fcbc-dfb5-4bc6-9c26-621230282fb4-kube-api-access-njfcd\") pod \"nova-metadata-0\" (UID: \"eca0fcbc-dfb5-4bc6-9c26-621230282fb4\") " pod="openstack/nova-metadata-0" Jan 26 17:10:50 crc kubenswrapper[4754]: I0126 17:10:50.994584 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Jan 26 17:10:50 crc kubenswrapper[4754]: I0126 17:10:50.995348 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3dbb5ed9-e2d4-45c9-90c2-3d56a4b4a6ba-ovsdbserver-nb\") pod \"dnsmasq-dns-757b4f8459-gr868\" (UID: \"3dbb5ed9-e2d4-45c9-90c2-3d56a4b4a6ba\") " pod="openstack/dnsmasq-dns-757b4f8459-gr868" Jan 26 17:10:50 crc kubenswrapper[4754]: I0126 17:10:50.995416 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3dbb5ed9-e2d4-45c9-90c2-3d56a4b4a6ba-dns-swift-storage-0\") pod \"dnsmasq-dns-757b4f8459-gr868\" (UID: \"3dbb5ed9-e2d4-45c9-90c2-3d56a4b4a6ba\") " pod="openstack/dnsmasq-dns-757b4f8459-gr868" Jan 26 17:10:50 crc kubenswrapper[4754]: I0126 17:10:50.995458 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3dbb5ed9-e2d4-45c9-90c2-3d56a4b4a6ba-ovsdbserver-sb\") pod \"dnsmasq-dns-757b4f8459-gr868\" (UID: \"3dbb5ed9-e2d4-45c9-90c2-3d56a4b4a6ba\") " pod="openstack/dnsmasq-dns-757b4f8459-gr868" Jan 26 17:10:50 crc kubenswrapper[4754]: I0126 17:10:50.995482 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h5psp\" (UniqueName: \"kubernetes.io/projected/3dbb5ed9-e2d4-45c9-90c2-3d56a4b4a6ba-kube-api-access-h5psp\") pod \"dnsmasq-dns-757b4f8459-gr868\" (UID: \"3dbb5ed9-e2d4-45c9-90c2-3d56a4b4a6ba\") " pod="openstack/dnsmasq-dns-757b4f8459-gr868" Jan 26 17:10:50 crc kubenswrapper[4754]: I0126 17:10:50.995509 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3dbb5ed9-e2d4-45c9-90c2-3d56a4b4a6ba-config\") pod \"dnsmasq-dns-757b4f8459-gr868\" (UID: \"3dbb5ed9-e2d4-45c9-90c2-3d56a4b4a6ba\") " pod="openstack/dnsmasq-dns-757b4f8459-gr868" Jan 26 17:10:50 crc kubenswrapper[4754]: I0126 17:10:50.995544 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3dbb5ed9-e2d4-45c9-90c2-3d56a4b4a6ba-dns-svc\") pod \"dnsmasq-dns-757b4f8459-gr868\" (UID: \"3dbb5ed9-e2d4-45c9-90c2-3d56a4b4a6ba\") " pod="openstack/dnsmasq-dns-757b4f8459-gr868" Jan 26 17:10:50 crc kubenswrapper[4754]: I0126 17:10:50.996636 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3dbb5ed9-e2d4-45c9-90c2-3d56a4b4a6ba-dns-svc\") pod \"dnsmasq-dns-757b4f8459-gr868\" (UID: \"3dbb5ed9-e2d4-45c9-90c2-3d56a4b4a6ba\") " pod="openstack/dnsmasq-dns-757b4f8459-gr868" Jan 26 17:10:50 crc kubenswrapper[4754]: I0126 17:10:50.997789 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3dbb5ed9-e2d4-45c9-90c2-3d56a4b4a6ba-ovsdbserver-nb\") pod \"dnsmasq-dns-757b4f8459-gr868\" (UID: \"3dbb5ed9-e2d4-45c9-90c2-3d56a4b4a6ba\") " pod="openstack/dnsmasq-dns-757b4f8459-gr868" Jan 26 17:10:50 crc kubenswrapper[4754]: I0126 17:10:50.999732 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3dbb5ed9-e2d4-45c9-90c2-3d56a4b4a6ba-ovsdbserver-sb\") pod \"dnsmasq-dns-757b4f8459-gr868\" (UID: \"3dbb5ed9-e2d4-45c9-90c2-3d56a4b4a6ba\") " pod="openstack/dnsmasq-dns-757b4f8459-gr868" Jan 26 17:10:51 crc kubenswrapper[4754]: I0126 17:10:51.002681 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3dbb5ed9-e2d4-45c9-90c2-3d56a4b4a6ba-config\") pod \"dnsmasq-dns-757b4f8459-gr868\" (UID: \"3dbb5ed9-e2d4-45c9-90c2-3d56a4b4a6ba\") " pod="openstack/dnsmasq-dns-757b4f8459-gr868" Jan 26 17:10:51 crc kubenswrapper[4754]: I0126 17:10:51.002708 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3dbb5ed9-e2d4-45c9-90c2-3d56a4b4a6ba-dns-swift-storage-0\") pod \"dnsmasq-dns-757b4f8459-gr868\" (UID: \"3dbb5ed9-e2d4-45c9-90c2-3d56a4b4a6ba\") " pod="openstack/dnsmasq-dns-757b4f8459-gr868" Jan 26 17:10:51 crc kubenswrapper[4754]: I0126 17:10:51.012386 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h5psp\" (UniqueName: \"kubernetes.io/projected/3dbb5ed9-e2d4-45c9-90c2-3d56a4b4a6ba-kube-api-access-h5psp\") pod \"dnsmasq-dns-757b4f8459-gr868\" (UID: \"3dbb5ed9-e2d4-45c9-90c2-3d56a4b4a6ba\") " pod="openstack/dnsmasq-dns-757b4f8459-gr868" Jan 26 17:10:51 crc kubenswrapper[4754]: I0126 17:10:51.021648 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Jan 26 17:10:51 crc kubenswrapper[4754]: I0126 17:10:51.081507 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Jan 26 17:10:51 crc kubenswrapper[4754]: I0126 17:10:51.242513 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-757b4f8459-gr868" Jan 26 17:10:51 crc kubenswrapper[4754]: I0126 17:10:51.298584 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-njgdn"] Jan 26 17:10:51 crc kubenswrapper[4754]: I0126 17:10:51.407056 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Jan 26 17:10:51 crc kubenswrapper[4754]: I0126 17:10:51.429462 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-55c6p"] Jan 26 17:10:51 crc kubenswrapper[4754]: I0126 17:10:51.431262 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-55c6p" Jan 26 17:10:51 crc kubenswrapper[4754]: I0126 17:10:51.438492 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Jan 26 17:10:51 crc kubenswrapper[4754]: I0126 17:10:51.438522 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Jan 26 17:10:51 crc kubenswrapper[4754]: I0126 17:10:51.456603 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-55c6p"] Jan 26 17:10:51 crc kubenswrapper[4754]: I0126 17:10:51.508693 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d40200db-b145-4a60-b337-82ae5032cc51-config-data\") pod \"nova-cell1-conductor-db-sync-55c6p\" (UID: \"d40200db-b145-4a60-b337-82ae5032cc51\") " pod="openstack/nova-cell1-conductor-db-sync-55c6p" Jan 26 17:10:51 crc kubenswrapper[4754]: I0126 17:10:51.508740 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d40200db-b145-4a60-b337-82ae5032cc51-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-55c6p\" (UID: \"d40200db-b145-4a60-b337-82ae5032cc51\") " pod="openstack/nova-cell1-conductor-db-sync-55c6p" Jan 26 17:10:51 crc kubenswrapper[4754]: I0126 17:10:51.508840 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7km8l\" (UniqueName: \"kubernetes.io/projected/d40200db-b145-4a60-b337-82ae5032cc51-kube-api-access-7km8l\") pod \"nova-cell1-conductor-db-sync-55c6p\" (UID: \"d40200db-b145-4a60-b337-82ae5032cc51\") " pod="openstack/nova-cell1-conductor-db-sync-55c6p" Jan 26 17:10:51 crc kubenswrapper[4754]: I0126 17:10:51.508895 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d40200db-b145-4a60-b337-82ae5032cc51-scripts\") pod \"nova-cell1-conductor-db-sync-55c6p\" (UID: \"d40200db-b145-4a60-b337-82ae5032cc51\") " pod="openstack/nova-cell1-conductor-db-sync-55c6p" Jan 26 17:10:51 crc kubenswrapper[4754]: I0126 17:10:51.528081 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Jan 26 17:10:51 crc kubenswrapper[4754]: I0126 17:10:51.611786 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7km8l\" (UniqueName: \"kubernetes.io/projected/d40200db-b145-4a60-b337-82ae5032cc51-kube-api-access-7km8l\") pod \"nova-cell1-conductor-db-sync-55c6p\" (UID: \"d40200db-b145-4a60-b337-82ae5032cc51\") " pod="openstack/nova-cell1-conductor-db-sync-55c6p" Jan 26 17:10:51 crc kubenswrapper[4754]: I0126 17:10:51.611846 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d40200db-b145-4a60-b337-82ae5032cc51-scripts\") pod \"nova-cell1-conductor-db-sync-55c6p\" (UID: \"d40200db-b145-4a60-b337-82ae5032cc51\") " pod="openstack/nova-cell1-conductor-db-sync-55c6p" Jan 26 17:10:51 crc kubenswrapper[4754]: I0126 17:10:51.611921 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d40200db-b145-4a60-b337-82ae5032cc51-config-data\") pod \"nova-cell1-conductor-db-sync-55c6p\" (UID: \"d40200db-b145-4a60-b337-82ae5032cc51\") " pod="openstack/nova-cell1-conductor-db-sync-55c6p" Jan 26 17:10:51 crc kubenswrapper[4754]: I0126 17:10:51.611944 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d40200db-b145-4a60-b337-82ae5032cc51-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-55c6p\" (UID: \"d40200db-b145-4a60-b337-82ae5032cc51\") " pod="openstack/nova-cell1-conductor-db-sync-55c6p" Jan 26 17:10:51 crc kubenswrapper[4754]: I0126 17:10:51.618752 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d40200db-b145-4a60-b337-82ae5032cc51-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-55c6p\" (UID: \"d40200db-b145-4a60-b337-82ae5032cc51\") " pod="openstack/nova-cell1-conductor-db-sync-55c6p" Jan 26 17:10:51 crc kubenswrapper[4754]: I0126 17:10:51.631818 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7km8l\" (UniqueName: \"kubernetes.io/projected/d40200db-b145-4a60-b337-82ae5032cc51-kube-api-access-7km8l\") pod \"nova-cell1-conductor-db-sync-55c6p\" (UID: \"d40200db-b145-4a60-b337-82ae5032cc51\") " pod="openstack/nova-cell1-conductor-db-sync-55c6p" Jan 26 17:10:51 crc kubenswrapper[4754]: I0126 17:10:51.634408 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d40200db-b145-4a60-b337-82ae5032cc51-scripts\") pod \"nova-cell1-conductor-db-sync-55c6p\" (UID: \"d40200db-b145-4a60-b337-82ae5032cc51\") " pod="openstack/nova-cell1-conductor-db-sync-55c6p" Jan 26 17:10:51 crc kubenswrapper[4754]: I0126 17:10:51.636605 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d40200db-b145-4a60-b337-82ae5032cc51-config-data\") pod \"nova-cell1-conductor-db-sync-55c6p\" (UID: \"d40200db-b145-4a60-b337-82ae5032cc51\") " pod="openstack/nova-cell1-conductor-db-sync-55c6p" Jan 26 17:10:51 crc kubenswrapper[4754]: I0126 17:10:51.664274 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Jan 26 17:10:51 crc kubenswrapper[4754]: I0126 17:10:51.722105 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-55c6p" Jan 26 17:10:51 crc kubenswrapper[4754]: I0126 17:10:51.734995 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Jan 26 17:10:51 crc kubenswrapper[4754]: W0126 17:10:51.766505 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod270ffa9f_f033_488c_887e_565ca34620c6.slice/crio-eb0592623072b15408efdbdfd5823d972d980b2bebd8e10f94507279c567ec4d WatchSource:0}: Error finding container eb0592623072b15408efdbdfd5823d972d980b2bebd8e10f94507279c567ec4d: Status 404 returned error can't find the container with id eb0592623072b15408efdbdfd5823d972d980b2bebd8e10f94507279c567ec4d Jan 26 17:10:51 crc kubenswrapper[4754]: I0126 17:10:51.782717 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"270ffa9f-f033-488c-887e-565ca34620c6","Type":"ContainerStarted","Data":"eb0592623072b15408efdbdfd5823d972d980b2bebd8e10f94507279c567ec4d"} Jan 26 17:10:51 crc kubenswrapper[4754]: I0126 17:10:51.783775 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"eca0fcbc-dfb5-4bc6-9c26-621230282fb4","Type":"ContainerStarted","Data":"00cfd6e217217ec30da0c6286e4beaf21f39d4f682b28d09e318642ab958b986"} Jan 26 17:10:51 crc kubenswrapper[4754]: I0126 17:10:51.784930 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"7922c387-0282-4b8e-a907-af058a637575","Type":"ContainerStarted","Data":"28065df71fcddc1b218655ca0c7410bcad8f81754d83496259ba407bb264c0f4"} Jan 26 17:10:51 crc kubenswrapper[4754]: I0126 17:10:51.786222 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-njgdn" event={"ID":"54617228-f6c0-4db0-a468-2b9283bcae21","Type":"ContainerStarted","Data":"e910b0c7089f8278c41b5fc19bdb746ba8ab6a228cf4defa17621078ef4cfdc9"} Jan 26 17:10:51 crc kubenswrapper[4754]: I0126 17:10:51.789021 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"9b50d057-68a1-44dc-8204-0366161e3cf4","Type":"ContainerStarted","Data":"a103081402c951965ad316a8311c8605e190face4d63fd1550a224a59959e289"} Jan 26 17:10:51 crc kubenswrapper[4754]: I0126 17:10:51.838919 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-gr868"] Jan 26 17:10:51 crc kubenswrapper[4754]: W0126 17:10:51.841185 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3dbb5ed9_e2d4_45c9_90c2_3d56a4b4a6ba.slice/crio-0b11e0d40d1a2ab4a954951a3251aa13984038b682c61e4503e1a16ee5d2c1e9 WatchSource:0}: Error finding container 0b11e0d40d1a2ab4a954951a3251aa13984038b682c61e4503e1a16ee5d2c1e9: Status 404 returned error can't find the container with id 0b11e0d40d1a2ab4a954951a3251aa13984038b682c61e4503e1a16ee5d2c1e9 Jan 26 17:10:52 crc kubenswrapper[4754]: I0126 17:10:52.194388 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-55c6p"] Jan 26 17:10:52 crc kubenswrapper[4754]: W0126 17:10:52.205602 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd40200db_b145_4a60_b337_82ae5032cc51.slice/crio-feef8482359e4251be97153f960b9e2ae8de7be83b6af60e9299abaaba674cf9 WatchSource:0}: Error finding container feef8482359e4251be97153f960b9e2ae8de7be83b6af60e9299abaaba674cf9: Status 404 returned error can't find the container with id feef8482359e4251be97153f960b9e2ae8de7be83b6af60e9299abaaba674cf9 Jan 26 17:10:52 crc kubenswrapper[4754]: I0126 17:10:52.799968 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-njgdn" event={"ID":"54617228-f6c0-4db0-a468-2b9283bcae21","Type":"ContainerStarted","Data":"81c700b92383b930ceaf735cdacfcdd6280d44d2817b951b810ae6a363a3efb3"} Jan 26 17:10:52 crc kubenswrapper[4754]: I0126 17:10:52.801580 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-757b4f8459-gr868" event={"ID":"3dbb5ed9-e2d4-45c9-90c2-3d56a4b4a6ba","Type":"ContainerStarted","Data":"0b11e0d40d1a2ab4a954951a3251aa13984038b682c61e4503e1a16ee5d2c1e9"} Jan 26 17:10:52 crc kubenswrapper[4754]: I0126 17:10:52.802647 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-55c6p" event={"ID":"d40200db-b145-4a60-b337-82ae5032cc51","Type":"ContainerStarted","Data":"feef8482359e4251be97153f960b9e2ae8de7be83b6af60e9299abaaba674cf9"} Jan 26 17:10:53 crc kubenswrapper[4754]: I0126 17:10:53.859764 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-55c6p" event={"ID":"d40200db-b145-4a60-b337-82ae5032cc51","Type":"ContainerStarted","Data":"f3fe6e1c3d0a565d5c2f44c4be05b0cb2ef674bd9aca961cfe8edcb5ff320e24"} Jan 26 17:10:53 crc kubenswrapper[4754]: I0126 17:10:53.875601 4754 generic.go:334] "Generic (PLEG): container finished" podID="3dbb5ed9-e2d4-45c9-90c2-3d56a4b4a6ba" containerID="ef702a387578e08af58300cf968ed63cc88a4db4cfedf1c3f865d2be42ec5add" exitCode=0 Jan 26 17:10:53 crc kubenswrapper[4754]: I0126 17:10:53.876821 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-757b4f8459-gr868" event={"ID":"3dbb5ed9-e2d4-45c9-90c2-3d56a4b4a6ba","Type":"ContainerDied","Data":"ef702a387578e08af58300cf968ed63cc88a4db4cfedf1c3f865d2be42ec5add"} Jan 26 17:10:54 crc kubenswrapper[4754]: I0126 17:10:54.003903 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-njgdn" podStartSLOduration=4.003883007 podStartE2EDuration="4.003883007s" podCreationTimestamp="2026-01-26 17:10:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 17:10:53.985723967 +0000 UTC m=+1420.509904401" watchObservedRunningTime="2026-01-26 17:10:54.003883007 +0000 UTC m=+1420.528063441" Jan 26 17:10:54 crc kubenswrapper[4754]: I0126 17:10:54.031842 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-55c6p" podStartSLOduration=3.031823067 podStartE2EDuration="3.031823067s" podCreationTimestamp="2026-01-26 17:10:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 17:10:54.002342434 +0000 UTC m=+1420.526522868" watchObservedRunningTime="2026-01-26 17:10:54.031823067 +0000 UTC m=+1420.556003501" Jan 26 17:10:54 crc kubenswrapper[4754]: I0126 17:10:54.103621 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Jan 26 17:10:54 crc kubenswrapper[4754]: I0126 17:10:54.188103 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Jan 26 17:10:54 crc kubenswrapper[4754]: I0126 17:10:54.891837 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-757b4f8459-gr868" event={"ID":"3dbb5ed9-e2d4-45c9-90c2-3d56a4b4a6ba","Type":"ContainerStarted","Data":"b7fce4148dc1c9bff90d8acb44402b775254d76641bf9c8076c267852e207723"} Jan 26 17:10:54 crc kubenswrapper[4754]: I0126 17:10:54.891910 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-757b4f8459-gr868" Jan 26 17:10:54 crc kubenswrapper[4754]: I0126 17:10:54.915478 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-757b4f8459-gr868" podStartSLOduration=4.915453758 podStartE2EDuration="4.915453758s" podCreationTimestamp="2026-01-26 17:10:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 17:10:54.910453841 +0000 UTC m=+1421.434634275" watchObservedRunningTime="2026-01-26 17:10:54.915453758 +0000 UTC m=+1421.439634202" Jan 26 17:10:57 crc kubenswrapper[4754]: I0126 17:10:57.948590 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Jan 26 17:10:59 crc kubenswrapper[4754]: I0126 17:10:59.561676 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-hxdzc"] Jan 26 17:10:59 crc kubenswrapper[4754]: I0126 17:10:59.565001 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hxdzc" Jan 26 17:10:59 crc kubenswrapper[4754]: I0126 17:10:59.584926 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-hxdzc"] Jan 26 17:10:59 crc kubenswrapper[4754]: I0126 17:10:59.670892 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e3881c45-db75-4ca1-bb8f-92f413446265-catalog-content\") pod \"redhat-operators-hxdzc\" (UID: \"e3881c45-db75-4ca1-bb8f-92f413446265\") " pod="openshift-marketplace/redhat-operators-hxdzc" Jan 26 17:10:59 crc kubenswrapper[4754]: I0126 17:10:59.670969 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6cc96\" (UniqueName: \"kubernetes.io/projected/e3881c45-db75-4ca1-bb8f-92f413446265-kube-api-access-6cc96\") pod \"redhat-operators-hxdzc\" (UID: \"e3881c45-db75-4ca1-bb8f-92f413446265\") " pod="openshift-marketplace/redhat-operators-hxdzc" Jan 26 17:10:59 crc kubenswrapper[4754]: I0126 17:10:59.671031 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e3881c45-db75-4ca1-bb8f-92f413446265-utilities\") pod \"redhat-operators-hxdzc\" (UID: \"e3881c45-db75-4ca1-bb8f-92f413446265\") " pod="openshift-marketplace/redhat-operators-hxdzc" Jan 26 17:10:59 crc kubenswrapper[4754]: I0126 17:10:59.772411 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e3881c45-db75-4ca1-bb8f-92f413446265-utilities\") pod \"redhat-operators-hxdzc\" (UID: \"e3881c45-db75-4ca1-bb8f-92f413446265\") " pod="openshift-marketplace/redhat-operators-hxdzc" Jan 26 17:10:59 crc kubenswrapper[4754]: I0126 17:10:59.772520 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e3881c45-db75-4ca1-bb8f-92f413446265-catalog-content\") pod \"redhat-operators-hxdzc\" (UID: \"e3881c45-db75-4ca1-bb8f-92f413446265\") " pod="openshift-marketplace/redhat-operators-hxdzc" Jan 26 17:10:59 crc kubenswrapper[4754]: I0126 17:10:59.772560 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6cc96\" (UniqueName: \"kubernetes.io/projected/e3881c45-db75-4ca1-bb8f-92f413446265-kube-api-access-6cc96\") pod \"redhat-operators-hxdzc\" (UID: \"e3881c45-db75-4ca1-bb8f-92f413446265\") " pod="openshift-marketplace/redhat-operators-hxdzc" Jan 26 17:10:59 crc kubenswrapper[4754]: I0126 17:10:59.773199 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e3881c45-db75-4ca1-bb8f-92f413446265-utilities\") pod \"redhat-operators-hxdzc\" (UID: \"e3881c45-db75-4ca1-bb8f-92f413446265\") " pod="openshift-marketplace/redhat-operators-hxdzc" Jan 26 17:10:59 crc kubenswrapper[4754]: I0126 17:10:59.773521 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e3881c45-db75-4ca1-bb8f-92f413446265-catalog-content\") pod \"redhat-operators-hxdzc\" (UID: \"e3881c45-db75-4ca1-bb8f-92f413446265\") " pod="openshift-marketplace/redhat-operators-hxdzc" Jan 26 17:10:59 crc kubenswrapper[4754]: I0126 17:10:59.810720 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6cc96\" (UniqueName: \"kubernetes.io/projected/e3881c45-db75-4ca1-bb8f-92f413446265-kube-api-access-6cc96\") pod \"redhat-operators-hxdzc\" (UID: \"e3881c45-db75-4ca1-bb8f-92f413446265\") " pod="openshift-marketplace/redhat-operators-hxdzc" Jan 26 17:10:59 crc kubenswrapper[4754]: I0126 17:10:59.909228 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hxdzc" Jan 26 17:10:59 crc kubenswrapper[4754]: I0126 17:10:59.934047 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"7922c387-0282-4b8e-a907-af058a637575","Type":"ContainerStarted","Data":"f05e7698a2c643b6ab539c22a2c1983a4bfc697ed974bcb9c7ba86d8a596f3e9"} Jan 26 17:10:59 crc kubenswrapper[4754]: I0126 17:10:59.934213 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="7922c387-0282-4b8e-a907-af058a637575" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://f05e7698a2c643b6ab539c22a2c1983a4bfc697ed974bcb9c7ba86d8a596f3e9" gracePeriod=30 Jan 26 17:10:59 crc kubenswrapper[4754]: I0126 17:10:59.935973 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"9b50d057-68a1-44dc-8204-0366161e3cf4","Type":"ContainerStarted","Data":"4c0f6a01ffaef43ea38f623749ef8fb2a10ea1b37845de8eae86906116547dc6"} Jan 26 17:10:59 crc kubenswrapper[4754]: I0126 17:10:59.936098 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"9b50d057-68a1-44dc-8204-0366161e3cf4","Type":"ContainerStarted","Data":"77aafa83659a5fb388cddee05c48d3c0ee10ca31c83aecb39a67800530e24940"} Jan 26 17:10:59 crc kubenswrapper[4754]: I0126 17:10:59.939165 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"270ffa9f-f033-488c-887e-565ca34620c6","Type":"ContainerStarted","Data":"dbafe75f4149545f66d7dcf67e5679e62b6cad3503ae5ede59a6619920905833"} Jan 26 17:10:59 crc kubenswrapper[4754]: I0126 17:10:59.941801 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"eca0fcbc-dfb5-4bc6-9c26-621230282fb4","Type":"ContainerStarted","Data":"d81cba4378ff840952fbd04ffd402cdc93da3a8fce7bfb7265942c4fad16f0f3"} Jan 26 17:10:59 crc kubenswrapper[4754]: I0126 17:10:59.941837 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"eca0fcbc-dfb5-4bc6-9c26-621230282fb4","Type":"ContainerStarted","Data":"7d9708cd4813226778c6001f4e7a76bf20c0d89caecec18eeae7090cd5058c94"} Jan 26 17:10:59 crc kubenswrapper[4754]: I0126 17:10:59.942021 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="eca0fcbc-dfb5-4bc6-9c26-621230282fb4" containerName="nova-metadata-log" containerID="cri-o://7d9708cd4813226778c6001f4e7a76bf20c0d89caecec18eeae7090cd5058c94" gracePeriod=30 Jan 26 17:10:59 crc kubenswrapper[4754]: I0126 17:10:59.942186 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="eca0fcbc-dfb5-4bc6-9c26-621230282fb4" containerName="nova-metadata-metadata" containerID="cri-o://d81cba4378ff840952fbd04ffd402cdc93da3a8fce7bfb7265942c4fad16f0f3" gracePeriod=30 Jan 26 17:10:59 crc kubenswrapper[4754]: I0126 17:10:59.954299 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=3.207277223 podStartE2EDuration="9.954285133s" podCreationTimestamp="2026-01-26 17:10:50 +0000 UTC" firstStartedPulling="2026-01-26 17:10:51.430340593 +0000 UTC m=+1417.954521017" lastFinishedPulling="2026-01-26 17:10:58.177348493 +0000 UTC m=+1424.701528927" observedRunningTime="2026-01-26 17:10:59.951347852 +0000 UTC m=+1426.475528276" watchObservedRunningTime="2026-01-26 17:10:59.954285133 +0000 UTC m=+1426.478465567" Jan 26 17:10:59 crc kubenswrapper[4754]: I0126 17:10:59.980023 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=3.573429394 podStartE2EDuration="9.980002422s" podCreationTimestamp="2026-01-26 17:10:50 +0000 UTC" firstStartedPulling="2026-01-26 17:10:51.769630843 +0000 UTC m=+1418.293811277" lastFinishedPulling="2026-01-26 17:10:58.176203871 +0000 UTC m=+1424.700384305" observedRunningTime="2026-01-26 17:10:59.970706946 +0000 UTC m=+1426.494887370" watchObservedRunningTime="2026-01-26 17:10:59.980002422 +0000 UTC m=+1426.504182856" Jan 26 17:11:00 crc kubenswrapper[4754]: I0126 17:11:00.002537 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=3.365764481 podStartE2EDuration="10.002516382s" podCreationTimestamp="2026-01-26 17:10:50 +0000 UTC" firstStartedPulling="2026-01-26 17:10:51.538622587 +0000 UTC m=+1418.062803021" lastFinishedPulling="2026-01-26 17:10:58.175374478 +0000 UTC m=+1424.699554922" observedRunningTime="2026-01-26 17:10:59.993851193 +0000 UTC m=+1426.518031637" watchObservedRunningTime="2026-01-26 17:11:00.002516382 +0000 UTC m=+1426.526696816" Jan 26 17:11:00 crc kubenswrapper[4754]: I0126 17:11:00.022024 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=3.517628105 podStartE2EDuration="10.022006689s" podCreationTimestamp="2026-01-26 17:10:50 +0000 UTC" firstStartedPulling="2026-01-26 17:10:51.671522139 +0000 UTC m=+1418.195702573" lastFinishedPulling="2026-01-26 17:10:58.175900723 +0000 UTC m=+1424.700081157" observedRunningTime="2026-01-26 17:11:00.014706169 +0000 UTC m=+1426.538886603" watchObservedRunningTime="2026-01-26 17:11:00.022006689 +0000 UTC m=+1426.546187123" Jan 26 17:11:00 crc kubenswrapper[4754]: I0126 17:11:00.746219 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Jan 26 17:11:00 crc kubenswrapper[4754]: I0126 17:11:00.952240 4754 generic.go:334] "Generic (PLEG): container finished" podID="eca0fcbc-dfb5-4bc6-9c26-621230282fb4" containerID="7d9708cd4813226778c6001f4e7a76bf20c0d89caecec18eeae7090cd5058c94" exitCode=143 Jan 26 17:11:00 crc kubenswrapper[4754]: I0126 17:11:00.952990 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"eca0fcbc-dfb5-4bc6-9c26-621230282fb4","Type":"ContainerDied","Data":"7d9708cd4813226778c6001f4e7a76bf20c0d89caecec18eeae7090cd5058c94"} Jan 26 17:11:00 crc kubenswrapper[4754]: I0126 17:11:00.994936 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Jan 26 17:11:00 crc kubenswrapper[4754]: I0126 17:11:00.994996 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Jan 26 17:11:01 crc kubenswrapper[4754]: I0126 17:11:01.023392 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Jan 26 17:11:01 crc kubenswrapper[4754]: I0126 17:11:01.023459 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Jan 26 17:11:01 crc kubenswrapper[4754]: I0126 17:11:01.082227 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Jan 26 17:11:01 crc kubenswrapper[4754]: I0126 17:11:01.082274 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Jan 26 17:11:01 crc kubenswrapper[4754]: I0126 17:11:01.111236 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Jan 26 17:11:01 crc kubenswrapper[4754]: I0126 17:11:01.158192 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-hxdzc"] Jan 26 17:11:01 crc kubenswrapper[4754]: I0126 17:11:01.244814 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-757b4f8459-gr868" Jan 26 17:11:01 crc kubenswrapper[4754]: I0126 17:11:01.322643 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-tnfmz"] Jan 26 17:11:01 crc kubenswrapper[4754]: I0126 17:11:01.322875 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5c9776ccc5-tnfmz" podUID="952095c1-837b-4ff5-b61c-ce1e42102b56" containerName="dnsmasq-dns" containerID="cri-o://a9080148a080f6c8a5352be74755999d3eb0d22a9a34cdd78cd8072a9a4d3f4f" gracePeriod=10 Jan 26 17:11:01 crc kubenswrapper[4754]: I0126 17:11:01.962833 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hxdzc" event={"ID":"e3881c45-db75-4ca1-bb8f-92f413446265","Type":"ContainerStarted","Data":"4185283ec6b8bfea7c969e1e116f6478ec1a3c39bee3af1453e9c98e0676dd40"} Jan 26 17:11:01 crc kubenswrapper[4754]: I0126 17:11:01.965011 4754 generic.go:334] "Generic (PLEG): container finished" podID="952095c1-837b-4ff5-b61c-ce1e42102b56" containerID="a9080148a080f6c8a5352be74755999d3eb0d22a9a34cdd78cd8072a9a4d3f4f" exitCode=0 Jan 26 17:11:01 crc kubenswrapper[4754]: I0126 17:11:01.965074 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-tnfmz" event={"ID":"952095c1-837b-4ff5-b61c-ce1e42102b56","Type":"ContainerDied","Data":"a9080148a080f6c8a5352be74755999d3eb0d22a9a34cdd78cd8072a9a4d3f4f"} Jan 26 17:11:01 crc kubenswrapper[4754]: I0126 17:11:01.966809 4754 generic.go:334] "Generic (PLEG): container finished" podID="eca0fcbc-dfb5-4bc6-9c26-621230282fb4" containerID="d81cba4378ff840952fbd04ffd402cdc93da3a8fce7bfb7265942c4fad16f0f3" exitCode=0 Jan 26 17:11:01 crc kubenswrapper[4754]: I0126 17:11:01.966862 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"eca0fcbc-dfb5-4bc6-9c26-621230282fb4","Type":"ContainerDied","Data":"d81cba4378ff840952fbd04ffd402cdc93da3a8fce7bfb7265942c4fad16f0f3"} Jan 26 17:11:01 crc kubenswrapper[4754]: I0126 17:11:01.968439 4754 generic.go:334] "Generic (PLEG): container finished" podID="54617228-f6c0-4db0-a468-2b9283bcae21" containerID="81c700b92383b930ceaf735cdacfcdd6280d44d2817b951b810ae6a363a3efb3" exitCode=0 Jan 26 17:11:01 crc kubenswrapper[4754]: I0126 17:11:01.969431 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-njgdn" event={"ID":"54617228-f6c0-4db0-a468-2b9283bcae21","Type":"ContainerDied","Data":"81c700b92383b930ceaf735cdacfcdd6280d44d2817b951b810ae6a363a3efb3"} Jan 26 17:11:02 crc kubenswrapper[4754]: I0126 17:11:02.006918 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Jan 26 17:11:02 crc kubenswrapper[4754]: I0126 17:11:02.076795 4754 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="9b50d057-68a1-44dc-8204-0366161e3cf4" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.193:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Jan 26 17:11:02 crc kubenswrapper[4754]: I0126 17:11:02.076886 4754 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="9b50d057-68a1-44dc-8204-0366161e3cf4" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.193:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Jan 26 17:11:02 crc kubenswrapper[4754]: I0126 17:11:02.549314 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Jan 26 17:11:02 crc kubenswrapper[4754]: I0126 17:11:02.563243 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-tnfmz" Jan 26 17:11:02 crc kubenswrapper[4754]: I0126 17:11:02.737569 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/952095c1-837b-4ff5-b61c-ce1e42102b56-dns-svc\") pod \"952095c1-837b-4ff5-b61c-ce1e42102b56\" (UID: \"952095c1-837b-4ff5-b61c-ce1e42102b56\") " Jan 26 17:11:02 crc kubenswrapper[4754]: I0126 17:11:02.737670 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-njfcd\" (UniqueName: \"kubernetes.io/projected/eca0fcbc-dfb5-4bc6-9c26-621230282fb4-kube-api-access-njfcd\") pod \"eca0fcbc-dfb5-4bc6-9c26-621230282fb4\" (UID: \"eca0fcbc-dfb5-4bc6-9c26-621230282fb4\") " Jan 26 17:11:02 crc kubenswrapper[4754]: I0126 17:11:02.737738 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kr2nr\" (UniqueName: \"kubernetes.io/projected/952095c1-837b-4ff5-b61c-ce1e42102b56-kube-api-access-kr2nr\") pod \"952095c1-837b-4ff5-b61c-ce1e42102b56\" (UID: \"952095c1-837b-4ff5-b61c-ce1e42102b56\") " Jan 26 17:11:02 crc kubenswrapper[4754]: I0126 17:11:02.737769 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eca0fcbc-dfb5-4bc6-9c26-621230282fb4-combined-ca-bundle\") pod \"eca0fcbc-dfb5-4bc6-9c26-621230282fb4\" (UID: \"eca0fcbc-dfb5-4bc6-9c26-621230282fb4\") " Jan 26 17:11:02 crc kubenswrapper[4754]: I0126 17:11:02.737794 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/952095c1-837b-4ff5-b61c-ce1e42102b56-dns-swift-storage-0\") pod \"952095c1-837b-4ff5-b61c-ce1e42102b56\" (UID: \"952095c1-837b-4ff5-b61c-ce1e42102b56\") " Jan 26 17:11:02 crc kubenswrapper[4754]: I0126 17:11:02.737847 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/952095c1-837b-4ff5-b61c-ce1e42102b56-config\") pod \"952095c1-837b-4ff5-b61c-ce1e42102b56\" (UID: \"952095c1-837b-4ff5-b61c-ce1e42102b56\") " Jan 26 17:11:02 crc kubenswrapper[4754]: I0126 17:11:02.737875 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eca0fcbc-dfb5-4bc6-9c26-621230282fb4-config-data\") pod \"eca0fcbc-dfb5-4bc6-9c26-621230282fb4\" (UID: \"eca0fcbc-dfb5-4bc6-9c26-621230282fb4\") " Jan 26 17:11:02 crc kubenswrapper[4754]: I0126 17:11:02.737934 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/eca0fcbc-dfb5-4bc6-9c26-621230282fb4-logs\") pod \"eca0fcbc-dfb5-4bc6-9c26-621230282fb4\" (UID: \"eca0fcbc-dfb5-4bc6-9c26-621230282fb4\") " Jan 26 17:11:02 crc kubenswrapper[4754]: I0126 17:11:02.737977 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/952095c1-837b-4ff5-b61c-ce1e42102b56-ovsdbserver-nb\") pod \"952095c1-837b-4ff5-b61c-ce1e42102b56\" (UID: \"952095c1-837b-4ff5-b61c-ce1e42102b56\") " Jan 26 17:11:02 crc kubenswrapper[4754]: I0126 17:11:02.738029 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/952095c1-837b-4ff5-b61c-ce1e42102b56-ovsdbserver-sb\") pod \"952095c1-837b-4ff5-b61c-ce1e42102b56\" (UID: \"952095c1-837b-4ff5-b61c-ce1e42102b56\") " Jan 26 17:11:02 crc kubenswrapper[4754]: I0126 17:11:02.738759 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eca0fcbc-dfb5-4bc6-9c26-621230282fb4-logs" (OuterVolumeSpecName: "logs") pod "eca0fcbc-dfb5-4bc6-9c26-621230282fb4" (UID: "eca0fcbc-dfb5-4bc6-9c26-621230282fb4"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:11:02 crc kubenswrapper[4754]: I0126 17:11:02.752869 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Jan 26 17:11:02 crc kubenswrapper[4754]: I0126 17:11:02.753083 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="c3c55551-6de7-403e-bbf5-46a2ec685c13" containerName="kube-state-metrics" containerID="cri-o://4774b9533f2557b915630e35e4b458e258217842d027febdef6f54ade9c3e270" gracePeriod=30 Jan 26 17:11:02 crc kubenswrapper[4754]: I0126 17:11:02.755926 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eca0fcbc-dfb5-4bc6-9c26-621230282fb4-kube-api-access-njfcd" (OuterVolumeSpecName: "kube-api-access-njfcd") pod "eca0fcbc-dfb5-4bc6-9c26-621230282fb4" (UID: "eca0fcbc-dfb5-4bc6-9c26-621230282fb4"). InnerVolumeSpecName "kube-api-access-njfcd". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:11:02 crc kubenswrapper[4754]: I0126 17:11:02.759542 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/952095c1-837b-4ff5-b61c-ce1e42102b56-kube-api-access-kr2nr" (OuterVolumeSpecName: "kube-api-access-kr2nr") pod "952095c1-837b-4ff5-b61c-ce1e42102b56" (UID: "952095c1-837b-4ff5-b61c-ce1e42102b56"). InnerVolumeSpecName "kube-api-access-kr2nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:11:02 crc kubenswrapper[4754]: I0126 17:11:02.817671 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eca0fcbc-dfb5-4bc6-9c26-621230282fb4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "eca0fcbc-dfb5-4bc6-9c26-621230282fb4" (UID: "eca0fcbc-dfb5-4bc6-9c26-621230282fb4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:11:02 crc kubenswrapper[4754]: I0126 17:11:02.840022 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-njfcd\" (UniqueName: \"kubernetes.io/projected/eca0fcbc-dfb5-4bc6-9c26-621230282fb4-kube-api-access-njfcd\") on node \"crc\" DevicePath \"\"" Jan 26 17:11:02 crc kubenswrapper[4754]: I0126 17:11:02.840061 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kr2nr\" (UniqueName: \"kubernetes.io/projected/952095c1-837b-4ff5-b61c-ce1e42102b56-kube-api-access-kr2nr\") on node \"crc\" DevicePath \"\"" Jan 26 17:11:02 crc kubenswrapper[4754]: I0126 17:11:02.840073 4754 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eca0fcbc-dfb5-4bc6-9c26-621230282fb4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 26 17:11:02 crc kubenswrapper[4754]: I0126 17:11:02.840093 4754 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/eca0fcbc-dfb5-4bc6-9c26-621230282fb4-logs\") on node \"crc\" DevicePath \"\"" Jan 26 17:11:02 crc kubenswrapper[4754]: I0126 17:11:02.844736 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eca0fcbc-dfb5-4bc6-9c26-621230282fb4-config-data" (OuterVolumeSpecName: "config-data") pod "eca0fcbc-dfb5-4bc6-9c26-621230282fb4" (UID: "eca0fcbc-dfb5-4bc6-9c26-621230282fb4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:11:02 crc kubenswrapper[4754]: I0126 17:11:02.853301 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/952095c1-837b-4ff5-b61c-ce1e42102b56-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "952095c1-837b-4ff5-b61c-ce1e42102b56" (UID: "952095c1-837b-4ff5-b61c-ce1e42102b56"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:11:02 crc kubenswrapper[4754]: I0126 17:11:02.855127 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/952095c1-837b-4ff5-b61c-ce1e42102b56-config" (OuterVolumeSpecName: "config") pod "952095c1-837b-4ff5-b61c-ce1e42102b56" (UID: "952095c1-837b-4ff5-b61c-ce1e42102b56"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:11:02 crc kubenswrapper[4754]: I0126 17:11:02.863173 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/952095c1-837b-4ff5-b61c-ce1e42102b56-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "952095c1-837b-4ff5-b61c-ce1e42102b56" (UID: "952095c1-837b-4ff5-b61c-ce1e42102b56"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:11:02 crc kubenswrapper[4754]: I0126 17:11:02.875464 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/952095c1-837b-4ff5-b61c-ce1e42102b56-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "952095c1-837b-4ff5-b61c-ce1e42102b56" (UID: "952095c1-837b-4ff5-b61c-ce1e42102b56"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:11:02 crc kubenswrapper[4754]: I0126 17:11:02.876424 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/952095c1-837b-4ff5-b61c-ce1e42102b56-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "952095c1-837b-4ff5-b61c-ce1e42102b56" (UID: "952095c1-837b-4ff5-b61c-ce1e42102b56"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:11:02 crc kubenswrapper[4754]: I0126 17:11:02.943300 4754 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/952095c1-837b-4ff5-b61c-ce1e42102b56-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Jan 26 17:11:02 crc kubenswrapper[4754]: I0126 17:11:02.943349 4754 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/952095c1-837b-4ff5-b61c-ce1e42102b56-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Jan 26 17:11:02 crc kubenswrapper[4754]: I0126 17:11:02.943361 4754 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/952095c1-837b-4ff5-b61c-ce1e42102b56-dns-svc\") on node \"crc\" DevicePath \"\"" Jan 26 17:11:02 crc kubenswrapper[4754]: I0126 17:11:02.943372 4754 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/952095c1-837b-4ff5-b61c-ce1e42102b56-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Jan 26 17:11:02 crc kubenswrapper[4754]: I0126 17:11:02.943383 4754 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/952095c1-837b-4ff5-b61c-ce1e42102b56-config\") on node \"crc\" DevicePath \"\"" Jan 26 17:11:02 crc kubenswrapper[4754]: I0126 17:11:02.943392 4754 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eca0fcbc-dfb5-4bc6-9c26-621230282fb4-config-data\") on node \"crc\" DevicePath \"\"" Jan 26 17:11:02 crc kubenswrapper[4754]: I0126 17:11:02.978910 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-tnfmz" event={"ID":"952095c1-837b-4ff5-b61c-ce1e42102b56","Type":"ContainerDied","Data":"373bbf9cbb4c96b3d33d178926090aed6931500656ad7c69799b9b4486f29eb0"} Jan 26 17:11:02 crc kubenswrapper[4754]: I0126 17:11:02.978970 4754 scope.go:117] "RemoveContainer" containerID="a9080148a080f6c8a5352be74755999d3eb0d22a9a34cdd78cd8072a9a4d3f4f" Jan 26 17:11:02 crc kubenswrapper[4754]: I0126 17:11:02.978924 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-tnfmz" Jan 26 17:11:02 crc kubenswrapper[4754]: I0126 17:11:02.983094 4754 generic.go:334] "Generic (PLEG): container finished" podID="c3c55551-6de7-403e-bbf5-46a2ec685c13" containerID="4774b9533f2557b915630e35e4b458e258217842d027febdef6f54ade9c3e270" exitCode=2 Jan 26 17:11:02 crc kubenswrapper[4754]: I0126 17:11:02.983180 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"c3c55551-6de7-403e-bbf5-46a2ec685c13","Type":"ContainerDied","Data":"4774b9533f2557b915630e35e4b458e258217842d027febdef6f54ade9c3e270"} Jan 26 17:11:02 crc kubenswrapper[4754]: I0126 17:11:02.986060 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"eca0fcbc-dfb5-4bc6-9c26-621230282fb4","Type":"ContainerDied","Data":"00cfd6e217217ec30da0c6286e4beaf21f39d4f682b28d09e318642ab958b986"} Jan 26 17:11:02 crc kubenswrapper[4754]: I0126 17:11:02.986194 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Jan 26 17:11:02 crc kubenswrapper[4754]: I0126 17:11:02.988227 4754 generic.go:334] "Generic (PLEG): container finished" podID="e3881c45-db75-4ca1-bb8f-92f413446265" containerID="6995e4eda258b809a08ef7ec99095737aca54eadc07eb6db51f9cb1dc5845921" exitCode=0 Jan 26 17:11:02 crc kubenswrapper[4754]: I0126 17:11:02.989590 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hxdzc" event={"ID":"e3881c45-db75-4ca1-bb8f-92f413446265","Type":"ContainerDied","Data":"6995e4eda258b809a08ef7ec99095737aca54eadc07eb6db51f9cb1dc5845921"} Jan 26 17:11:03 crc kubenswrapper[4754]: I0126 17:11:03.014776 4754 scope.go:117] "RemoveContainer" containerID="86bc9c5562701990e9174a443eecc059bb82afee194d1bfe69d2a52c8e8dc8ed" Jan 26 17:11:03 crc kubenswrapper[4754]: I0126 17:11:03.064318 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-tnfmz"] Jan 26 17:11:03 crc kubenswrapper[4754]: I0126 17:11:03.080008 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-tnfmz"] Jan 26 17:11:03 crc kubenswrapper[4754]: I0126 17:11:03.094413 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Jan 26 17:11:03 crc kubenswrapper[4754]: I0126 17:11:03.103140 4754 scope.go:117] "RemoveContainer" containerID="d81cba4378ff840952fbd04ffd402cdc93da3a8fce7bfb7265942c4fad16f0f3" Jan 26 17:11:03 crc kubenswrapper[4754]: I0126 17:11:03.105257 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Jan 26 17:11:03 crc kubenswrapper[4754]: I0126 17:11:03.115419 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Jan 26 17:11:03 crc kubenswrapper[4754]: E0126 17:11:03.115838 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eca0fcbc-dfb5-4bc6-9c26-621230282fb4" containerName="nova-metadata-metadata" Jan 26 17:11:03 crc kubenswrapper[4754]: I0126 17:11:03.115855 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="eca0fcbc-dfb5-4bc6-9c26-621230282fb4" containerName="nova-metadata-metadata" Jan 26 17:11:03 crc kubenswrapper[4754]: E0126 17:11:03.115871 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eca0fcbc-dfb5-4bc6-9c26-621230282fb4" containerName="nova-metadata-log" Jan 26 17:11:03 crc kubenswrapper[4754]: I0126 17:11:03.115878 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="eca0fcbc-dfb5-4bc6-9c26-621230282fb4" containerName="nova-metadata-log" Jan 26 17:11:03 crc kubenswrapper[4754]: E0126 17:11:03.115898 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="952095c1-837b-4ff5-b61c-ce1e42102b56" containerName="dnsmasq-dns" Jan 26 17:11:03 crc kubenswrapper[4754]: I0126 17:11:03.115904 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="952095c1-837b-4ff5-b61c-ce1e42102b56" containerName="dnsmasq-dns" Jan 26 17:11:03 crc kubenswrapper[4754]: E0126 17:11:03.115914 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="952095c1-837b-4ff5-b61c-ce1e42102b56" containerName="init" Jan 26 17:11:03 crc kubenswrapper[4754]: I0126 17:11:03.115919 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="952095c1-837b-4ff5-b61c-ce1e42102b56" containerName="init" Jan 26 17:11:03 crc kubenswrapper[4754]: I0126 17:11:03.116217 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="952095c1-837b-4ff5-b61c-ce1e42102b56" containerName="dnsmasq-dns" Jan 26 17:11:03 crc kubenswrapper[4754]: I0126 17:11:03.116266 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="eca0fcbc-dfb5-4bc6-9c26-621230282fb4" containerName="nova-metadata-log" Jan 26 17:11:03 crc kubenswrapper[4754]: I0126 17:11:03.116282 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="eca0fcbc-dfb5-4bc6-9c26-621230282fb4" containerName="nova-metadata-metadata" Jan 26 17:11:03 crc kubenswrapper[4754]: I0126 17:11:03.117194 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Jan 26 17:11:03 crc kubenswrapper[4754]: I0126 17:11:03.119540 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Jan 26 17:11:03 crc kubenswrapper[4754]: I0126 17:11:03.120119 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Jan 26 17:11:03 crc kubenswrapper[4754]: I0126 17:11:03.128460 4754 scope.go:117] "RemoveContainer" containerID="7d9708cd4813226778c6001f4e7a76bf20c0d89caecec18eeae7090cd5058c94" Jan 26 17:11:03 crc kubenswrapper[4754]: I0126 17:11:03.130506 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Jan 26 17:11:03 crc kubenswrapper[4754]: I0126 17:11:03.248499 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/a5a89a1e-ce5c-4e31-9919-c43880fe6c04-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"a5a89a1e-ce5c-4e31-9919-c43880fe6c04\") " pod="openstack/nova-metadata-0" Jan 26 17:11:03 crc kubenswrapper[4754]: I0126 17:11:03.248590 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bgxwk\" (UniqueName: \"kubernetes.io/projected/a5a89a1e-ce5c-4e31-9919-c43880fe6c04-kube-api-access-bgxwk\") pod \"nova-metadata-0\" (UID: \"a5a89a1e-ce5c-4e31-9919-c43880fe6c04\") " pod="openstack/nova-metadata-0" Jan 26 17:11:03 crc kubenswrapper[4754]: I0126 17:11:03.248645 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a5a89a1e-ce5c-4e31-9919-c43880fe6c04-config-data\") pod \"nova-metadata-0\" (UID: \"a5a89a1e-ce5c-4e31-9919-c43880fe6c04\") " pod="openstack/nova-metadata-0" Jan 26 17:11:03 crc kubenswrapper[4754]: I0126 17:11:03.249445 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5a89a1e-ce5c-4e31-9919-c43880fe6c04-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"a5a89a1e-ce5c-4e31-9919-c43880fe6c04\") " pod="openstack/nova-metadata-0" Jan 26 17:11:03 crc kubenswrapper[4754]: I0126 17:11:03.249740 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a5a89a1e-ce5c-4e31-9919-c43880fe6c04-logs\") pod \"nova-metadata-0\" (UID: \"a5a89a1e-ce5c-4e31-9919-c43880fe6c04\") " pod="openstack/nova-metadata-0" Jan 26 17:11:03 crc kubenswrapper[4754]: I0126 17:11:03.282154 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Jan 26 17:11:03 crc kubenswrapper[4754]: I0126 17:11:03.351560 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a5a89a1e-ce5c-4e31-9919-c43880fe6c04-logs\") pod \"nova-metadata-0\" (UID: \"a5a89a1e-ce5c-4e31-9919-c43880fe6c04\") " pod="openstack/nova-metadata-0" Jan 26 17:11:03 crc kubenswrapper[4754]: I0126 17:11:03.351637 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/a5a89a1e-ce5c-4e31-9919-c43880fe6c04-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"a5a89a1e-ce5c-4e31-9919-c43880fe6c04\") " pod="openstack/nova-metadata-0" Jan 26 17:11:03 crc kubenswrapper[4754]: I0126 17:11:03.351662 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bgxwk\" (UniqueName: \"kubernetes.io/projected/a5a89a1e-ce5c-4e31-9919-c43880fe6c04-kube-api-access-bgxwk\") pod \"nova-metadata-0\" (UID: \"a5a89a1e-ce5c-4e31-9919-c43880fe6c04\") " pod="openstack/nova-metadata-0" Jan 26 17:11:03 crc kubenswrapper[4754]: I0126 17:11:03.351708 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a5a89a1e-ce5c-4e31-9919-c43880fe6c04-config-data\") pod \"nova-metadata-0\" (UID: \"a5a89a1e-ce5c-4e31-9919-c43880fe6c04\") " pod="openstack/nova-metadata-0" Jan 26 17:11:03 crc kubenswrapper[4754]: I0126 17:11:03.351792 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5a89a1e-ce5c-4e31-9919-c43880fe6c04-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"a5a89a1e-ce5c-4e31-9919-c43880fe6c04\") " pod="openstack/nova-metadata-0" Jan 26 17:11:03 crc kubenswrapper[4754]: I0126 17:11:03.355294 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a5a89a1e-ce5c-4e31-9919-c43880fe6c04-logs\") pod \"nova-metadata-0\" (UID: \"a5a89a1e-ce5c-4e31-9919-c43880fe6c04\") " pod="openstack/nova-metadata-0" Jan 26 17:11:03 crc kubenswrapper[4754]: I0126 17:11:03.358718 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/a5a89a1e-ce5c-4e31-9919-c43880fe6c04-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"a5a89a1e-ce5c-4e31-9919-c43880fe6c04\") " pod="openstack/nova-metadata-0" Jan 26 17:11:03 crc kubenswrapper[4754]: I0126 17:11:03.361660 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a5a89a1e-ce5c-4e31-9919-c43880fe6c04-config-data\") pod \"nova-metadata-0\" (UID: \"a5a89a1e-ce5c-4e31-9919-c43880fe6c04\") " pod="openstack/nova-metadata-0" Jan 26 17:11:03 crc kubenswrapper[4754]: I0126 17:11:03.362301 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5a89a1e-ce5c-4e31-9919-c43880fe6c04-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"a5a89a1e-ce5c-4e31-9919-c43880fe6c04\") " pod="openstack/nova-metadata-0" Jan 26 17:11:03 crc kubenswrapper[4754]: I0126 17:11:03.376364 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bgxwk\" (UniqueName: \"kubernetes.io/projected/a5a89a1e-ce5c-4e31-9919-c43880fe6c04-kube-api-access-bgxwk\") pod \"nova-metadata-0\" (UID: \"a5a89a1e-ce5c-4e31-9919-c43880fe6c04\") " pod="openstack/nova-metadata-0" Jan 26 17:11:03 crc kubenswrapper[4754]: I0126 17:11:03.403185 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-njgdn" Jan 26 17:11:03 crc kubenswrapper[4754]: I0126 17:11:03.444764 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Jan 26 17:11:03 crc kubenswrapper[4754]: I0126 17:11:03.453779 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5lddx\" (UniqueName: \"kubernetes.io/projected/c3c55551-6de7-403e-bbf5-46a2ec685c13-kube-api-access-5lddx\") pod \"c3c55551-6de7-403e-bbf5-46a2ec685c13\" (UID: \"c3c55551-6de7-403e-bbf5-46a2ec685c13\") " Jan 26 17:11:03 crc kubenswrapper[4754]: I0126 17:11:03.459301 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c3c55551-6de7-403e-bbf5-46a2ec685c13-kube-api-access-5lddx" (OuterVolumeSpecName: "kube-api-access-5lddx") pod "c3c55551-6de7-403e-bbf5-46a2ec685c13" (UID: "c3c55551-6de7-403e-bbf5-46a2ec685c13"). InnerVolumeSpecName "kube-api-access-5lddx". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:11:03 crc kubenswrapper[4754]: I0126 17:11:03.556208 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/54617228-f6c0-4db0-a468-2b9283bcae21-scripts\") pod \"54617228-f6c0-4db0-a468-2b9283bcae21\" (UID: \"54617228-f6c0-4db0-a468-2b9283bcae21\") " Jan 26 17:11:03 crc kubenswrapper[4754]: I0126 17:11:03.556280 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54617228-f6c0-4db0-a468-2b9283bcae21-combined-ca-bundle\") pod \"54617228-f6c0-4db0-a468-2b9283bcae21\" (UID: \"54617228-f6c0-4db0-a468-2b9283bcae21\") " Jan 26 17:11:03 crc kubenswrapper[4754]: I0126 17:11:03.556346 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/54617228-f6c0-4db0-a468-2b9283bcae21-config-data\") pod \"54617228-f6c0-4db0-a468-2b9283bcae21\" (UID: \"54617228-f6c0-4db0-a468-2b9283bcae21\") " Jan 26 17:11:03 crc kubenswrapper[4754]: I0126 17:11:03.556379 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9br5h\" (UniqueName: \"kubernetes.io/projected/54617228-f6c0-4db0-a468-2b9283bcae21-kube-api-access-9br5h\") pod \"54617228-f6c0-4db0-a468-2b9283bcae21\" (UID: \"54617228-f6c0-4db0-a468-2b9283bcae21\") " Jan 26 17:11:03 crc kubenswrapper[4754]: I0126 17:11:03.557031 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5lddx\" (UniqueName: \"kubernetes.io/projected/c3c55551-6de7-403e-bbf5-46a2ec685c13-kube-api-access-5lddx\") on node \"crc\" DevicePath \"\"" Jan 26 17:11:03 crc kubenswrapper[4754]: I0126 17:11:03.561886 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/54617228-f6c0-4db0-a468-2b9283bcae21-kube-api-access-9br5h" (OuterVolumeSpecName: "kube-api-access-9br5h") pod "54617228-f6c0-4db0-a468-2b9283bcae21" (UID: "54617228-f6c0-4db0-a468-2b9283bcae21"). InnerVolumeSpecName "kube-api-access-9br5h". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:11:03 crc kubenswrapper[4754]: I0126 17:11:03.565232 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/54617228-f6c0-4db0-a468-2b9283bcae21-scripts" (OuterVolumeSpecName: "scripts") pod "54617228-f6c0-4db0-a468-2b9283bcae21" (UID: "54617228-f6c0-4db0-a468-2b9283bcae21"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:11:03 crc kubenswrapper[4754]: I0126 17:11:03.590222 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/54617228-f6c0-4db0-a468-2b9283bcae21-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "54617228-f6c0-4db0-a468-2b9283bcae21" (UID: "54617228-f6c0-4db0-a468-2b9283bcae21"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:11:03 crc kubenswrapper[4754]: I0126 17:11:03.598352 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/54617228-f6c0-4db0-a468-2b9283bcae21-config-data" (OuterVolumeSpecName: "config-data") pod "54617228-f6c0-4db0-a468-2b9283bcae21" (UID: "54617228-f6c0-4db0-a468-2b9283bcae21"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:11:03 crc kubenswrapper[4754]: I0126 17:11:03.658926 4754 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54617228-f6c0-4db0-a468-2b9283bcae21-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 26 17:11:03 crc kubenswrapper[4754]: I0126 17:11:03.658956 4754 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/54617228-f6c0-4db0-a468-2b9283bcae21-config-data\") on node \"crc\" DevicePath \"\"" Jan 26 17:11:03 crc kubenswrapper[4754]: I0126 17:11:03.658967 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9br5h\" (UniqueName: \"kubernetes.io/projected/54617228-f6c0-4db0-a468-2b9283bcae21-kube-api-access-9br5h\") on node \"crc\" DevicePath \"\"" Jan 26 17:11:03 crc kubenswrapper[4754]: I0126 17:11:03.658975 4754 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/54617228-f6c0-4db0-a468-2b9283bcae21-scripts\") on node \"crc\" DevicePath \"\"" Jan 26 17:11:03 crc kubenswrapper[4754]: I0126 17:11:03.778351 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="952095c1-837b-4ff5-b61c-ce1e42102b56" path="/var/lib/kubelet/pods/952095c1-837b-4ff5-b61c-ce1e42102b56/volumes" Jan 26 17:11:03 crc kubenswrapper[4754]: I0126 17:11:03.779416 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eca0fcbc-dfb5-4bc6-9c26-621230282fb4" path="/var/lib/kubelet/pods/eca0fcbc-dfb5-4bc6-9c26-621230282fb4/volumes" Jan 26 17:11:03 crc kubenswrapper[4754]: I0126 17:11:03.939940 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Jan 26 17:11:03 crc kubenswrapper[4754]: W0126 17:11:03.942894 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda5a89a1e_ce5c_4e31_9919_c43880fe6c04.slice/crio-652d72177417a349bcc1adadeda2b20ab4edd0e2ad27917d3be625238cd214bf WatchSource:0}: Error finding container 652d72177417a349bcc1adadeda2b20ab4edd0e2ad27917d3be625238cd214bf: Status 404 returned error can't find the container with id 652d72177417a349bcc1adadeda2b20ab4edd0e2ad27917d3be625238cd214bf Jan 26 17:11:04 crc kubenswrapper[4754]: I0126 17:11:04.000258 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a5a89a1e-ce5c-4e31-9919-c43880fe6c04","Type":"ContainerStarted","Data":"652d72177417a349bcc1adadeda2b20ab4edd0e2ad27917d3be625238cd214bf"} Jan 26 17:11:04 crc kubenswrapper[4754]: I0126 17:11:04.004136 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"c3c55551-6de7-403e-bbf5-46a2ec685c13","Type":"ContainerDied","Data":"f43ccdd4b774f3fdbbf1ff52ffd4e67686633eb09604f68fd5c9a4b779e1f724"} Jan 26 17:11:04 crc kubenswrapper[4754]: I0126 17:11:04.004204 4754 scope.go:117] "RemoveContainer" containerID="4774b9533f2557b915630e35e4b458e258217842d027febdef6f54ade9c3e270" Jan 26 17:11:04 crc kubenswrapper[4754]: I0126 17:11:04.004157 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Jan 26 17:11:04 crc kubenswrapper[4754]: I0126 17:11:04.011459 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-njgdn" event={"ID":"54617228-f6c0-4db0-a468-2b9283bcae21","Type":"ContainerDied","Data":"e910b0c7089f8278c41b5fc19bdb746ba8ab6a228cf4defa17621078ef4cfdc9"} Jan 26 17:11:04 crc kubenswrapper[4754]: I0126 17:11:04.011506 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e910b0c7089f8278c41b5fc19bdb746ba8ab6a228cf4defa17621078ef4cfdc9" Jan 26 17:11:04 crc kubenswrapper[4754]: I0126 17:11:04.012368 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-njgdn" Jan 26 17:11:04 crc kubenswrapper[4754]: I0126 17:11:04.040562 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hxdzc" event={"ID":"e3881c45-db75-4ca1-bb8f-92f413446265","Type":"ContainerStarted","Data":"df4175ce477fff9b340f1ab44f40a75ad91b861e26473ef821f2251779a39ea4"} Jan 26 17:11:04 crc kubenswrapper[4754]: I0126 17:11:04.046744 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Jan 26 17:11:04 crc kubenswrapper[4754]: I0126 17:11:04.057739 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Jan 26 17:11:04 crc kubenswrapper[4754]: I0126 17:11:04.107380 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Jan 26 17:11:04 crc kubenswrapper[4754]: E0126 17:11:04.108303 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54617228-f6c0-4db0-a468-2b9283bcae21" containerName="nova-manage" Jan 26 17:11:04 crc kubenswrapper[4754]: I0126 17:11:04.108328 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="54617228-f6c0-4db0-a468-2b9283bcae21" containerName="nova-manage" Jan 26 17:11:04 crc kubenswrapper[4754]: E0126 17:11:04.108365 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c3c55551-6de7-403e-bbf5-46a2ec685c13" containerName="kube-state-metrics" Jan 26 17:11:04 crc kubenswrapper[4754]: I0126 17:11:04.108374 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="c3c55551-6de7-403e-bbf5-46a2ec685c13" containerName="kube-state-metrics" Jan 26 17:11:04 crc kubenswrapper[4754]: I0126 17:11:04.128209 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="54617228-f6c0-4db0-a468-2b9283bcae21" containerName="nova-manage" Jan 26 17:11:04 crc kubenswrapper[4754]: I0126 17:11:04.128275 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="c3c55551-6de7-403e-bbf5-46a2ec685c13" containerName="kube-state-metrics" Jan 26 17:11:04 crc kubenswrapper[4754]: I0126 17:11:04.130027 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Jan 26 17:11:04 crc kubenswrapper[4754]: I0126 17:11:04.130141 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Jan 26 17:11:04 crc kubenswrapper[4754]: I0126 17:11:04.134920 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Jan 26 17:11:04 crc kubenswrapper[4754]: I0126 17:11:04.135182 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Jan 26 17:11:04 crc kubenswrapper[4754]: I0126 17:11:04.189151 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Jan 26 17:11:04 crc kubenswrapper[4754]: I0126 17:11:04.189424 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="9b50d057-68a1-44dc-8204-0366161e3cf4" containerName="nova-api-log" containerID="cri-o://77aafa83659a5fb388cddee05c48d3c0ee10ca31c83aecb39a67800530e24940" gracePeriod=30 Jan 26 17:11:04 crc kubenswrapper[4754]: I0126 17:11:04.189842 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="9b50d057-68a1-44dc-8204-0366161e3cf4" containerName="nova-api-api" containerID="cri-o://4c0f6a01ffaef43ea38f623749ef8fb2a10ea1b37845de8eae86906116547dc6" gracePeriod=30 Jan 26 17:11:04 crc kubenswrapper[4754]: I0126 17:11:04.211520 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Jan 26 17:11:04 crc kubenswrapper[4754]: I0126 17:11:04.225055 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Jan 26 17:11:04 crc kubenswrapper[4754]: I0126 17:11:04.272272 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/6be68290-e783-4f39-8257-d5dc4051447a-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"6be68290-e783-4f39-8257-d5dc4051447a\") " pod="openstack/kube-state-metrics-0" Jan 26 17:11:04 crc kubenswrapper[4754]: I0126 17:11:04.272359 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6be68290-e783-4f39-8257-d5dc4051447a-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"6be68290-e783-4f39-8257-d5dc4051447a\") " pod="openstack/kube-state-metrics-0" Jan 26 17:11:04 crc kubenswrapper[4754]: I0126 17:11:04.272457 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/6be68290-e783-4f39-8257-d5dc4051447a-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"6be68290-e783-4f39-8257-d5dc4051447a\") " pod="openstack/kube-state-metrics-0" Jan 26 17:11:04 crc kubenswrapper[4754]: I0126 17:11:04.272579 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c27dk\" (UniqueName: \"kubernetes.io/projected/6be68290-e783-4f39-8257-d5dc4051447a-kube-api-access-c27dk\") pod \"kube-state-metrics-0\" (UID: \"6be68290-e783-4f39-8257-d5dc4051447a\") " pod="openstack/kube-state-metrics-0" Jan 26 17:11:04 crc kubenswrapper[4754]: I0126 17:11:04.374566 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/6be68290-e783-4f39-8257-d5dc4051447a-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"6be68290-e783-4f39-8257-d5dc4051447a\") " pod="openstack/kube-state-metrics-0" Jan 26 17:11:04 crc kubenswrapper[4754]: I0126 17:11:04.374959 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c27dk\" (UniqueName: \"kubernetes.io/projected/6be68290-e783-4f39-8257-d5dc4051447a-kube-api-access-c27dk\") pod \"kube-state-metrics-0\" (UID: \"6be68290-e783-4f39-8257-d5dc4051447a\") " pod="openstack/kube-state-metrics-0" Jan 26 17:11:04 crc kubenswrapper[4754]: I0126 17:11:04.375050 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/6be68290-e783-4f39-8257-d5dc4051447a-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"6be68290-e783-4f39-8257-d5dc4051447a\") " pod="openstack/kube-state-metrics-0" Jan 26 17:11:04 crc kubenswrapper[4754]: I0126 17:11:04.375108 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6be68290-e783-4f39-8257-d5dc4051447a-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"6be68290-e783-4f39-8257-d5dc4051447a\") " pod="openstack/kube-state-metrics-0" Jan 26 17:11:04 crc kubenswrapper[4754]: I0126 17:11:04.380215 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/6be68290-e783-4f39-8257-d5dc4051447a-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"6be68290-e783-4f39-8257-d5dc4051447a\") " pod="openstack/kube-state-metrics-0" Jan 26 17:11:04 crc kubenswrapper[4754]: I0126 17:11:04.380285 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6be68290-e783-4f39-8257-d5dc4051447a-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"6be68290-e783-4f39-8257-d5dc4051447a\") " pod="openstack/kube-state-metrics-0" Jan 26 17:11:04 crc kubenswrapper[4754]: I0126 17:11:04.383159 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/6be68290-e783-4f39-8257-d5dc4051447a-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"6be68290-e783-4f39-8257-d5dc4051447a\") " pod="openstack/kube-state-metrics-0" Jan 26 17:11:04 crc kubenswrapper[4754]: I0126 17:11:04.397811 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c27dk\" (UniqueName: \"kubernetes.io/projected/6be68290-e783-4f39-8257-d5dc4051447a-kube-api-access-c27dk\") pod \"kube-state-metrics-0\" (UID: \"6be68290-e783-4f39-8257-d5dc4051447a\") " pod="openstack/kube-state-metrics-0" Jan 26 17:11:04 crc kubenswrapper[4754]: I0126 17:11:04.470856 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Jan 26 17:11:04 crc kubenswrapper[4754]: I0126 17:11:04.895694 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Jan 26 17:11:04 crc kubenswrapper[4754]: I0126 17:11:04.896129 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="6d0563ff-f0f4-45b7-aaba-014f4426f4b0" containerName="ceilometer-central-agent" containerID="cri-o://eb80b00f37f12db37a6f667c832c82d481c8513f0c10b367ce2bb2cb977c988b" gracePeriod=30 Jan 26 17:11:04 crc kubenswrapper[4754]: I0126 17:11:04.896496 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="6d0563ff-f0f4-45b7-aaba-014f4426f4b0" containerName="proxy-httpd" containerID="cri-o://ebd4a3041094da9f63be25f198192702e8ece75b22e3c5c564ca86dfb8a0148c" gracePeriod=30 Jan 26 17:11:04 crc kubenswrapper[4754]: I0126 17:11:04.896551 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="6d0563ff-f0f4-45b7-aaba-014f4426f4b0" containerName="sg-core" containerID="cri-o://2450afc5412f20e482cf129b8e70df60fec15191d846f728199beaa14ae587fe" gracePeriod=30 Jan 26 17:11:04 crc kubenswrapper[4754]: I0126 17:11:04.896583 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="6d0563ff-f0f4-45b7-aaba-014f4426f4b0" containerName="ceilometer-notification-agent" containerID="cri-o://234f898b8e364e60561ed34cd7475fe072e0bd9490b1145b1909804af99b448d" gracePeriod=30 Jan 26 17:11:05 crc kubenswrapper[4754]: I0126 17:11:05.007347 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Jan 26 17:11:05 crc kubenswrapper[4754]: W0126 17:11:05.008907 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6be68290_e783_4f39_8257_d5dc4051447a.slice/crio-23cdee3c120f6fd2489d53ef5942c6a7c77db47e3e108936ffdd5134c682f120 WatchSource:0}: Error finding container 23cdee3c120f6fd2489d53ef5942c6a7c77db47e3e108936ffdd5134c682f120: Status 404 returned error can't find the container with id 23cdee3c120f6fd2489d53ef5942c6a7c77db47e3e108936ffdd5134c682f120 Jan 26 17:11:05 crc kubenswrapper[4754]: I0126 17:11:05.058171 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a5a89a1e-ce5c-4e31-9919-c43880fe6c04","Type":"ContainerStarted","Data":"3bda1766505c4134ecf969f9f1a8caf229decec57840fbbcfd89a88471c78aa2"} Jan 26 17:11:05 crc kubenswrapper[4754]: I0126 17:11:05.072954 4754 generic.go:334] "Generic (PLEG): container finished" podID="9b50d057-68a1-44dc-8204-0366161e3cf4" containerID="77aafa83659a5fb388cddee05c48d3c0ee10ca31c83aecb39a67800530e24940" exitCode=143 Jan 26 17:11:05 crc kubenswrapper[4754]: I0126 17:11:05.073036 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"9b50d057-68a1-44dc-8204-0366161e3cf4","Type":"ContainerDied","Data":"77aafa83659a5fb388cddee05c48d3c0ee10ca31c83aecb39a67800530e24940"} Jan 26 17:11:05 crc kubenswrapper[4754]: I0126 17:11:05.077854 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"6be68290-e783-4f39-8257-d5dc4051447a","Type":"ContainerStarted","Data":"23cdee3c120f6fd2489d53ef5942c6a7c77db47e3e108936ffdd5134c682f120"} Jan 26 17:11:05 crc kubenswrapper[4754]: I0126 17:11:05.078082 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="270ffa9f-f033-488c-887e-565ca34620c6" containerName="nova-scheduler-scheduler" containerID="cri-o://dbafe75f4149545f66d7dcf67e5679e62b6cad3503ae5ede59a6619920905833" gracePeriod=30 Jan 26 17:11:05 crc kubenswrapper[4754]: I0126 17:11:05.781099 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c3c55551-6de7-403e-bbf5-46a2ec685c13" path="/var/lib/kubelet/pods/c3c55551-6de7-403e-bbf5-46a2ec685c13/volumes" Jan 26 17:11:06 crc kubenswrapper[4754]: E0126 17:11:06.089255 4754 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="dbafe75f4149545f66d7dcf67e5679e62b6cad3503ae5ede59a6619920905833" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Jan 26 17:11:06 crc kubenswrapper[4754]: E0126 17:11:06.093921 4754 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="dbafe75f4149545f66d7dcf67e5679e62b6cad3503ae5ede59a6619920905833" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Jan 26 17:11:06 crc kubenswrapper[4754]: I0126 17:11:06.093989 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a5a89a1e-ce5c-4e31-9919-c43880fe6c04","Type":"ContainerStarted","Data":"0b335ff1e7d7ca9a9013152a903eb9bc32541c7f911804f273535fefecd8f2fc"} Jan 26 17:11:06 crc kubenswrapper[4754]: E0126 17:11:06.102846 4754 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="dbafe75f4149545f66d7dcf67e5679e62b6cad3503ae5ede59a6619920905833" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Jan 26 17:11:06 crc kubenswrapper[4754]: E0126 17:11:06.102916 4754 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="270ffa9f-f033-488c-887e-565ca34620c6" containerName="nova-scheduler-scheduler" Jan 26 17:11:07 crc kubenswrapper[4754]: I0126 17:11:07.105824 4754 generic.go:334] "Generic (PLEG): container finished" podID="6d0563ff-f0f4-45b7-aaba-014f4426f4b0" containerID="ebd4a3041094da9f63be25f198192702e8ece75b22e3c5c564ca86dfb8a0148c" exitCode=0 Jan 26 17:11:07 crc kubenswrapper[4754]: I0126 17:11:07.105859 4754 generic.go:334] "Generic (PLEG): container finished" podID="6d0563ff-f0f4-45b7-aaba-014f4426f4b0" containerID="2450afc5412f20e482cf129b8e70df60fec15191d846f728199beaa14ae587fe" exitCode=2 Jan 26 17:11:07 crc kubenswrapper[4754]: I0126 17:11:07.106797 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6d0563ff-f0f4-45b7-aaba-014f4426f4b0","Type":"ContainerDied","Data":"ebd4a3041094da9f63be25f198192702e8ece75b22e3c5c564ca86dfb8a0148c"} Jan 26 17:11:07 crc kubenswrapper[4754]: I0126 17:11:07.106916 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6d0563ff-f0f4-45b7-aaba-014f4426f4b0","Type":"ContainerDied","Data":"2450afc5412f20e482cf129b8e70df60fec15191d846f728199beaa14ae587fe"} Jan 26 17:11:07 crc kubenswrapper[4754]: E0126 17:11:07.299827 4754 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6d0563ff_f0f4_45b7_aaba_014f4426f4b0.slice/crio-conmon-eb80b00f37f12db37a6f667c832c82d481c8513f0c10b367ce2bb2cb977c988b.scope\": RecentStats: unable to find data in memory cache]" Jan 26 17:11:08 crc kubenswrapper[4754]: I0126 17:11:08.121281 4754 generic.go:334] "Generic (PLEG): container finished" podID="6d0563ff-f0f4-45b7-aaba-014f4426f4b0" containerID="eb80b00f37f12db37a6f667c832c82d481c8513f0c10b367ce2bb2cb977c988b" exitCode=0 Jan 26 17:11:08 crc kubenswrapper[4754]: I0126 17:11:08.121320 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6d0563ff-f0f4-45b7-aaba-014f4426f4b0","Type":"ContainerDied","Data":"eb80b00f37f12db37a6f667c832c82d481c8513f0c10b367ce2bb2cb977c988b"} Jan 26 17:11:08 crc kubenswrapper[4754]: I0126 17:11:08.125447 4754 generic.go:334] "Generic (PLEG): container finished" podID="9b50d057-68a1-44dc-8204-0366161e3cf4" containerID="4c0f6a01ffaef43ea38f623749ef8fb2a10ea1b37845de8eae86906116547dc6" exitCode=0 Jan 26 17:11:08 crc kubenswrapper[4754]: I0126 17:11:08.125509 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"9b50d057-68a1-44dc-8204-0366161e3cf4","Type":"ContainerDied","Data":"4c0f6a01ffaef43ea38f623749ef8fb2a10ea1b37845de8eae86906116547dc6"} Jan 26 17:11:08 crc kubenswrapper[4754]: I0126 17:11:08.128235 4754 generic.go:334] "Generic (PLEG): container finished" podID="e3881c45-db75-4ca1-bb8f-92f413446265" containerID="df4175ce477fff9b340f1ab44f40a75ad91b861e26473ef821f2251779a39ea4" exitCode=0 Jan 26 17:11:08 crc kubenswrapper[4754]: I0126 17:11:08.128286 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hxdzc" event={"ID":"e3881c45-db75-4ca1-bb8f-92f413446265","Type":"ContainerDied","Data":"df4175ce477fff9b340f1ab44f40a75ad91b861e26473ef821f2251779a39ea4"} Jan 26 17:11:08 crc kubenswrapper[4754]: I0126 17:11:08.128427 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="a5a89a1e-ce5c-4e31-9919-c43880fe6c04" containerName="nova-metadata-log" containerID="cri-o://3bda1766505c4134ecf969f9f1a8caf229decec57840fbbcfd89a88471c78aa2" gracePeriod=30 Jan 26 17:11:08 crc kubenswrapper[4754]: I0126 17:11:08.128532 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="a5a89a1e-ce5c-4e31-9919-c43880fe6c04" containerName="nova-metadata-metadata" containerID="cri-o://0b335ff1e7d7ca9a9013152a903eb9bc32541c7f911804f273535fefecd8f2fc" gracePeriod=30 Jan 26 17:11:08 crc kubenswrapper[4754]: I0126 17:11:08.157860 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=5.157839084 podStartE2EDuration="5.157839084s" podCreationTimestamp="2026-01-26 17:11:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 17:11:08.151441537 +0000 UTC m=+1434.675622001" watchObservedRunningTime="2026-01-26 17:11:08.157839084 +0000 UTC m=+1434.682019518" Jan 26 17:11:08 crc kubenswrapper[4754]: I0126 17:11:08.169527 4754 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/kube-state-metrics-0" podUID="c3c55551-6de7-403e-bbf5-46a2ec685c13" containerName="kube-state-metrics" probeResult="failure" output="Get \"http://10.217.0.103:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Jan 26 17:11:08 crc kubenswrapper[4754]: I0126 17:11:08.445811 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Jan 26 17:11:08 crc kubenswrapper[4754]: I0126 17:11:08.446056 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Jan 26 17:11:08 crc kubenswrapper[4754]: I0126 17:11:08.531770 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Jan 26 17:11:08 crc kubenswrapper[4754]: I0126 17:11:08.653075 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9b50d057-68a1-44dc-8204-0366161e3cf4-config-data\") pod \"9b50d057-68a1-44dc-8204-0366161e3cf4\" (UID: \"9b50d057-68a1-44dc-8204-0366161e3cf4\") " Jan 26 17:11:08 crc kubenswrapper[4754]: I0126 17:11:08.653138 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gfxp5\" (UniqueName: \"kubernetes.io/projected/9b50d057-68a1-44dc-8204-0366161e3cf4-kube-api-access-gfxp5\") pod \"9b50d057-68a1-44dc-8204-0366161e3cf4\" (UID: \"9b50d057-68a1-44dc-8204-0366161e3cf4\") " Jan 26 17:11:08 crc kubenswrapper[4754]: I0126 17:11:08.653207 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9b50d057-68a1-44dc-8204-0366161e3cf4-logs\") pod \"9b50d057-68a1-44dc-8204-0366161e3cf4\" (UID: \"9b50d057-68a1-44dc-8204-0366161e3cf4\") " Jan 26 17:11:08 crc kubenswrapper[4754]: I0126 17:11:08.653422 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9b50d057-68a1-44dc-8204-0366161e3cf4-combined-ca-bundle\") pod \"9b50d057-68a1-44dc-8204-0366161e3cf4\" (UID: \"9b50d057-68a1-44dc-8204-0366161e3cf4\") " Jan 26 17:11:08 crc kubenswrapper[4754]: I0126 17:11:08.654814 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9b50d057-68a1-44dc-8204-0366161e3cf4-logs" (OuterVolumeSpecName: "logs") pod "9b50d057-68a1-44dc-8204-0366161e3cf4" (UID: "9b50d057-68a1-44dc-8204-0366161e3cf4"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:11:08 crc kubenswrapper[4754]: I0126 17:11:08.659619 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9b50d057-68a1-44dc-8204-0366161e3cf4-kube-api-access-gfxp5" (OuterVolumeSpecName: "kube-api-access-gfxp5") pod "9b50d057-68a1-44dc-8204-0366161e3cf4" (UID: "9b50d057-68a1-44dc-8204-0366161e3cf4"). InnerVolumeSpecName "kube-api-access-gfxp5". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:11:08 crc kubenswrapper[4754]: I0126 17:11:08.691459 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9b50d057-68a1-44dc-8204-0366161e3cf4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9b50d057-68a1-44dc-8204-0366161e3cf4" (UID: "9b50d057-68a1-44dc-8204-0366161e3cf4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:11:08 crc kubenswrapper[4754]: I0126 17:11:08.693579 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9b50d057-68a1-44dc-8204-0366161e3cf4-config-data" (OuterVolumeSpecName: "config-data") pod "9b50d057-68a1-44dc-8204-0366161e3cf4" (UID: "9b50d057-68a1-44dc-8204-0366161e3cf4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:11:08 crc kubenswrapper[4754]: I0126 17:11:08.755932 4754 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9b50d057-68a1-44dc-8204-0366161e3cf4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 26 17:11:08 crc kubenswrapper[4754]: I0126 17:11:08.755962 4754 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9b50d057-68a1-44dc-8204-0366161e3cf4-config-data\") on node \"crc\" DevicePath \"\"" Jan 26 17:11:08 crc kubenswrapper[4754]: I0126 17:11:08.755971 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gfxp5\" (UniqueName: \"kubernetes.io/projected/9b50d057-68a1-44dc-8204-0366161e3cf4-kube-api-access-gfxp5\") on node \"crc\" DevicePath \"\"" Jan 26 17:11:08 crc kubenswrapper[4754]: I0126 17:11:08.755982 4754 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9b50d057-68a1-44dc-8204-0366161e3cf4-logs\") on node \"crc\" DevicePath \"\"" Jan 26 17:11:08 crc kubenswrapper[4754]: I0126 17:11:08.865632 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 26 17:11:08 crc kubenswrapper[4754]: I0126 17:11:08.888770 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Jan 26 17:11:08 crc kubenswrapper[4754]: I0126 17:11:08.898683 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Jan 26 17:11:08 crc kubenswrapper[4754]: I0126 17:11:08.958881 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/270ffa9f-f033-488c-887e-565ca34620c6-combined-ca-bundle\") pod \"270ffa9f-f033-488c-887e-565ca34620c6\" (UID: \"270ffa9f-f033-488c-887e-565ca34620c6\") " Jan 26 17:11:08 crc kubenswrapper[4754]: I0126 17:11:08.959091 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a5a89a1e-ce5c-4e31-9919-c43880fe6c04-config-data\") pod \"a5a89a1e-ce5c-4e31-9919-c43880fe6c04\" (UID: \"a5a89a1e-ce5c-4e31-9919-c43880fe6c04\") " Jan 26 17:11:08 crc kubenswrapper[4754]: I0126 17:11:08.959195 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/6d0563ff-f0f4-45b7-aaba-014f4426f4b0-sg-core-conf-yaml\") pod \"6d0563ff-f0f4-45b7-aaba-014f4426f4b0\" (UID: \"6d0563ff-f0f4-45b7-aaba-014f4426f4b0\") " Jan 26 17:11:08 crc kubenswrapper[4754]: I0126 17:11:08.959247 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a5a89a1e-ce5c-4e31-9919-c43880fe6c04-logs\") pod \"a5a89a1e-ce5c-4e31-9919-c43880fe6c04\" (UID: \"a5a89a1e-ce5c-4e31-9919-c43880fe6c04\") " Jan 26 17:11:08 crc kubenswrapper[4754]: I0126 17:11:08.959281 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6d0563ff-f0f4-45b7-aaba-014f4426f4b0-config-data\") pod \"6d0563ff-f0f4-45b7-aaba-014f4426f4b0\" (UID: \"6d0563ff-f0f4-45b7-aaba-014f4426f4b0\") " Jan 26 17:11:08 crc kubenswrapper[4754]: I0126 17:11:08.959318 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d0563ff-f0f4-45b7-aaba-014f4426f4b0-combined-ca-bundle\") pod \"6d0563ff-f0f4-45b7-aaba-014f4426f4b0\" (UID: \"6d0563ff-f0f4-45b7-aaba-014f4426f4b0\") " Jan 26 17:11:08 crc kubenswrapper[4754]: I0126 17:11:08.959343 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5g8db\" (UniqueName: \"kubernetes.io/projected/6d0563ff-f0f4-45b7-aaba-014f4426f4b0-kube-api-access-5g8db\") pod \"6d0563ff-f0f4-45b7-aaba-014f4426f4b0\" (UID: \"6d0563ff-f0f4-45b7-aaba-014f4426f4b0\") " Jan 26 17:11:08 crc kubenswrapper[4754]: I0126 17:11:08.959380 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bgxwk\" (UniqueName: \"kubernetes.io/projected/a5a89a1e-ce5c-4e31-9919-c43880fe6c04-kube-api-access-bgxwk\") pod \"a5a89a1e-ce5c-4e31-9919-c43880fe6c04\" (UID: \"a5a89a1e-ce5c-4e31-9919-c43880fe6c04\") " Jan 26 17:11:08 crc kubenswrapper[4754]: I0126 17:11:08.959417 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6d0563ff-f0f4-45b7-aaba-014f4426f4b0-run-httpd\") pod \"6d0563ff-f0f4-45b7-aaba-014f4426f4b0\" (UID: \"6d0563ff-f0f4-45b7-aaba-014f4426f4b0\") " Jan 26 17:11:08 crc kubenswrapper[4754]: I0126 17:11:08.959457 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qh4h8\" (UniqueName: \"kubernetes.io/projected/270ffa9f-f033-488c-887e-565ca34620c6-kube-api-access-qh4h8\") pod \"270ffa9f-f033-488c-887e-565ca34620c6\" (UID: \"270ffa9f-f033-488c-887e-565ca34620c6\") " Jan 26 17:11:08 crc kubenswrapper[4754]: I0126 17:11:08.959494 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6d0563ff-f0f4-45b7-aaba-014f4426f4b0-scripts\") pod \"6d0563ff-f0f4-45b7-aaba-014f4426f4b0\" (UID: \"6d0563ff-f0f4-45b7-aaba-014f4426f4b0\") " Jan 26 17:11:08 crc kubenswrapper[4754]: I0126 17:11:08.959523 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6d0563ff-f0f4-45b7-aaba-014f4426f4b0-log-httpd\") pod \"6d0563ff-f0f4-45b7-aaba-014f4426f4b0\" (UID: \"6d0563ff-f0f4-45b7-aaba-014f4426f4b0\") " Jan 26 17:11:08 crc kubenswrapper[4754]: I0126 17:11:08.959568 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/270ffa9f-f033-488c-887e-565ca34620c6-config-data\") pod \"270ffa9f-f033-488c-887e-565ca34620c6\" (UID: \"270ffa9f-f033-488c-887e-565ca34620c6\") " Jan 26 17:11:08 crc kubenswrapper[4754]: I0126 17:11:08.959583 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a5a89a1e-ce5c-4e31-9919-c43880fe6c04-logs" (OuterVolumeSpecName: "logs") pod "a5a89a1e-ce5c-4e31-9919-c43880fe6c04" (UID: "a5a89a1e-ce5c-4e31-9919-c43880fe6c04"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:11:08 crc kubenswrapper[4754]: I0126 17:11:08.959596 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5a89a1e-ce5c-4e31-9919-c43880fe6c04-combined-ca-bundle\") pod \"a5a89a1e-ce5c-4e31-9919-c43880fe6c04\" (UID: \"a5a89a1e-ce5c-4e31-9919-c43880fe6c04\") " Jan 26 17:11:08 crc kubenswrapper[4754]: I0126 17:11:08.959616 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/a5a89a1e-ce5c-4e31-9919-c43880fe6c04-nova-metadata-tls-certs\") pod \"a5a89a1e-ce5c-4e31-9919-c43880fe6c04\" (UID: \"a5a89a1e-ce5c-4e31-9919-c43880fe6c04\") " Jan 26 17:11:08 crc kubenswrapper[4754]: I0126 17:11:08.959996 4754 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a5a89a1e-ce5c-4e31-9919-c43880fe6c04-logs\") on node \"crc\" DevicePath \"\"" Jan 26 17:11:08 crc kubenswrapper[4754]: I0126 17:11:08.961457 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6d0563ff-f0f4-45b7-aaba-014f4426f4b0-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "6d0563ff-f0f4-45b7-aaba-014f4426f4b0" (UID: "6d0563ff-f0f4-45b7-aaba-014f4426f4b0"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:11:08 crc kubenswrapper[4754]: I0126 17:11:08.965862 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a5a89a1e-ce5c-4e31-9919-c43880fe6c04-kube-api-access-bgxwk" (OuterVolumeSpecName: "kube-api-access-bgxwk") pod "a5a89a1e-ce5c-4e31-9919-c43880fe6c04" (UID: "a5a89a1e-ce5c-4e31-9919-c43880fe6c04"). InnerVolumeSpecName "kube-api-access-bgxwk". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:11:08 crc kubenswrapper[4754]: I0126 17:11:08.966165 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6d0563ff-f0f4-45b7-aaba-014f4426f4b0-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "6d0563ff-f0f4-45b7-aaba-014f4426f4b0" (UID: "6d0563ff-f0f4-45b7-aaba-014f4426f4b0"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:11:08 crc kubenswrapper[4754]: I0126 17:11:08.970103 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/270ffa9f-f033-488c-887e-565ca34620c6-kube-api-access-qh4h8" (OuterVolumeSpecName: "kube-api-access-qh4h8") pod "270ffa9f-f033-488c-887e-565ca34620c6" (UID: "270ffa9f-f033-488c-887e-565ca34620c6"). InnerVolumeSpecName "kube-api-access-qh4h8". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:11:08 crc kubenswrapper[4754]: I0126 17:11:08.975918 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6d0563ff-f0f4-45b7-aaba-014f4426f4b0-scripts" (OuterVolumeSpecName: "scripts") pod "6d0563ff-f0f4-45b7-aaba-014f4426f4b0" (UID: "6d0563ff-f0f4-45b7-aaba-014f4426f4b0"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:11:08 crc kubenswrapper[4754]: I0126 17:11:08.975932 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6d0563ff-f0f4-45b7-aaba-014f4426f4b0-kube-api-access-5g8db" (OuterVolumeSpecName: "kube-api-access-5g8db") pod "6d0563ff-f0f4-45b7-aaba-014f4426f4b0" (UID: "6d0563ff-f0f4-45b7-aaba-014f4426f4b0"). InnerVolumeSpecName "kube-api-access-5g8db". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:11:08 crc kubenswrapper[4754]: I0126 17:11:08.997543 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6d0563ff-f0f4-45b7-aaba-014f4426f4b0-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "6d0563ff-f0f4-45b7-aaba-014f4426f4b0" (UID: "6d0563ff-f0f4-45b7-aaba-014f4426f4b0"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.014320 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a5a89a1e-ce5c-4e31-9919-c43880fe6c04-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a5a89a1e-ce5c-4e31-9919-c43880fe6c04" (UID: "a5a89a1e-ce5c-4e31-9919-c43880fe6c04"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.014579 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/270ffa9f-f033-488c-887e-565ca34620c6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "270ffa9f-f033-488c-887e-565ca34620c6" (UID: "270ffa9f-f033-488c-887e-565ca34620c6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.017330 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a5a89a1e-ce5c-4e31-9919-c43880fe6c04-config-data" (OuterVolumeSpecName: "config-data") pod "a5a89a1e-ce5c-4e31-9919-c43880fe6c04" (UID: "a5a89a1e-ce5c-4e31-9919-c43880fe6c04"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.029923 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/270ffa9f-f033-488c-887e-565ca34620c6-config-data" (OuterVolumeSpecName: "config-data") pod "270ffa9f-f033-488c-887e-565ca34620c6" (UID: "270ffa9f-f033-488c-887e-565ca34620c6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.034059 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a5a89a1e-ce5c-4e31-9919-c43880fe6c04-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "a5a89a1e-ce5c-4e31-9919-c43880fe6c04" (UID: "a5a89a1e-ce5c-4e31-9919-c43880fe6c04"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.061276 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6d0563ff-f0f4-45b7-aaba-014f4426f4b0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6d0563ff-f0f4-45b7-aaba-014f4426f4b0" (UID: "6d0563ff-f0f4-45b7-aaba-014f4426f4b0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.062757 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5g8db\" (UniqueName: \"kubernetes.io/projected/6d0563ff-f0f4-45b7-aaba-014f4426f4b0-kube-api-access-5g8db\") on node \"crc\" DevicePath \"\"" Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.062905 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bgxwk\" (UniqueName: \"kubernetes.io/projected/a5a89a1e-ce5c-4e31-9919-c43880fe6c04-kube-api-access-bgxwk\") on node \"crc\" DevicePath \"\"" Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.062999 4754 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6d0563ff-f0f4-45b7-aaba-014f4426f4b0-run-httpd\") on node \"crc\" DevicePath \"\"" Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.063132 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qh4h8\" (UniqueName: \"kubernetes.io/projected/270ffa9f-f033-488c-887e-565ca34620c6-kube-api-access-qh4h8\") on node \"crc\" DevicePath \"\"" Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.063241 4754 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6d0563ff-f0f4-45b7-aaba-014f4426f4b0-scripts\") on node \"crc\" DevicePath \"\"" Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.063310 4754 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6d0563ff-f0f4-45b7-aaba-014f4426f4b0-log-httpd\") on node \"crc\" DevicePath \"\"" Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.063387 4754 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/270ffa9f-f033-488c-887e-565ca34620c6-config-data\") on node \"crc\" DevicePath \"\"" Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.063458 4754 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5a89a1e-ce5c-4e31-9919-c43880fe6c04-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.063529 4754 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/a5a89a1e-ce5c-4e31-9919-c43880fe6c04-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.063608 4754 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/270ffa9f-f033-488c-887e-565ca34620c6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.063722 4754 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a5a89a1e-ce5c-4e31-9919-c43880fe6c04-config-data\") on node \"crc\" DevicePath \"\"" Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.063857 4754 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/6d0563ff-f0f4-45b7-aaba-014f4426f4b0-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.063983 4754 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d0563ff-f0f4-45b7-aaba-014f4426f4b0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.082818 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6d0563ff-f0f4-45b7-aaba-014f4426f4b0-config-data" (OuterVolumeSpecName: "config-data") pod "6d0563ff-f0f4-45b7-aaba-014f4426f4b0" (UID: "6d0563ff-f0f4-45b7-aaba-014f4426f4b0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.138691 4754 generic.go:334] "Generic (PLEG): container finished" podID="a5a89a1e-ce5c-4e31-9919-c43880fe6c04" containerID="0b335ff1e7d7ca9a9013152a903eb9bc32541c7f911804f273535fefecd8f2fc" exitCode=0 Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.138727 4754 generic.go:334] "Generic (PLEG): container finished" podID="a5a89a1e-ce5c-4e31-9919-c43880fe6c04" containerID="3bda1766505c4134ecf969f9f1a8caf229decec57840fbbcfd89a88471c78aa2" exitCode=143 Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.138771 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a5a89a1e-ce5c-4e31-9919-c43880fe6c04","Type":"ContainerDied","Data":"0b335ff1e7d7ca9a9013152a903eb9bc32541c7f911804f273535fefecd8f2fc"} Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.138801 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a5a89a1e-ce5c-4e31-9919-c43880fe6c04","Type":"ContainerDied","Data":"3bda1766505c4134ecf969f9f1a8caf229decec57840fbbcfd89a88471c78aa2"} Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.138815 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a5a89a1e-ce5c-4e31-9919-c43880fe6c04","Type":"ContainerDied","Data":"652d72177417a349bcc1adadeda2b20ab4edd0e2ad27917d3be625238cd214bf"} Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.138833 4754 scope.go:117] "RemoveContainer" containerID="0b335ff1e7d7ca9a9013152a903eb9bc32541c7f911804f273535fefecd8f2fc" Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.138941 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.151007 4754 generic.go:334] "Generic (PLEG): container finished" podID="6d0563ff-f0f4-45b7-aaba-014f4426f4b0" containerID="234f898b8e364e60561ed34cd7475fe072e0bd9490b1145b1909804af99b448d" exitCode=0 Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.151089 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6d0563ff-f0f4-45b7-aaba-014f4426f4b0","Type":"ContainerDied","Data":"234f898b8e364e60561ed34cd7475fe072e0bd9490b1145b1909804af99b448d"} Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.151121 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6d0563ff-f0f4-45b7-aaba-014f4426f4b0","Type":"ContainerDied","Data":"01491f02574e1817229e56b88ccabe69e44826d913624b64f6546ef1af57aee7"} Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.151194 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.157561 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"9b50d057-68a1-44dc-8204-0366161e3cf4","Type":"ContainerDied","Data":"a103081402c951965ad316a8311c8605e190face4d63fd1550a224a59959e289"} Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.158089 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.168911 4754 scope.go:117] "RemoveContainer" containerID="3bda1766505c4134ecf969f9f1a8caf229decec57840fbbcfd89a88471c78aa2" Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.169626 4754 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6d0563ff-f0f4-45b7-aaba-014f4426f4b0-config-data\") on node \"crc\" DevicePath \"\"" Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.173767 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hxdzc" event={"ID":"e3881c45-db75-4ca1-bb8f-92f413446265","Type":"ContainerStarted","Data":"9e2c96a294df3c2e892df642a032ac13972ed913ae08a5cc7874baec5dd4fc6b"} Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.176906 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"6be68290-e783-4f39-8257-d5dc4051447a","Type":"ContainerStarted","Data":"0fcab80f376b5d00ae89d5db8aeaadf44c1928109a5a0e72eda564d3a79df8b7"} Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.177034 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.187038 4754 generic.go:334] "Generic (PLEG): container finished" podID="270ffa9f-f033-488c-887e-565ca34620c6" containerID="dbafe75f4149545f66d7dcf67e5679e62b6cad3503ae5ede59a6619920905833" exitCode=0 Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.187092 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"270ffa9f-f033-488c-887e-565ca34620c6","Type":"ContainerDied","Data":"dbafe75f4149545f66d7dcf67e5679e62b6cad3503ae5ede59a6619920905833"} Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.187123 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"270ffa9f-f033-488c-887e-565ca34620c6","Type":"ContainerDied","Data":"eb0592623072b15408efdbdfd5823d972d980b2bebd8e10f94507279c567ec4d"} Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.187191 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.206252 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.208532 4754 scope.go:117] "RemoveContainer" containerID="0b335ff1e7d7ca9a9013152a903eb9bc32541c7f911804f273535fefecd8f2fc" Jan 26 17:11:09 crc kubenswrapper[4754]: E0126 17:11:09.209165 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0b335ff1e7d7ca9a9013152a903eb9bc32541c7f911804f273535fefecd8f2fc\": container with ID starting with 0b335ff1e7d7ca9a9013152a903eb9bc32541c7f911804f273535fefecd8f2fc not found: ID does not exist" containerID="0b335ff1e7d7ca9a9013152a903eb9bc32541c7f911804f273535fefecd8f2fc" Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.209199 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0b335ff1e7d7ca9a9013152a903eb9bc32541c7f911804f273535fefecd8f2fc"} err="failed to get container status \"0b335ff1e7d7ca9a9013152a903eb9bc32541c7f911804f273535fefecd8f2fc\": rpc error: code = NotFound desc = could not find container \"0b335ff1e7d7ca9a9013152a903eb9bc32541c7f911804f273535fefecd8f2fc\": container with ID starting with 0b335ff1e7d7ca9a9013152a903eb9bc32541c7f911804f273535fefecd8f2fc not found: ID does not exist" Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.209226 4754 scope.go:117] "RemoveContainer" containerID="3bda1766505c4134ecf969f9f1a8caf229decec57840fbbcfd89a88471c78aa2" Jan 26 17:11:09 crc kubenswrapper[4754]: E0126 17:11:09.210832 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3bda1766505c4134ecf969f9f1a8caf229decec57840fbbcfd89a88471c78aa2\": container with ID starting with 3bda1766505c4134ecf969f9f1a8caf229decec57840fbbcfd89a88471c78aa2 not found: ID does not exist" containerID="3bda1766505c4134ecf969f9f1a8caf229decec57840fbbcfd89a88471c78aa2" Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.210863 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3bda1766505c4134ecf969f9f1a8caf229decec57840fbbcfd89a88471c78aa2"} err="failed to get container status \"3bda1766505c4134ecf969f9f1a8caf229decec57840fbbcfd89a88471c78aa2\": rpc error: code = NotFound desc = could not find container \"3bda1766505c4134ecf969f9f1a8caf229decec57840fbbcfd89a88471c78aa2\": container with ID starting with 3bda1766505c4134ecf969f9f1a8caf229decec57840fbbcfd89a88471c78aa2 not found: ID does not exist" Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.210885 4754 scope.go:117] "RemoveContainer" containerID="0b335ff1e7d7ca9a9013152a903eb9bc32541c7f911804f273535fefecd8f2fc" Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.213937 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0b335ff1e7d7ca9a9013152a903eb9bc32541c7f911804f273535fefecd8f2fc"} err="failed to get container status \"0b335ff1e7d7ca9a9013152a903eb9bc32541c7f911804f273535fefecd8f2fc\": rpc error: code = NotFound desc = could not find container \"0b335ff1e7d7ca9a9013152a903eb9bc32541c7f911804f273535fefecd8f2fc\": container with ID starting with 0b335ff1e7d7ca9a9013152a903eb9bc32541c7f911804f273535fefecd8f2fc not found: ID does not exist" Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.213963 4754 scope.go:117] "RemoveContainer" containerID="3bda1766505c4134ecf969f9f1a8caf229decec57840fbbcfd89a88471c78aa2" Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.215505 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3bda1766505c4134ecf969f9f1a8caf229decec57840fbbcfd89a88471c78aa2"} err="failed to get container status \"3bda1766505c4134ecf969f9f1a8caf229decec57840fbbcfd89a88471c78aa2\": rpc error: code = NotFound desc = could not find container \"3bda1766505c4134ecf969f9f1a8caf229decec57840fbbcfd89a88471c78aa2\": container with ID starting with 3bda1766505c4134ecf969f9f1a8caf229decec57840fbbcfd89a88471c78aa2 not found: ID does not exist" Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.215528 4754 scope.go:117] "RemoveContainer" containerID="ebd4a3041094da9f63be25f198192702e8ece75b22e3c5c564ca86dfb8a0148c" Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.222740 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.236812 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Jan 26 17:11:09 crc kubenswrapper[4754]: E0126 17:11:09.237281 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d0563ff-f0f4-45b7-aaba-014f4426f4b0" containerName="ceilometer-central-agent" Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.237301 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d0563ff-f0f4-45b7-aaba-014f4426f4b0" containerName="ceilometer-central-agent" Jan 26 17:11:09 crc kubenswrapper[4754]: E0126 17:11:09.237315 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d0563ff-f0f4-45b7-aaba-014f4426f4b0" containerName="sg-core" Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.237323 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d0563ff-f0f4-45b7-aaba-014f4426f4b0" containerName="sg-core" Jan 26 17:11:09 crc kubenswrapper[4754]: E0126 17:11:09.237334 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b50d057-68a1-44dc-8204-0366161e3cf4" containerName="nova-api-log" Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.237342 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b50d057-68a1-44dc-8204-0366161e3cf4" containerName="nova-api-log" Jan 26 17:11:09 crc kubenswrapper[4754]: E0126 17:11:09.237356 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5a89a1e-ce5c-4e31-9919-c43880fe6c04" containerName="nova-metadata-log" Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.237363 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5a89a1e-ce5c-4e31-9919-c43880fe6c04" containerName="nova-metadata-log" Jan 26 17:11:09 crc kubenswrapper[4754]: E0126 17:11:09.237393 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d0563ff-f0f4-45b7-aaba-014f4426f4b0" containerName="proxy-httpd" Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.237400 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d0563ff-f0f4-45b7-aaba-014f4426f4b0" containerName="proxy-httpd" Jan 26 17:11:09 crc kubenswrapper[4754]: E0126 17:11:09.237413 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b50d057-68a1-44dc-8204-0366161e3cf4" containerName="nova-api-api" Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.237420 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b50d057-68a1-44dc-8204-0366161e3cf4" containerName="nova-api-api" Jan 26 17:11:09 crc kubenswrapper[4754]: E0126 17:11:09.237430 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5a89a1e-ce5c-4e31-9919-c43880fe6c04" containerName="nova-metadata-metadata" Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.237438 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5a89a1e-ce5c-4e31-9919-c43880fe6c04" containerName="nova-metadata-metadata" Jan 26 17:11:09 crc kubenswrapper[4754]: E0126 17:11:09.237453 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="270ffa9f-f033-488c-887e-565ca34620c6" containerName="nova-scheduler-scheduler" Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.237460 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="270ffa9f-f033-488c-887e-565ca34620c6" containerName="nova-scheduler-scheduler" Jan 26 17:11:09 crc kubenswrapper[4754]: E0126 17:11:09.237476 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d0563ff-f0f4-45b7-aaba-014f4426f4b0" containerName="ceilometer-notification-agent" Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.237485 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d0563ff-f0f4-45b7-aaba-014f4426f4b0" containerName="ceilometer-notification-agent" Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.237686 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="9b50d057-68a1-44dc-8204-0366161e3cf4" containerName="nova-api-api" Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.237718 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="6d0563ff-f0f4-45b7-aaba-014f4426f4b0" containerName="ceilometer-central-agent" Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.237728 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="270ffa9f-f033-488c-887e-565ca34620c6" containerName="nova-scheduler-scheduler" Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.237751 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="6d0563ff-f0f4-45b7-aaba-014f4426f4b0" containerName="sg-core" Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.237762 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="a5a89a1e-ce5c-4e31-9919-c43880fe6c04" containerName="nova-metadata-log" Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.237771 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="6d0563ff-f0f4-45b7-aaba-014f4426f4b0" containerName="ceilometer-notification-agent" Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.237782 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="6d0563ff-f0f4-45b7-aaba-014f4426f4b0" containerName="proxy-httpd" Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.237797 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="a5a89a1e-ce5c-4e31-9919-c43880fe6c04" containerName="nova-metadata-metadata" Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.237810 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="9b50d057-68a1-44dc-8204-0366161e3cf4" containerName="nova-api-log" Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.239173 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.246799 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-hxdzc" podStartSLOduration=4.702374506 podStartE2EDuration="10.246777734s" podCreationTimestamp="2026-01-26 17:10:59 +0000 UTC" firstStartedPulling="2026-01-26 17:11:02.994817417 +0000 UTC m=+1429.518997861" lastFinishedPulling="2026-01-26 17:11:08.539220655 +0000 UTC m=+1435.063401089" observedRunningTime="2026-01-26 17:11:09.215908643 +0000 UTC m=+1435.740089087" watchObservedRunningTime="2026-01-26 17:11:09.246777734 +0000 UTC m=+1435.770958168" Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.248907 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.250440 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.267154 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.267298 4754 scope.go:117] "RemoveContainer" containerID="2450afc5412f20e482cf129b8e70df60fec15191d846f728199beaa14ae587fe" Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.280719 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.293381 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.302171 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=2.391624019 podStartE2EDuration="5.302132499s" podCreationTimestamp="2026-01-26 17:11:04 +0000 UTC" firstStartedPulling="2026-01-26 17:11:05.016937764 +0000 UTC m=+1431.541118208" lastFinishedPulling="2026-01-26 17:11:07.927446264 +0000 UTC m=+1434.451626688" observedRunningTime="2026-01-26 17:11:09.2702151 +0000 UTC m=+1435.794395554" watchObservedRunningTime="2026-01-26 17:11:09.302132499 +0000 UTC m=+1435.826312933" Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.325996 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.332983 4754 scope.go:117] "RemoveContainer" containerID="234f898b8e364e60561ed34cd7475fe072e0bd9490b1145b1909804af99b448d" Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.364237 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.380810 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/37f5b6db-8ef3-4e4f-9763-b150188c1466-logs\") pod \"nova-metadata-0\" (UID: \"37f5b6db-8ef3-4e4f-9763-b150188c1466\") " pod="openstack/nova-metadata-0" Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.380907 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hf9q7\" (UniqueName: \"kubernetes.io/projected/37f5b6db-8ef3-4e4f-9763-b150188c1466-kube-api-access-hf9q7\") pod \"nova-metadata-0\" (UID: \"37f5b6db-8ef3-4e4f-9763-b150188c1466\") " pod="openstack/nova-metadata-0" Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.380952 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37f5b6db-8ef3-4e4f-9763-b150188c1466-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"37f5b6db-8ef3-4e4f-9763-b150188c1466\") " pod="openstack/nova-metadata-0" Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.381065 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37f5b6db-8ef3-4e4f-9763-b150188c1466-config-data\") pod \"nova-metadata-0\" (UID: \"37f5b6db-8ef3-4e4f-9763-b150188c1466\") " pod="openstack/nova-metadata-0" Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.381117 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/37f5b6db-8ef3-4e4f-9763-b150188c1466-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"37f5b6db-8ef3-4e4f-9763-b150188c1466\") " pod="openstack/nova-metadata-0" Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.402031 4754 scope.go:117] "RemoveContainer" containerID="eb80b00f37f12db37a6f667c832c82d481c8513f0c10b367ce2bb2cb977c988b" Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.411829 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.419304 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.447406 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.447488 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.484764 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.486695 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/37f5b6db-8ef3-4e4f-9763-b150188c1466-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"37f5b6db-8ef3-4e4f-9763-b150188c1466\") " pod="openstack/nova-metadata-0" Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.486870 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/37f5b6db-8ef3-4e4f-9763-b150188c1466-logs\") pod \"nova-metadata-0\" (UID: \"37f5b6db-8ef3-4e4f-9763-b150188c1466\") " pod="openstack/nova-metadata-0" Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.486974 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hf9q7\" (UniqueName: \"kubernetes.io/projected/37f5b6db-8ef3-4e4f-9763-b150188c1466-kube-api-access-hf9q7\") pod \"nova-metadata-0\" (UID: \"37f5b6db-8ef3-4e4f-9763-b150188c1466\") " pod="openstack/nova-metadata-0" Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.487048 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37f5b6db-8ef3-4e4f-9763-b150188c1466-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"37f5b6db-8ef3-4e4f-9763-b150188c1466\") " pod="openstack/nova-metadata-0" Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.487121 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d2701786-b1e9-4a86-b751-63e1dd98eeba-config-data\") pod \"nova-api-0\" (UID: \"d2701786-b1e9-4a86-b751-63e1dd98eeba\") " pod="openstack/nova-api-0" Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.487188 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h547r\" (UniqueName: \"kubernetes.io/projected/d2701786-b1e9-4a86-b751-63e1dd98eeba-kube-api-access-h547r\") pod \"nova-api-0\" (UID: \"d2701786-b1e9-4a86-b751-63e1dd98eeba\") " pod="openstack/nova-api-0" Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.487222 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d2701786-b1e9-4a86-b751-63e1dd98eeba-logs\") pod \"nova-api-0\" (UID: \"d2701786-b1e9-4a86-b751-63e1dd98eeba\") " pod="openstack/nova-api-0" Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.487273 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2701786-b1e9-4a86-b751-63e1dd98eeba-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"d2701786-b1e9-4a86-b751-63e1dd98eeba\") " pod="openstack/nova-api-0" Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.487309 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37f5b6db-8ef3-4e4f-9763-b150188c1466-config-data\") pod \"nova-metadata-0\" (UID: \"37f5b6db-8ef3-4e4f-9763-b150188c1466\") " pod="openstack/nova-metadata-0" Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.496071 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.497163 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/37f5b6db-8ef3-4e4f-9763-b150188c1466-logs\") pod \"nova-metadata-0\" (UID: \"37f5b6db-8ef3-4e4f-9763-b150188c1466\") " pod="openstack/nova-metadata-0" Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.500725 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37f5b6db-8ef3-4e4f-9763-b150188c1466-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"37f5b6db-8ef3-4e4f-9763-b150188c1466\") " pod="openstack/nova-metadata-0" Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.509393 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37f5b6db-8ef3-4e4f-9763-b150188c1466-config-data\") pod \"nova-metadata-0\" (UID: \"37f5b6db-8ef3-4e4f-9763-b150188c1466\") " pod="openstack/nova-metadata-0" Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.509594 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/37f5b6db-8ef3-4e4f-9763-b150188c1466-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"37f5b6db-8ef3-4e4f-9763-b150188c1466\") " pod="openstack/nova-metadata-0" Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.511390 4754 scope.go:117] "RemoveContainer" containerID="ebd4a3041094da9f63be25f198192702e8ece75b22e3c5c564ca86dfb8a0148c" Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.515898 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hf9q7\" (UniqueName: \"kubernetes.io/projected/37f5b6db-8ef3-4e4f-9763-b150188c1466-kube-api-access-hf9q7\") pod \"nova-metadata-0\" (UID: \"37f5b6db-8ef3-4e4f-9763-b150188c1466\") " pod="openstack/nova-metadata-0" Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.515970 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.519954 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 26 17:11:09 crc kubenswrapper[4754]: E0126 17:11:09.519971 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ebd4a3041094da9f63be25f198192702e8ece75b22e3c5c564ca86dfb8a0148c\": container with ID starting with ebd4a3041094da9f63be25f198192702e8ece75b22e3c5c564ca86dfb8a0148c not found: ID does not exist" containerID="ebd4a3041094da9f63be25f198192702e8ece75b22e3c5c564ca86dfb8a0148c" Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.520026 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ebd4a3041094da9f63be25f198192702e8ece75b22e3c5c564ca86dfb8a0148c"} err="failed to get container status \"ebd4a3041094da9f63be25f198192702e8ece75b22e3c5c564ca86dfb8a0148c\": rpc error: code = NotFound desc = could not find container \"ebd4a3041094da9f63be25f198192702e8ece75b22e3c5c564ca86dfb8a0148c\": container with ID starting with ebd4a3041094da9f63be25f198192702e8ece75b22e3c5c564ca86dfb8a0148c not found: ID does not exist" Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.520056 4754 scope.go:117] "RemoveContainer" containerID="2450afc5412f20e482cf129b8e70df60fec15191d846f728199beaa14ae587fe" Jan 26 17:11:09 crc kubenswrapper[4754]: E0126 17:11:09.520954 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2450afc5412f20e482cf129b8e70df60fec15191d846f728199beaa14ae587fe\": container with ID starting with 2450afc5412f20e482cf129b8e70df60fec15191d846f728199beaa14ae587fe not found: ID does not exist" containerID="2450afc5412f20e482cf129b8e70df60fec15191d846f728199beaa14ae587fe" Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.520992 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2450afc5412f20e482cf129b8e70df60fec15191d846f728199beaa14ae587fe"} err="failed to get container status \"2450afc5412f20e482cf129b8e70df60fec15191d846f728199beaa14ae587fe\": rpc error: code = NotFound desc = could not find container \"2450afc5412f20e482cf129b8e70df60fec15191d846f728199beaa14ae587fe\": container with ID starting with 2450afc5412f20e482cf129b8e70df60fec15191d846f728199beaa14ae587fe not found: ID does not exist" Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.521022 4754 scope.go:117] "RemoveContainer" containerID="234f898b8e364e60561ed34cd7475fe072e0bd9490b1145b1909804af99b448d" Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.522981 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.526951 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.527183 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.533107 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.534449 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.536880 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Jan 26 17:11:09 crc kubenswrapper[4754]: E0126 17:11:09.544833 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"234f898b8e364e60561ed34cd7475fe072e0bd9490b1145b1909804af99b448d\": container with ID starting with 234f898b8e364e60561ed34cd7475fe072e0bd9490b1145b1909804af99b448d not found: ID does not exist" containerID="234f898b8e364e60561ed34cd7475fe072e0bd9490b1145b1909804af99b448d" Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.544873 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"234f898b8e364e60561ed34cd7475fe072e0bd9490b1145b1909804af99b448d"} err="failed to get container status \"234f898b8e364e60561ed34cd7475fe072e0bd9490b1145b1909804af99b448d\": rpc error: code = NotFound desc = could not find container \"234f898b8e364e60561ed34cd7475fe072e0bd9490b1145b1909804af99b448d\": container with ID starting with 234f898b8e364e60561ed34cd7475fe072e0bd9490b1145b1909804af99b448d not found: ID does not exist" Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.544895 4754 scope.go:117] "RemoveContainer" containerID="eb80b00f37f12db37a6f667c832c82d481c8513f0c10b367ce2bb2cb977c988b" Jan 26 17:11:09 crc kubenswrapper[4754]: E0126 17:11:09.545474 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eb80b00f37f12db37a6f667c832c82d481c8513f0c10b367ce2bb2cb977c988b\": container with ID starting with eb80b00f37f12db37a6f667c832c82d481c8513f0c10b367ce2bb2cb977c988b not found: ID does not exist" containerID="eb80b00f37f12db37a6f667c832c82d481c8513f0c10b367ce2bb2cb977c988b" Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.545533 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eb80b00f37f12db37a6f667c832c82d481c8513f0c10b367ce2bb2cb977c988b"} err="failed to get container status \"eb80b00f37f12db37a6f667c832c82d481c8513f0c10b367ce2bb2cb977c988b\": rpc error: code = NotFound desc = could not find container \"eb80b00f37f12db37a6f667c832c82d481c8513f0c10b367ce2bb2cb977c988b\": container with ID starting with eb80b00f37f12db37a6f667c832c82d481c8513f0c10b367ce2bb2cb977c988b not found: ID does not exist" Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.545569 4754 scope.go:117] "RemoveContainer" containerID="4c0f6a01ffaef43ea38f623749ef8fb2a10ea1b37845de8eae86906116547dc6" Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.572367 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.573688 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.588951 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h547r\" (UniqueName: \"kubernetes.io/projected/d2701786-b1e9-4a86-b751-63e1dd98eeba-kube-api-access-h547r\") pod \"nova-api-0\" (UID: \"d2701786-b1e9-4a86-b751-63e1dd98eeba\") " pod="openstack/nova-api-0" Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.588994 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d2701786-b1e9-4a86-b751-63e1dd98eeba-logs\") pod \"nova-api-0\" (UID: \"d2701786-b1e9-4a86-b751-63e1dd98eeba\") " pod="openstack/nova-api-0" Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.589016 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2701786-b1e9-4a86-b751-63e1dd98eeba-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"d2701786-b1e9-4a86-b751-63e1dd98eeba\") " pod="openstack/nova-api-0" Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.589048 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2cff5231-483a-4d14-abf2-5ae85efc4edf-run-httpd\") pod \"ceilometer-0\" (UID: \"2cff5231-483a-4d14-abf2-5ae85efc4edf\") " pod="openstack/ceilometer-0" Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.589087 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/2cff5231-483a-4d14-abf2-5ae85efc4edf-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"2cff5231-483a-4d14-abf2-5ae85efc4edf\") " pod="openstack/ceilometer-0" Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.589106 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2cff5231-483a-4d14-abf2-5ae85efc4edf-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2cff5231-483a-4d14-abf2-5ae85efc4edf\") " pod="openstack/ceilometer-0" Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.589142 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2cff5231-483a-4d14-abf2-5ae85efc4edf-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2cff5231-483a-4d14-abf2-5ae85efc4edf\") " pod="openstack/ceilometer-0" Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.589156 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2cff5231-483a-4d14-abf2-5ae85efc4edf-config-data\") pod \"ceilometer-0\" (UID: \"2cff5231-483a-4d14-abf2-5ae85efc4edf\") " pod="openstack/ceilometer-0" Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.589195 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-frlx6\" (UniqueName: \"kubernetes.io/projected/2cff5231-483a-4d14-abf2-5ae85efc4edf-kube-api-access-frlx6\") pod \"ceilometer-0\" (UID: \"2cff5231-483a-4d14-abf2-5ae85efc4edf\") " pod="openstack/ceilometer-0" Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.589230 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2cff5231-483a-4d14-abf2-5ae85efc4edf-scripts\") pod \"ceilometer-0\" (UID: \"2cff5231-483a-4d14-abf2-5ae85efc4edf\") " pod="openstack/ceilometer-0" Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.589248 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d2701786-b1e9-4a86-b751-63e1dd98eeba-config-data\") pod \"nova-api-0\" (UID: \"d2701786-b1e9-4a86-b751-63e1dd98eeba\") " pod="openstack/nova-api-0" Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.589266 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2cff5231-483a-4d14-abf2-5ae85efc4edf-log-httpd\") pod \"ceilometer-0\" (UID: \"2cff5231-483a-4d14-abf2-5ae85efc4edf\") " pod="openstack/ceilometer-0" Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.589989 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d2701786-b1e9-4a86-b751-63e1dd98eeba-logs\") pod \"nova-api-0\" (UID: \"d2701786-b1e9-4a86-b751-63e1dd98eeba\") " pod="openstack/nova-api-0" Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.592640 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.596457 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2701786-b1e9-4a86-b751-63e1dd98eeba-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"d2701786-b1e9-4a86-b751-63e1dd98eeba\") " pod="openstack/nova-api-0" Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.607332 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d2701786-b1e9-4a86-b751-63e1dd98eeba-config-data\") pod \"nova-api-0\" (UID: \"d2701786-b1e9-4a86-b751-63e1dd98eeba\") " pod="openstack/nova-api-0" Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.615228 4754 scope.go:117] "RemoveContainer" containerID="77aafa83659a5fb388cddee05c48d3c0ee10ca31c83aecb39a67800530e24940" Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.620566 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h547r\" (UniqueName: \"kubernetes.io/projected/d2701786-b1e9-4a86-b751-63e1dd98eeba-kube-api-access-h547r\") pod \"nova-api-0\" (UID: \"d2701786-b1e9-4a86-b751-63e1dd98eeba\") " pod="openstack/nova-api-0" Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.690775 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2cff5231-483a-4d14-abf2-5ae85efc4edf-log-httpd\") pod \"ceilometer-0\" (UID: \"2cff5231-483a-4d14-abf2-5ae85efc4edf\") " pod="openstack/ceilometer-0" Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.691096 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wj9p7\" (UniqueName: \"kubernetes.io/projected/ef3b4399-44ec-476a-a0d5-e727b76c624e-kube-api-access-wj9p7\") pod \"nova-scheduler-0\" (UID: \"ef3b4399-44ec-476a-a0d5-e727b76c624e\") " pod="openstack/nova-scheduler-0" Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.691288 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2cff5231-483a-4d14-abf2-5ae85efc4edf-run-httpd\") pod \"ceilometer-0\" (UID: \"2cff5231-483a-4d14-abf2-5ae85efc4edf\") " pod="openstack/ceilometer-0" Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.691373 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2cff5231-483a-4d14-abf2-5ae85efc4edf-log-httpd\") pod \"ceilometer-0\" (UID: \"2cff5231-483a-4d14-abf2-5ae85efc4edf\") " pod="openstack/ceilometer-0" Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.691543 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/2cff5231-483a-4d14-abf2-5ae85efc4edf-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"2cff5231-483a-4d14-abf2-5ae85efc4edf\") " pod="openstack/ceilometer-0" Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.691663 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2cff5231-483a-4d14-abf2-5ae85efc4edf-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2cff5231-483a-4d14-abf2-5ae85efc4edf\") " pod="openstack/ceilometer-0" Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.691787 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2cff5231-483a-4d14-abf2-5ae85efc4edf-run-httpd\") pod \"ceilometer-0\" (UID: \"2cff5231-483a-4d14-abf2-5ae85efc4edf\") " pod="openstack/ceilometer-0" Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.691936 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2cff5231-483a-4d14-abf2-5ae85efc4edf-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2cff5231-483a-4d14-abf2-5ae85efc4edf\") " pod="openstack/ceilometer-0" Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.692068 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2cff5231-483a-4d14-abf2-5ae85efc4edf-config-data\") pod \"ceilometer-0\" (UID: \"2cff5231-483a-4d14-abf2-5ae85efc4edf\") " pod="openstack/ceilometer-0" Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.692701 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-frlx6\" (UniqueName: \"kubernetes.io/projected/2cff5231-483a-4d14-abf2-5ae85efc4edf-kube-api-access-frlx6\") pod \"ceilometer-0\" (UID: \"2cff5231-483a-4d14-abf2-5ae85efc4edf\") " pod="openstack/ceilometer-0" Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.692882 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ef3b4399-44ec-476a-a0d5-e727b76c624e-config-data\") pod \"nova-scheduler-0\" (UID: \"ef3b4399-44ec-476a-a0d5-e727b76c624e\") " pod="openstack/nova-scheduler-0" Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.693017 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2cff5231-483a-4d14-abf2-5ae85efc4edf-scripts\") pod \"ceilometer-0\" (UID: \"2cff5231-483a-4d14-abf2-5ae85efc4edf\") " pod="openstack/ceilometer-0" Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.693137 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef3b4399-44ec-476a-a0d5-e727b76c624e-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"ef3b4399-44ec-476a-a0d5-e727b76c624e\") " pod="openstack/nova-scheduler-0" Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.695873 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/2cff5231-483a-4d14-abf2-5ae85efc4edf-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"2cff5231-483a-4d14-abf2-5ae85efc4edf\") " pod="openstack/ceilometer-0" Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.696647 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2cff5231-483a-4d14-abf2-5ae85efc4edf-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2cff5231-483a-4d14-abf2-5ae85efc4edf\") " pod="openstack/ceilometer-0" Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.701627 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2cff5231-483a-4d14-abf2-5ae85efc4edf-scripts\") pod \"ceilometer-0\" (UID: \"2cff5231-483a-4d14-abf2-5ae85efc4edf\") " pod="openstack/ceilometer-0" Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.702053 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2cff5231-483a-4d14-abf2-5ae85efc4edf-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2cff5231-483a-4d14-abf2-5ae85efc4edf\") " pod="openstack/ceilometer-0" Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.704437 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2cff5231-483a-4d14-abf2-5ae85efc4edf-config-data\") pod \"ceilometer-0\" (UID: \"2cff5231-483a-4d14-abf2-5ae85efc4edf\") " pod="openstack/ceilometer-0" Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.736592 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-frlx6\" (UniqueName: \"kubernetes.io/projected/2cff5231-483a-4d14-abf2-5ae85efc4edf-kube-api-access-frlx6\") pod \"ceilometer-0\" (UID: \"2cff5231-483a-4d14-abf2-5ae85efc4edf\") " pod="openstack/ceilometer-0" Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.772788 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.794807 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ef3b4399-44ec-476a-a0d5-e727b76c624e-config-data\") pod \"nova-scheduler-0\" (UID: \"ef3b4399-44ec-476a-a0d5-e727b76c624e\") " pod="openstack/nova-scheduler-0" Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.794870 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef3b4399-44ec-476a-a0d5-e727b76c624e-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"ef3b4399-44ec-476a-a0d5-e727b76c624e\") " pod="openstack/nova-scheduler-0" Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.794900 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wj9p7\" (UniqueName: \"kubernetes.io/projected/ef3b4399-44ec-476a-a0d5-e727b76c624e-kube-api-access-wj9p7\") pod \"nova-scheduler-0\" (UID: \"ef3b4399-44ec-476a-a0d5-e727b76c624e\") " pod="openstack/nova-scheduler-0" Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.801016 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ef3b4399-44ec-476a-a0d5-e727b76c624e-config-data\") pod \"nova-scheduler-0\" (UID: \"ef3b4399-44ec-476a-a0d5-e727b76c624e\") " pod="openstack/nova-scheduler-0" Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.802438 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef3b4399-44ec-476a-a0d5-e727b76c624e-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"ef3b4399-44ec-476a-a0d5-e727b76c624e\") " pod="openstack/nova-scheduler-0" Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.805524 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="270ffa9f-f033-488c-887e-565ca34620c6" path="/var/lib/kubelet/pods/270ffa9f-f033-488c-887e-565ca34620c6/volumes" Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.806435 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6d0563ff-f0f4-45b7-aaba-014f4426f4b0" path="/var/lib/kubelet/pods/6d0563ff-f0f4-45b7-aaba-014f4426f4b0/volumes" Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.807234 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9b50d057-68a1-44dc-8204-0366161e3cf4" path="/var/lib/kubelet/pods/9b50d057-68a1-44dc-8204-0366161e3cf4/volumes" Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.808446 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a5a89a1e-ce5c-4e31-9919-c43880fe6c04" path="/var/lib/kubelet/pods/a5a89a1e-ce5c-4e31-9919-c43880fe6c04/volumes" Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.818013 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wj9p7\" (UniqueName: \"kubernetes.io/projected/ef3b4399-44ec-476a-a0d5-e727b76c624e-kube-api-access-wj9p7\") pod \"nova-scheduler-0\" (UID: \"ef3b4399-44ec-476a-a0d5-e727b76c624e\") " pod="openstack/nova-scheduler-0" Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.861373 4754 scope.go:117] "RemoveContainer" containerID="dbafe75f4149545f66d7dcf67e5679e62b6cad3503ae5ede59a6619920905833" Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.869241 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.883952 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.885882 4754 scope.go:117] "RemoveContainer" containerID="dbafe75f4149545f66d7dcf67e5679e62b6cad3503ae5ede59a6619920905833" Jan 26 17:11:09 crc kubenswrapper[4754]: E0126 17:11:09.889725 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dbafe75f4149545f66d7dcf67e5679e62b6cad3503ae5ede59a6619920905833\": container with ID starting with dbafe75f4149545f66d7dcf67e5679e62b6cad3503ae5ede59a6619920905833 not found: ID does not exist" containerID="dbafe75f4149545f66d7dcf67e5679e62b6cad3503ae5ede59a6619920905833" Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.889753 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dbafe75f4149545f66d7dcf67e5679e62b6cad3503ae5ede59a6619920905833"} err="failed to get container status \"dbafe75f4149545f66d7dcf67e5679e62b6cad3503ae5ede59a6619920905833\": rpc error: code = NotFound desc = could not find container \"dbafe75f4149545f66d7dcf67e5679e62b6cad3503ae5ede59a6619920905833\": container with ID starting with dbafe75f4149545f66d7dcf67e5679e62b6cad3503ae5ede59a6619920905833 not found: ID does not exist" Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.910298 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-hxdzc" Jan 26 17:11:09 crc kubenswrapper[4754]: I0126 17:11:09.910863 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-hxdzc" Jan 26 17:11:10 crc kubenswrapper[4754]: I0126 17:11:10.116384 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Jan 26 17:11:10 crc kubenswrapper[4754]: I0126 17:11:10.200287 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"37f5b6db-8ef3-4e4f-9763-b150188c1466","Type":"ContainerStarted","Data":"40200c2ad928b0e5d0cf5e04a79d89f21e722b73f5f6bee5ec83eb7ebff6d363"} Jan 26 17:11:10 crc kubenswrapper[4754]: W0126 17:11:10.282584 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd2701786_b1e9_4a86_b751_63e1dd98eeba.slice/crio-fdc2672d5c94d24c79accfc673effaff68e21c4e594087180df6a01190672fab WatchSource:0}: Error finding container fdc2672d5c94d24c79accfc673effaff68e21c4e594087180df6a01190672fab: Status 404 returned error can't find the container with id fdc2672d5c94d24c79accfc673effaff68e21c4e594087180df6a01190672fab Jan 26 17:11:10 crc kubenswrapper[4754]: I0126 17:11:10.284628 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Jan 26 17:11:10 crc kubenswrapper[4754]: I0126 17:11:10.418897 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Jan 26 17:11:10 crc kubenswrapper[4754]: I0126 17:11:10.440204 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Jan 26 17:11:10 crc kubenswrapper[4754]: I0126 17:11:10.956465 4754 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-hxdzc" podUID="e3881c45-db75-4ca1-bb8f-92f413446265" containerName="registry-server" probeResult="failure" output=< Jan 26 17:11:10 crc kubenswrapper[4754]: timeout: failed to connect service ":50051" within 1s Jan 26 17:11:10 crc kubenswrapper[4754]: > Jan 26 17:11:11 crc kubenswrapper[4754]: I0126 17:11:11.213933 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2cff5231-483a-4d14-abf2-5ae85efc4edf","Type":"ContainerStarted","Data":"89811eb1ead413cb52a39b5cd2b62eee48456da05d7ecc6dda828260057c51b0"} Jan 26 17:11:11 crc kubenswrapper[4754]: I0126 17:11:11.217545 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"37f5b6db-8ef3-4e4f-9763-b150188c1466","Type":"ContainerStarted","Data":"b8bcc64a29d1de162bd574631509ab962e70a5eba797922712a474b28c35c163"} Jan 26 17:11:11 crc kubenswrapper[4754]: I0126 17:11:11.217586 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"37f5b6db-8ef3-4e4f-9763-b150188c1466","Type":"ContainerStarted","Data":"911ce027833e9fd70123c130dde99ab0e73b4f39eb0c6df15227bfe3a7a4e4cf"} Jan 26 17:11:11 crc kubenswrapper[4754]: I0126 17:11:11.219814 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"ef3b4399-44ec-476a-a0d5-e727b76c624e","Type":"ContainerStarted","Data":"5f0b0d2d8d721b498cb7253818255e0158226a13e1da30cdda99a766305b2734"} Jan 26 17:11:11 crc kubenswrapper[4754]: I0126 17:11:11.219859 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"ef3b4399-44ec-476a-a0d5-e727b76c624e","Type":"ContainerStarted","Data":"fca9d9cf754082914a4799e437c7f2999dfdb540035e937df1644ce078be05b6"} Jan 26 17:11:11 crc kubenswrapper[4754]: I0126 17:11:11.225480 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d2701786-b1e9-4a86-b751-63e1dd98eeba","Type":"ContainerStarted","Data":"86fe60aad1ad15e9ae3a4c78134c0e5a58adf5afaa53f605d1679062f022d2e7"} Jan 26 17:11:11 crc kubenswrapper[4754]: I0126 17:11:11.225534 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d2701786-b1e9-4a86-b751-63e1dd98eeba","Type":"ContainerStarted","Data":"3fac24e179c3d5fb3303205cbc6c04bd843a7190a169d7f411fe4d9ede9cd718"} Jan 26 17:11:11 crc kubenswrapper[4754]: I0126 17:11:11.225549 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d2701786-b1e9-4a86-b751-63e1dd98eeba","Type":"ContainerStarted","Data":"fdc2672d5c94d24c79accfc673effaff68e21c4e594087180df6a01190672fab"} Jan 26 17:11:11 crc kubenswrapper[4754]: I0126 17:11:11.240901 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.240882239 podStartE2EDuration="2.240882239s" podCreationTimestamp="2026-01-26 17:11:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 17:11:11.238642867 +0000 UTC m=+1437.762823311" watchObservedRunningTime="2026-01-26 17:11:11.240882239 +0000 UTC m=+1437.765062683" Jan 26 17:11:11 crc kubenswrapper[4754]: I0126 17:11:11.279719 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.279698859 podStartE2EDuration="2.279698859s" podCreationTimestamp="2026-01-26 17:11:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 17:11:11.277709024 +0000 UTC m=+1437.801889468" watchObservedRunningTime="2026-01-26 17:11:11.279698859 +0000 UTC m=+1437.803879313" Jan 26 17:11:11 crc kubenswrapper[4754]: I0126 17:11:11.287775 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.287753501 podStartE2EDuration="2.287753501s" podCreationTimestamp="2026-01-26 17:11:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 17:11:11.262160256 +0000 UTC m=+1437.786340710" watchObservedRunningTime="2026-01-26 17:11:11.287753501 +0000 UTC m=+1437.811933945" Jan 26 17:11:12 crc kubenswrapper[4754]: I0126 17:11:12.245769 4754 generic.go:334] "Generic (PLEG): container finished" podID="d40200db-b145-4a60-b337-82ae5032cc51" containerID="f3fe6e1c3d0a565d5c2f44c4be05b0cb2ef674bd9aca961cfe8edcb5ff320e24" exitCode=0 Jan 26 17:11:12 crc kubenswrapper[4754]: I0126 17:11:12.245855 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-55c6p" event={"ID":"d40200db-b145-4a60-b337-82ae5032cc51","Type":"ContainerDied","Data":"f3fe6e1c3d0a565d5c2f44c4be05b0cb2ef674bd9aca961cfe8edcb5ff320e24"} Jan 26 17:11:12 crc kubenswrapper[4754]: I0126 17:11:12.249093 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2cff5231-483a-4d14-abf2-5ae85efc4edf","Type":"ContainerStarted","Data":"bbb85d0bafc9230cf8a2edca595fa74b484d83abcf59bfc91946a805f5f6685a"} Jan 26 17:11:13 crc kubenswrapper[4754]: I0126 17:11:13.263067 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2cff5231-483a-4d14-abf2-5ae85efc4edf","Type":"ContainerStarted","Data":"e03fe24cc4619f6479549af2e568a0b9eb6dfb1b18e2f3a703e6d81495aa1029"} Jan 26 17:11:13 crc kubenswrapper[4754]: I0126 17:11:13.724634 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-55c6p" Jan 26 17:11:13 crc kubenswrapper[4754]: I0126 17:11:13.791920 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d40200db-b145-4a60-b337-82ae5032cc51-config-data\") pod \"d40200db-b145-4a60-b337-82ae5032cc51\" (UID: \"d40200db-b145-4a60-b337-82ae5032cc51\") " Jan 26 17:11:13 crc kubenswrapper[4754]: I0126 17:11:13.792145 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7km8l\" (UniqueName: \"kubernetes.io/projected/d40200db-b145-4a60-b337-82ae5032cc51-kube-api-access-7km8l\") pod \"d40200db-b145-4a60-b337-82ae5032cc51\" (UID: \"d40200db-b145-4a60-b337-82ae5032cc51\") " Jan 26 17:11:13 crc kubenswrapper[4754]: I0126 17:11:13.792233 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d40200db-b145-4a60-b337-82ae5032cc51-scripts\") pod \"d40200db-b145-4a60-b337-82ae5032cc51\" (UID: \"d40200db-b145-4a60-b337-82ae5032cc51\") " Jan 26 17:11:13 crc kubenswrapper[4754]: I0126 17:11:13.792282 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d40200db-b145-4a60-b337-82ae5032cc51-combined-ca-bundle\") pod \"d40200db-b145-4a60-b337-82ae5032cc51\" (UID: \"d40200db-b145-4a60-b337-82ae5032cc51\") " Jan 26 17:11:13 crc kubenswrapper[4754]: I0126 17:11:13.815365 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d40200db-b145-4a60-b337-82ae5032cc51-scripts" (OuterVolumeSpecName: "scripts") pod "d40200db-b145-4a60-b337-82ae5032cc51" (UID: "d40200db-b145-4a60-b337-82ae5032cc51"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:11:13 crc kubenswrapper[4754]: I0126 17:11:13.815687 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d40200db-b145-4a60-b337-82ae5032cc51-kube-api-access-7km8l" (OuterVolumeSpecName: "kube-api-access-7km8l") pod "d40200db-b145-4a60-b337-82ae5032cc51" (UID: "d40200db-b145-4a60-b337-82ae5032cc51"). InnerVolumeSpecName "kube-api-access-7km8l". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:11:13 crc kubenswrapper[4754]: I0126 17:11:13.843712 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d40200db-b145-4a60-b337-82ae5032cc51-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d40200db-b145-4a60-b337-82ae5032cc51" (UID: "d40200db-b145-4a60-b337-82ae5032cc51"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:11:13 crc kubenswrapper[4754]: I0126 17:11:13.856171 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d40200db-b145-4a60-b337-82ae5032cc51-config-data" (OuterVolumeSpecName: "config-data") pod "d40200db-b145-4a60-b337-82ae5032cc51" (UID: "d40200db-b145-4a60-b337-82ae5032cc51"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:11:13 crc kubenswrapper[4754]: I0126 17:11:13.895077 4754 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d40200db-b145-4a60-b337-82ae5032cc51-scripts\") on node \"crc\" DevicePath \"\"" Jan 26 17:11:13 crc kubenswrapper[4754]: I0126 17:11:13.895119 4754 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d40200db-b145-4a60-b337-82ae5032cc51-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 26 17:11:13 crc kubenswrapper[4754]: I0126 17:11:13.895135 4754 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d40200db-b145-4a60-b337-82ae5032cc51-config-data\") on node \"crc\" DevicePath \"\"" Jan 26 17:11:13 crc kubenswrapper[4754]: I0126 17:11:13.895150 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7km8l\" (UniqueName: \"kubernetes.io/projected/d40200db-b145-4a60-b337-82ae5032cc51-kube-api-access-7km8l\") on node \"crc\" DevicePath \"\"" Jan 26 17:11:14 crc kubenswrapper[4754]: I0126 17:11:14.273759 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2cff5231-483a-4d14-abf2-5ae85efc4edf","Type":"ContainerStarted","Data":"c702ed2bd4c2c40a50755ee03ac9ea27984331034e2ade38f8e5bca2f359f2be"} Jan 26 17:11:14 crc kubenswrapper[4754]: I0126 17:11:14.276043 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-55c6p" event={"ID":"d40200db-b145-4a60-b337-82ae5032cc51","Type":"ContainerDied","Data":"feef8482359e4251be97153f960b9e2ae8de7be83b6af60e9299abaaba674cf9"} Jan 26 17:11:14 crc kubenswrapper[4754]: I0126 17:11:14.276073 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="feef8482359e4251be97153f960b9e2ae8de7be83b6af60e9299abaaba674cf9" Jan 26 17:11:14 crc kubenswrapper[4754]: I0126 17:11:14.276126 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-55c6p" Jan 26 17:11:14 crc kubenswrapper[4754]: I0126 17:11:14.347114 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Jan 26 17:11:14 crc kubenswrapper[4754]: E0126 17:11:14.347653 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d40200db-b145-4a60-b337-82ae5032cc51" containerName="nova-cell1-conductor-db-sync" Jan 26 17:11:14 crc kubenswrapper[4754]: I0126 17:11:14.347692 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="d40200db-b145-4a60-b337-82ae5032cc51" containerName="nova-cell1-conductor-db-sync" Jan 26 17:11:14 crc kubenswrapper[4754]: I0126 17:11:14.347915 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="d40200db-b145-4a60-b337-82ae5032cc51" containerName="nova-cell1-conductor-db-sync" Jan 26 17:11:14 crc kubenswrapper[4754]: I0126 17:11:14.348632 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Jan 26 17:11:14 crc kubenswrapper[4754]: I0126 17:11:14.355610 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Jan 26 17:11:14 crc kubenswrapper[4754]: I0126 17:11:14.375641 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Jan 26 17:11:14 crc kubenswrapper[4754]: I0126 17:11:14.487965 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Jan 26 17:11:14 crc kubenswrapper[4754]: I0126 17:11:14.505426 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a81d671-1d63-4eac-a3fd-bf0e9bf38cc2-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"7a81d671-1d63-4eac-a3fd-bf0e9bf38cc2\") " pod="openstack/nova-cell1-conductor-0" Jan 26 17:11:14 crc kubenswrapper[4754]: I0126 17:11:14.505610 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7a81d671-1d63-4eac-a3fd-bf0e9bf38cc2-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"7a81d671-1d63-4eac-a3fd-bf0e9bf38cc2\") " pod="openstack/nova-cell1-conductor-0" Jan 26 17:11:14 crc kubenswrapper[4754]: I0126 17:11:14.505735 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jqx25\" (UniqueName: \"kubernetes.io/projected/7a81d671-1d63-4eac-a3fd-bf0e9bf38cc2-kube-api-access-jqx25\") pod \"nova-cell1-conductor-0\" (UID: \"7a81d671-1d63-4eac-a3fd-bf0e9bf38cc2\") " pod="openstack/nova-cell1-conductor-0" Jan 26 17:11:14 crc kubenswrapper[4754]: I0126 17:11:14.573462 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Jan 26 17:11:14 crc kubenswrapper[4754]: I0126 17:11:14.573813 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Jan 26 17:11:14 crc kubenswrapper[4754]: I0126 17:11:14.616007 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jqx25\" (UniqueName: \"kubernetes.io/projected/7a81d671-1d63-4eac-a3fd-bf0e9bf38cc2-kube-api-access-jqx25\") pod \"nova-cell1-conductor-0\" (UID: \"7a81d671-1d63-4eac-a3fd-bf0e9bf38cc2\") " pod="openstack/nova-cell1-conductor-0" Jan 26 17:11:14 crc kubenswrapper[4754]: I0126 17:11:14.616137 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a81d671-1d63-4eac-a3fd-bf0e9bf38cc2-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"7a81d671-1d63-4eac-a3fd-bf0e9bf38cc2\") " pod="openstack/nova-cell1-conductor-0" Jan 26 17:11:14 crc kubenswrapper[4754]: I0126 17:11:14.616967 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7a81d671-1d63-4eac-a3fd-bf0e9bf38cc2-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"7a81d671-1d63-4eac-a3fd-bf0e9bf38cc2\") " pod="openstack/nova-cell1-conductor-0" Jan 26 17:11:14 crc kubenswrapper[4754]: I0126 17:11:14.623651 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7a81d671-1d63-4eac-a3fd-bf0e9bf38cc2-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"7a81d671-1d63-4eac-a3fd-bf0e9bf38cc2\") " pod="openstack/nova-cell1-conductor-0" Jan 26 17:11:14 crc kubenswrapper[4754]: I0126 17:11:14.633800 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jqx25\" (UniqueName: \"kubernetes.io/projected/7a81d671-1d63-4eac-a3fd-bf0e9bf38cc2-kube-api-access-jqx25\") pod \"nova-cell1-conductor-0\" (UID: \"7a81d671-1d63-4eac-a3fd-bf0e9bf38cc2\") " pod="openstack/nova-cell1-conductor-0" Jan 26 17:11:14 crc kubenswrapper[4754]: I0126 17:11:14.640376 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a81d671-1d63-4eac-a3fd-bf0e9bf38cc2-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"7a81d671-1d63-4eac-a3fd-bf0e9bf38cc2\") " pod="openstack/nova-cell1-conductor-0" Jan 26 17:11:14 crc kubenswrapper[4754]: I0126 17:11:14.687736 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Jan 26 17:11:14 crc kubenswrapper[4754]: I0126 17:11:14.884606 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Jan 26 17:11:15 crc kubenswrapper[4754]: W0126 17:11:15.186230 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7a81d671_1d63_4eac_a3fd_bf0e9bf38cc2.slice/crio-52aba0d5faebb6ab3944da1d1ab352d7df2a763fb0de56533bb614b5ab07a765 WatchSource:0}: Error finding container 52aba0d5faebb6ab3944da1d1ab352d7df2a763fb0de56533bb614b5ab07a765: Status 404 returned error can't find the container with id 52aba0d5faebb6ab3944da1d1ab352d7df2a763fb0de56533bb614b5ab07a765 Jan 26 17:11:15 crc kubenswrapper[4754]: I0126 17:11:15.188275 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Jan 26 17:11:15 crc kubenswrapper[4754]: I0126 17:11:15.288614 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"7a81d671-1d63-4eac-a3fd-bf0e9bf38cc2","Type":"ContainerStarted","Data":"52aba0d5faebb6ab3944da1d1ab352d7df2a763fb0de56533bb614b5ab07a765"} Jan 26 17:11:16 crc kubenswrapper[4754]: I0126 17:11:16.301081 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2cff5231-483a-4d14-abf2-5ae85efc4edf","Type":"ContainerStarted","Data":"856f0015d714ff8b75734961028e9b2aa5c3cf587b1a971f3716bcc5f5112156"} Jan 26 17:11:16 crc kubenswrapper[4754]: I0126 17:11:16.301431 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Jan 26 17:11:16 crc kubenswrapper[4754]: I0126 17:11:16.302757 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"7a81d671-1d63-4eac-a3fd-bf0e9bf38cc2","Type":"ContainerStarted","Data":"b112be6ddcf84555b99e53b67b3e84d8d2f58484030af80d95bda8ff81f10692"} Jan 26 17:11:16 crc kubenswrapper[4754]: I0126 17:11:16.302912 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Jan 26 17:11:16 crc kubenswrapper[4754]: I0126 17:11:16.333451 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.608691575 podStartE2EDuration="7.333430004s" podCreationTimestamp="2026-01-26 17:11:09 +0000 UTC" firstStartedPulling="2026-01-26 17:11:10.437371975 +0000 UTC m=+1436.961552409" lastFinishedPulling="2026-01-26 17:11:15.162110404 +0000 UTC m=+1441.686290838" observedRunningTime="2026-01-26 17:11:16.322963415 +0000 UTC m=+1442.847143859" watchObservedRunningTime="2026-01-26 17:11:16.333430004 +0000 UTC m=+1442.857610438" Jan 26 17:11:16 crc kubenswrapper[4754]: I0126 17:11:16.347348 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.347331957 podStartE2EDuration="2.347331957s" podCreationTimestamp="2026-01-26 17:11:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 17:11:16.345558408 +0000 UTC m=+1442.869738852" watchObservedRunningTime="2026-01-26 17:11:16.347331957 +0000 UTC m=+1442.871512391" Jan 26 17:11:19 crc kubenswrapper[4754]: I0126 17:11:19.573859 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Jan 26 17:11:19 crc kubenswrapper[4754]: I0126 17:11:19.574604 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Jan 26 17:11:19 crc kubenswrapper[4754]: I0126 17:11:19.781832 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Jan 26 17:11:19 crc kubenswrapper[4754]: I0126 17:11:19.781875 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Jan 26 17:11:19 crc kubenswrapper[4754]: I0126 17:11:19.884379 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Jan 26 17:11:19 crc kubenswrapper[4754]: I0126 17:11:19.915981 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Jan 26 17:11:19 crc kubenswrapper[4754]: I0126 17:11:19.966558 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-hxdzc" Jan 26 17:11:20 crc kubenswrapper[4754]: I0126 17:11:20.022628 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-hxdzc" Jan 26 17:11:20 crc kubenswrapper[4754]: I0126 17:11:20.214390 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-hxdzc"] Jan 26 17:11:20 crc kubenswrapper[4754]: I0126 17:11:20.369199 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Jan 26 17:11:20 crc kubenswrapper[4754]: I0126 17:11:20.597840 4754 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="37f5b6db-8ef3-4e4f-9763-b150188c1466" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.201:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Jan 26 17:11:20 crc kubenswrapper[4754]: I0126 17:11:20.597840 4754 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="37f5b6db-8ef3-4e4f-9763-b150188c1466" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.201:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Jan 26 17:11:20 crc kubenswrapper[4754]: I0126 17:11:20.857876 4754 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="d2701786-b1e9-4a86-b751-63e1dd98eeba" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.202:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Jan 26 17:11:20 crc kubenswrapper[4754]: I0126 17:11:20.857900 4754 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="d2701786-b1e9-4a86-b751-63e1dd98eeba" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.202:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Jan 26 17:11:21 crc kubenswrapper[4754]: I0126 17:11:21.348854 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-hxdzc" podUID="e3881c45-db75-4ca1-bb8f-92f413446265" containerName="registry-server" containerID="cri-o://9e2c96a294df3c2e892df642a032ac13972ed913ae08a5cc7874baec5dd4fc6b" gracePeriod=2 Jan 26 17:11:22 crc kubenswrapper[4754]: I0126 17:11:22.052716 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hxdzc" Jan 26 17:11:22 crc kubenswrapper[4754]: I0126 17:11:22.170539 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e3881c45-db75-4ca1-bb8f-92f413446265-catalog-content\") pod \"e3881c45-db75-4ca1-bb8f-92f413446265\" (UID: \"e3881c45-db75-4ca1-bb8f-92f413446265\") " Jan 26 17:11:22 crc kubenswrapper[4754]: I0126 17:11:22.170609 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e3881c45-db75-4ca1-bb8f-92f413446265-utilities\") pod \"e3881c45-db75-4ca1-bb8f-92f413446265\" (UID: \"e3881c45-db75-4ca1-bb8f-92f413446265\") " Jan 26 17:11:22 crc kubenswrapper[4754]: I0126 17:11:22.170779 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6cc96\" (UniqueName: \"kubernetes.io/projected/e3881c45-db75-4ca1-bb8f-92f413446265-kube-api-access-6cc96\") pod \"e3881c45-db75-4ca1-bb8f-92f413446265\" (UID: \"e3881c45-db75-4ca1-bb8f-92f413446265\") " Jan 26 17:11:22 crc kubenswrapper[4754]: I0126 17:11:22.171344 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e3881c45-db75-4ca1-bb8f-92f413446265-utilities" (OuterVolumeSpecName: "utilities") pod "e3881c45-db75-4ca1-bb8f-92f413446265" (UID: "e3881c45-db75-4ca1-bb8f-92f413446265"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:11:22 crc kubenswrapper[4754]: I0126 17:11:22.171446 4754 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e3881c45-db75-4ca1-bb8f-92f413446265-utilities\") on node \"crc\" DevicePath \"\"" Jan 26 17:11:22 crc kubenswrapper[4754]: I0126 17:11:22.175884 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e3881c45-db75-4ca1-bb8f-92f413446265-kube-api-access-6cc96" (OuterVolumeSpecName: "kube-api-access-6cc96") pod "e3881c45-db75-4ca1-bb8f-92f413446265" (UID: "e3881c45-db75-4ca1-bb8f-92f413446265"). InnerVolumeSpecName "kube-api-access-6cc96". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:11:22 crc kubenswrapper[4754]: I0126 17:11:22.273304 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6cc96\" (UniqueName: \"kubernetes.io/projected/e3881c45-db75-4ca1-bb8f-92f413446265-kube-api-access-6cc96\") on node \"crc\" DevicePath \"\"" Jan 26 17:11:22 crc kubenswrapper[4754]: I0126 17:11:22.282515 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e3881c45-db75-4ca1-bb8f-92f413446265-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e3881c45-db75-4ca1-bb8f-92f413446265" (UID: "e3881c45-db75-4ca1-bb8f-92f413446265"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:11:22 crc kubenswrapper[4754]: I0126 17:11:22.357945 4754 generic.go:334] "Generic (PLEG): container finished" podID="e3881c45-db75-4ca1-bb8f-92f413446265" containerID="9e2c96a294df3c2e892df642a032ac13972ed913ae08a5cc7874baec5dd4fc6b" exitCode=0 Jan 26 17:11:22 crc kubenswrapper[4754]: I0126 17:11:22.357984 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hxdzc" event={"ID":"e3881c45-db75-4ca1-bb8f-92f413446265","Type":"ContainerDied","Data":"9e2c96a294df3c2e892df642a032ac13972ed913ae08a5cc7874baec5dd4fc6b"} Jan 26 17:11:22 crc kubenswrapper[4754]: I0126 17:11:22.358010 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hxdzc" event={"ID":"e3881c45-db75-4ca1-bb8f-92f413446265","Type":"ContainerDied","Data":"4185283ec6b8bfea7c969e1e116f6478ec1a3c39bee3af1453e9c98e0676dd40"} Jan 26 17:11:22 crc kubenswrapper[4754]: I0126 17:11:22.358007 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hxdzc" Jan 26 17:11:22 crc kubenswrapper[4754]: I0126 17:11:22.358036 4754 scope.go:117] "RemoveContainer" containerID="9e2c96a294df3c2e892df642a032ac13972ed913ae08a5cc7874baec5dd4fc6b" Jan 26 17:11:22 crc kubenswrapper[4754]: I0126 17:11:22.374537 4754 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e3881c45-db75-4ca1-bb8f-92f413446265-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 26 17:11:22 crc kubenswrapper[4754]: I0126 17:11:22.380610 4754 scope.go:117] "RemoveContainer" containerID="df4175ce477fff9b340f1ab44f40a75ad91b861e26473ef821f2251779a39ea4" Jan 26 17:11:22 crc kubenswrapper[4754]: I0126 17:11:22.392649 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-hxdzc"] Jan 26 17:11:22 crc kubenswrapper[4754]: I0126 17:11:22.402742 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-hxdzc"] Jan 26 17:11:22 crc kubenswrapper[4754]: I0126 17:11:22.406393 4754 scope.go:117] "RemoveContainer" containerID="6995e4eda258b809a08ef7ec99095737aca54eadc07eb6db51f9cb1dc5845921" Jan 26 17:11:22 crc kubenswrapper[4754]: I0126 17:11:22.441342 4754 scope.go:117] "RemoveContainer" containerID="9e2c96a294df3c2e892df642a032ac13972ed913ae08a5cc7874baec5dd4fc6b" Jan 26 17:11:22 crc kubenswrapper[4754]: E0126 17:11:22.441926 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9e2c96a294df3c2e892df642a032ac13972ed913ae08a5cc7874baec5dd4fc6b\": container with ID starting with 9e2c96a294df3c2e892df642a032ac13972ed913ae08a5cc7874baec5dd4fc6b not found: ID does not exist" containerID="9e2c96a294df3c2e892df642a032ac13972ed913ae08a5cc7874baec5dd4fc6b" Jan 26 17:11:22 crc kubenswrapper[4754]: I0126 17:11:22.442035 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9e2c96a294df3c2e892df642a032ac13972ed913ae08a5cc7874baec5dd4fc6b"} err="failed to get container status \"9e2c96a294df3c2e892df642a032ac13972ed913ae08a5cc7874baec5dd4fc6b\": rpc error: code = NotFound desc = could not find container \"9e2c96a294df3c2e892df642a032ac13972ed913ae08a5cc7874baec5dd4fc6b\": container with ID starting with 9e2c96a294df3c2e892df642a032ac13972ed913ae08a5cc7874baec5dd4fc6b not found: ID does not exist" Jan 26 17:11:22 crc kubenswrapper[4754]: I0126 17:11:22.442136 4754 scope.go:117] "RemoveContainer" containerID="df4175ce477fff9b340f1ab44f40a75ad91b861e26473ef821f2251779a39ea4" Jan 26 17:11:22 crc kubenswrapper[4754]: E0126 17:11:22.442590 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"df4175ce477fff9b340f1ab44f40a75ad91b861e26473ef821f2251779a39ea4\": container with ID starting with df4175ce477fff9b340f1ab44f40a75ad91b861e26473ef821f2251779a39ea4 not found: ID does not exist" containerID="df4175ce477fff9b340f1ab44f40a75ad91b861e26473ef821f2251779a39ea4" Jan 26 17:11:22 crc kubenswrapper[4754]: I0126 17:11:22.442621 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"df4175ce477fff9b340f1ab44f40a75ad91b861e26473ef821f2251779a39ea4"} err="failed to get container status \"df4175ce477fff9b340f1ab44f40a75ad91b861e26473ef821f2251779a39ea4\": rpc error: code = NotFound desc = could not find container \"df4175ce477fff9b340f1ab44f40a75ad91b861e26473ef821f2251779a39ea4\": container with ID starting with df4175ce477fff9b340f1ab44f40a75ad91b861e26473ef821f2251779a39ea4 not found: ID does not exist" Jan 26 17:11:22 crc kubenswrapper[4754]: I0126 17:11:22.442643 4754 scope.go:117] "RemoveContainer" containerID="6995e4eda258b809a08ef7ec99095737aca54eadc07eb6db51f9cb1dc5845921" Jan 26 17:11:22 crc kubenswrapper[4754]: E0126 17:11:22.442876 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6995e4eda258b809a08ef7ec99095737aca54eadc07eb6db51f9cb1dc5845921\": container with ID starting with 6995e4eda258b809a08ef7ec99095737aca54eadc07eb6db51f9cb1dc5845921 not found: ID does not exist" containerID="6995e4eda258b809a08ef7ec99095737aca54eadc07eb6db51f9cb1dc5845921" Jan 26 17:11:22 crc kubenswrapper[4754]: I0126 17:11:22.442976 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6995e4eda258b809a08ef7ec99095737aca54eadc07eb6db51f9cb1dc5845921"} err="failed to get container status \"6995e4eda258b809a08ef7ec99095737aca54eadc07eb6db51f9cb1dc5845921\": rpc error: code = NotFound desc = could not find container \"6995e4eda258b809a08ef7ec99095737aca54eadc07eb6db51f9cb1dc5845921\": container with ID starting with 6995e4eda258b809a08ef7ec99095737aca54eadc07eb6db51f9cb1dc5845921 not found: ID does not exist" Jan 26 17:11:23 crc kubenswrapper[4754]: I0126 17:11:23.781897 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e3881c45-db75-4ca1-bb8f-92f413446265" path="/var/lib/kubelet/pods/e3881c45-db75-4ca1-bb8f-92f413446265/volumes" Jan 26 17:11:24 crc kubenswrapper[4754]: I0126 17:11:24.716575 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Jan 26 17:11:29 crc kubenswrapper[4754]: I0126 17:11:29.579034 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Jan 26 17:11:29 crc kubenswrapper[4754]: I0126 17:11:29.579648 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Jan 26 17:11:29 crc kubenswrapper[4754]: I0126 17:11:29.584813 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Jan 26 17:11:29 crc kubenswrapper[4754]: I0126 17:11:29.586346 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Jan 26 17:11:29 crc kubenswrapper[4754]: I0126 17:11:29.793839 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Jan 26 17:11:29 crc kubenswrapper[4754]: I0126 17:11:29.793953 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Jan 26 17:11:29 crc kubenswrapper[4754]: I0126 17:11:29.796055 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Jan 26 17:11:29 crc kubenswrapper[4754]: I0126 17:11:29.796074 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Jan 26 17:11:29 crc kubenswrapper[4754]: I0126 17:11:29.799541 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Jan 26 17:11:29 crc kubenswrapper[4754]: I0126 17:11:29.802050 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Jan 26 17:11:29 crc kubenswrapper[4754]: I0126 17:11:29.976387 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-89c5cd4d5-8wrps"] Jan 26 17:11:29 crc kubenswrapper[4754]: E0126 17:11:29.976890 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3881c45-db75-4ca1-bb8f-92f413446265" containerName="extract-content" Jan 26 17:11:29 crc kubenswrapper[4754]: I0126 17:11:29.976915 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3881c45-db75-4ca1-bb8f-92f413446265" containerName="extract-content" Jan 26 17:11:29 crc kubenswrapper[4754]: E0126 17:11:29.976931 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3881c45-db75-4ca1-bb8f-92f413446265" containerName="extract-utilities" Jan 26 17:11:29 crc kubenswrapper[4754]: I0126 17:11:29.976939 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3881c45-db75-4ca1-bb8f-92f413446265" containerName="extract-utilities" Jan 26 17:11:29 crc kubenswrapper[4754]: E0126 17:11:29.976955 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3881c45-db75-4ca1-bb8f-92f413446265" containerName="registry-server" Jan 26 17:11:29 crc kubenswrapper[4754]: I0126 17:11:29.976962 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3881c45-db75-4ca1-bb8f-92f413446265" containerName="registry-server" Jan 26 17:11:29 crc kubenswrapper[4754]: I0126 17:11:29.977196 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="e3881c45-db75-4ca1-bb8f-92f413446265" containerName="registry-server" Jan 26 17:11:29 crc kubenswrapper[4754]: I0126 17:11:29.978247 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-89c5cd4d5-8wrps" Jan 26 17:11:29 crc kubenswrapper[4754]: I0126 17:11:29.983512 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-89c5cd4d5-8wrps"] Jan 26 17:11:30 crc kubenswrapper[4754]: I0126 17:11:30.115992 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/443a6550-c50d-453e-9f10-3875fcb41c18-ovsdbserver-nb\") pod \"dnsmasq-dns-89c5cd4d5-8wrps\" (UID: \"443a6550-c50d-453e-9f10-3875fcb41c18\") " pod="openstack/dnsmasq-dns-89c5cd4d5-8wrps" Jan 26 17:11:30 crc kubenswrapper[4754]: I0126 17:11:30.116314 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bstvk\" (UniqueName: \"kubernetes.io/projected/443a6550-c50d-453e-9f10-3875fcb41c18-kube-api-access-bstvk\") pod \"dnsmasq-dns-89c5cd4d5-8wrps\" (UID: \"443a6550-c50d-453e-9f10-3875fcb41c18\") " pod="openstack/dnsmasq-dns-89c5cd4d5-8wrps" Jan 26 17:11:30 crc kubenswrapper[4754]: I0126 17:11:30.116379 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/443a6550-c50d-453e-9f10-3875fcb41c18-dns-svc\") pod \"dnsmasq-dns-89c5cd4d5-8wrps\" (UID: \"443a6550-c50d-453e-9f10-3875fcb41c18\") " pod="openstack/dnsmasq-dns-89c5cd4d5-8wrps" Jan 26 17:11:30 crc kubenswrapper[4754]: I0126 17:11:30.116421 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/443a6550-c50d-453e-9f10-3875fcb41c18-config\") pod \"dnsmasq-dns-89c5cd4d5-8wrps\" (UID: \"443a6550-c50d-453e-9f10-3875fcb41c18\") " pod="openstack/dnsmasq-dns-89c5cd4d5-8wrps" Jan 26 17:11:30 crc kubenswrapper[4754]: I0126 17:11:30.116464 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/443a6550-c50d-453e-9f10-3875fcb41c18-dns-swift-storage-0\") pod \"dnsmasq-dns-89c5cd4d5-8wrps\" (UID: \"443a6550-c50d-453e-9f10-3875fcb41c18\") " pod="openstack/dnsmasq-dns-89c5cd4d5-8wrps" Jan 26 17:11:30 crc kubenswrapper[4754]: I0126 17:11:30.116493 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/443a6550-c50d-453e-9f10-3875fcb41c18-ovsdbserver-sb\") pod \"dnsmasq-dns-89c5cd4d5-8wrps\" (UID: \"443a6550-c50d-453e-9f10-3875fcb41c18\") " pod="openstack/dnsmasq-dns-89c5cd4d5-8wrps" Jan 26 17:11:30 crc kubenswrapper[4754]: I0126 17:11:30.217948 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/443a6550-c50d-453e-9f10-3875fcb41c18-ovsdbserver-sb\") pod \"dnsmasq-dns-89c5cd4d5-8wrps\" (UID: \"443a6550-c50d-453e-9f10-3875fcb41c18\") " pod="openstack/dnsmasq-dns-89c5cd4d5-8wrps" Jan 26 17:11:30 crc kubenswrapper[4754]: I0126 17:11:30.218021 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/443a6550-c50d-453e-9f10-3875fcb41c18-ovsdbserver-nb\") pod \"dnsmasq-dns-89c5cd4d5-8wrps\" (UID: \"443a6550-c50d-453e-9f10-3875fcb41c18\") " pod="openstack/dnsmasq-dns-89c5cd4d5-8wrps" Jan 26 17:11:30 crc kubenswrapper[4754]: I0126 17:11:30.218082 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bstvk\" (UniqueName: \"kubernetes.io/projected/443a6550-c50d-453e-9f10-3875fcb41c18-kube-api-access-bstvk\") pod \"dnsmasq-dns-89c5cd4d5-8wrps\" (UID: \"443a6550-c50d-453e-9f10-3875fcb41c18\") " pod="openstack/dnsmasq-dns-89c5cd4d5-8wrps" Jan 26 17:11:30 crc kubenswrapper[4754]: I0126 17:11:30.218157 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/443a6550-c50d-453e-9f10-3875fcb41c18-dns-svc\") pod \"dnsmasq-dns-89c5cd4d5-8wrps\" (UID: \"443a6550-c50d-453e-9f10-3875fcb41c18\") " pod="openstack/dnsmasq-dns-89c5cd4d5-8wrps" Jan 26 17:11:30 crc kubenswrapper[4754]: I0126 17:11:30.218205 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/443a6550-c50d-453e-9f10-3875fcb41c18-config\") pod \"dnsmasq-dns-89c5cd4d5-8wrps\" (UID: \"443a6550-c50d-453e-9f10-3875fcb41c18\") " pod="openstack/dnsmasq-dns-89c5cd4d5-8wrps" Jan 26 17:11:30 crc kubenswrapper[4754]: I0126 17:11:30.218257 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/443a6550-c50d-453e-9f10-3875fcb41c18-dns-swift-storage-0\") pod \"dnsmasq-dns-89c5cd4d5-8wrps\" (UID: \"443a6550-c50d-453e-9f10-3875fcb41c18\") " pod="openstack/dnsmasq-dns-89c5cd4d5-8wrps" Jan 26 17:11:30 crc kubenswrapper[4754]: I0126 17:11:30.219125 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/443a6550-c50d-453e-9f10-3875fcb41c18-dns-swift-storage-0\") pod \"dnsmasq-dns-89c5cd4d5-8wrps\" (UID: \"443a6550-c50d-453e-9f10-3875fcb41c18\") " pod="openstack/dnsmasq-dns-89c5cd4d5-8wrps" Jan 26 17:11:30 crc kubenswrapper[4754]: I0126 17:11:30.219129 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/443a6550-c50d-453e-9f10-3875fcb41c18-ovsdbserver-sb\") pod \"dnsmasq-dns-89c5cd4d5-8wrps\" (UID: \"443a6550-c50d-453e-9f10-3875fcb41c18\") " pod="openstack/dnsmasq-dns-89c5cd4d5-8wrps" Jan 26 17:11:30 crc kubenswrapper[4754]: I0126 17:11:30.219603 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/443a6550-c50d-453e-9f10-3875fcb41c18-ovsdbserver-nb\") pod \"dnsmasq-dns-89c5cd4d5-8wrps\" (UID: \"443a6550-c50d-453e-9f10-3875fcb41c18\") " pod="openstack/dnsmasq-dns-89c5cd4d5-8wrps" Jan 26 17:11:30 crc kubenswrapper[4754]: I0126 17:11:30.219938 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/443a6550-c50d-453e-9f10-3875fcb41c18-config\") pod \"dnsmasq-dns-89c5cd4d5-8wrps\" (UID: \"443a6550-c50d-453e-9f10-3875fcb41c18\") " pod="openstack/dnsmasq-dns-89c5cd4d5-8wrps" Jan 26 17:11:30 crc kubenswrapper[4754]: I0126 17:11:30.219938 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/443a6550-c50d-453e-9f10-3875fcb41c18-dns-svc\") pod \"dnsmasq-dns-89c5cd4d5-8wrps\" (UID: \"443a6550-c50d-453e-9f10-3875fcb41c18\") " pod="openstack/dnsmasq-dns-89c5cd4d5-8wrps" Jan 26 17:11:30 crc kubenswrapper[4754]: I0126 17:11:30.241027 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bstvk\" (UniqueName: \"kubernetes.io/projected/443a6550-c50d-453e-9f10-3875fcb41c18-kube-api-access-bstvk\") pod \"dnsmasq-dns-89c5cd4d5-8wrps\" (UID: \"443a6550-c50d-453e-9f10-3875fcb41c18\") " pod="openstack/dnsmasq-dns-89c5cd4d5-8wrps" Jan 26 17:11:30 crc kubenswrapper[4754]: I0126 17:11:30.313158 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-89c5cd4d5-8wrps" Jan 26 17:11:30 crc kubenswrapper[4754]: I0126 17:11:30.431408 4754 generic.go:334] "Generic (PLEG): container finished" podID="7922c387-0282-4b8e-a907-af058a637575" containerID="f05e7698a2c643b6ab539c22a2c1983a4bfc697ed974bcb9c7ba86d8a596f3e9" exitCode=137 Jan 26 17:11:30 crc kubenswrapper[4754]: I0126 17:11:30.432500 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"7922c387-0282-4b8e-a907-af058a637575","Type":"ContainerDied","Data":"f05e7698a2c643b6ab539c22a2c1983a4bfc697ed974bcb9c7ba86d8a596f3e9"} Jan 26 17:11:30 crc kubenswrapper[4754]: I0126 17:11:30.432557 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"7922c387-0282-4b8e-a907-af058a637575","Type":"ContainerDied","Data":"28065df71fcddc1b218655ca0c7410bcad8f81754d83496259ba407bb264c0f4"} Jan 26 17:11:30 crc kubenswrapper[4754]: I0126 17:11:30.432573 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="28065df71fcddc1b218655ca0c7410bcad8f81754d83496259ba407bb264c0f4" Jan 26 17:11:30 crc kubenswrapper[4754]: I0126 17:11:30.485019 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Jan 26 17:11:30 crc kubenswrapper[4754]: I0126 17:11:30.628020 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4q7bk\" (UniqueName: \"kubernetes.io/projected/7922c387-0282-4b8e-a907-af058a637575-kube-api-access-4q7bk\") pod \"7922c387-0282-4b8e-a907-af058a637575\" (UID: \"7922c387-0282-4b8e-a907-af058a637575\") " Jan 26 17:11:30 crc kubenswrapper[4754]: I0126 17:11:30.628127 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7922c387-0282-4b8e-a907-af058a637575-config-data\") pod \"7922c387-0282-4b8e-a907-af058a637575\" (UID: \"7922c387-0282-4b8e-a907-af058a637575\") " Jan 26 17:11:30 crc kubenswrapper[4754]: I0126 17:11:30.628164 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7922c387-0282-4b8e-a907-af058a637575-combined-ca-bundle\") pod \"7922c387-0282-4b8e-a907-af058a637575\" (UID: \"7922c387-0282-4b8e-a907-af058a637575\") " Jan 26 17:11:30 crc kubenswrapper[4754]: I0126 17:11:30.632469 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7922c387-0282-4b8e-a907-af058a637575-kube-api-access-4q7bk" (OuterVolumeSpecName: "kube-api-access-4q7bk") pod "7922c387-0282-4b8e-a907-af058a637575" (UID: "7922c387-0282-4b8e-a907-af058a637575"). InnerVolumeSpecName "kube-api-access-4q7bk". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:11:30 crc kubenswrapper[4754]: I0126 17:11:30.657199 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7922c387-0282-4b8e-a907-af058a637575-config-data" (OuterVolumeSpecName: "config-data") pod "7922c387-0282-4b8e-a907-af058a637575" (UID: "7922c387-0282-4b8e-a907-af058a637575"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:11:30 crc kubenswrapper[4754]: I0126 17:11:30.661907 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7922c387-0282-4b8e-a907-af058a637575-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7922c387-0282-4b8e-a907-af058a637575" (UID: "7922c387-0282-4b8e-a907-af058a637575"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:11:30 crc kubenswrapper[4754]: I0126 17:11:30.730421 4754 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7922c387-0282-4b8e-a907-af058a637575-config-data\") on node \"crc\" DevicePath \"\"" Jan 26 17:11:30 crc kubenswrapper[4754]: I0126 17:11:30.730467 4754 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7922c387-0282-4b8e-a907-af058a637575-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 26 17:11:30 crc kubenswrapper[4754]: I0126 17:11:30.730483 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4q7bk\" (UniqueName: \"kubernetes.io/projected/7922c387-0282-4b8e-a907-af058a637575-kube-api-access-4q7bk\") on node \"crc\" DevicePath \"\"" Jan 26 17:11:30 crc kubenswrapper[4754]: W0126 17:11:30.810965 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod443a6550_c50d_453e_9f10_3875fcb41c18.slice/crio-d4197cb2cfe9741f52a6fe12eb036913fc38177359ef56817c5baf702af7553d WatchSource:0}: Error finding container d4197cb2cfe9741f52a6fe12eb036913fc38177359ef56817c5baf702af7553d: Status 404 returned error can't find the container with id d4197cb2cfe9741f52a6fe12eb036913fc38177359ef56817c5baf702af7553d Jan 26 17:11:30 crc kubenswrapper[4754]: I0126 17:11:30.811756 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-89c5cd4d5-8wrps"] Jan 26 17:11:31 crc kubenswrapper[4754]: I0126 17:11:31.439932 4754 generic.go:334] "Generic (PLEG): container finished" podID="443a6550-c50d-453e-9f10-3875fcb41c18" containerID="121ef5f4c966fb80a1b1b516ebd2e15d37b4b339bcf34b10842645bd2d6f26a4" exitCode=0 Jan 26 17:11:31 crc kubenswrapper[4754]: I0126 17:11:31.440065 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-89c5cd4d5-8wrps" event={"ID":"443a6550-c50d-453e-9f10-3875fcb41c18","Type":"ContainerDied","Data":"121ef5f4c966fb80a1b1b516ebd2e15d37b4b339bcf34b10842645bd2d6f26a4"} Jan 26 17:11:31 crc kubenswrapper[4754]: I0126 17:11:31.440454 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Jan 26 17:11:31 crc kubenswrapper[4754]: I0126 17:11:31.440478 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-89c5cd4d5-8wrps" event={"ID":"443a6550-c50d-453e-9f10-3875fcb41c18","Type":"ContainerStarted","Data":"d4197cb2cfe9741f52a6fe12eb036913fc38177359ef56817c5baf702af7553d"} Jan 26 17:11:31 crc kubenswrapper[4754]: I0126 17:11:31.660164 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Jan 26 17:11:31 crc kubenswrapper[4754]: I0126 17:11:31.698092 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Jan 26 17:11:31 crc kubenswrapper[4754]: I0126 17:11:31.705508 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Jan 26 17:11:31 crc kubenswrapper[4754]: E0126 17:11:31.706077 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7922c387-0282-4b8e-a907-af058a637575" containerName="nova-cell1-novncproxy-novncproxy" Jan 26 17:11:31 crc kubenswrapper[4754]: I0126 17:11:31.706099 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="7922c387-0282-4b8e-a907-af058a637575" containerName="nova-cell1-novncproxy-novncproxy" Jan 26 17:11:31 crc kubenswrapper[4754]: I0126 17:11:31.707948 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="7922c387-0282-4b8e-a907-af058a637575" containerName="nova-cell1-novncproxy-novncproxy" Jan 26 17:11:31 crc kubenswrapper[4754]: I0126 17:11:31.709463 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Jan 26 17:11:31 crc kubenswrapper[4754]: I0126 17:11:31.714081 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Jan 26 17:11:31 crc kubenswrapper[4754]: I0126 17:11:31.714326 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Jan 26 17:11:31 crc kubenswrapper[4754]: I0126 17:11:31.714479 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Jan 26 17:11:31 crc kubenswrapper[4754]: I0126 17:11:31.715160 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Jan 26 17:11:31 crc kubenswrapper[4754]: I0126 17:11:31.778580 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7922c387-0282-4b8e-a907-af058a637575" path="/var/lib/kubelet/pods/7922c387-0282-4b8e-a907-af058a637575/volumes" Jan 26 17:11:31 crc kubenswrapper[4754]: I0126 17:11:31.859197 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fbd0b5ca-aa69-4a3a-a81d-5b912f5db6a1-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"fbd0b5ca-aa69-4a3a-a81d-5b912f5db6a1\") " pod="openstack/nova-cell1-novncproxy-0" Jan 26 17:11:31 crc kubenswrapper[4754]: I0126 17:11:31.859296 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fbd0b5ca-aa69-4a3a-a81d-5b912f5db6a1-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"fbd0b5ca-aa69-4a3a-a81d-5b912f5db6a1\") " pod="openstack/nova-cell1-novncproxy-0" Jan 26 17:11:31 crc kubenswrapper[4754]: I0126 17:11:31.859337 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/fbd0b5ca-aa69-4a3a-a81d-5b912f5db6a1-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"fbd0b5ca-aa69-4a3a-a81d-5b912f5db6a1\") " pod="openstack/nova-cell1-novncproxy-0" Jan 26 17:11:31 crc kubenswrapper[4754]: I0126 17:11:31.860119 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/fbd0b5ca-aa69-4a3a-a81d-5b912f5db6a1-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"fbd0b5ca-aa69-4a3a-a81d-5b912f5db6a1\") " pod="openstack/nova-cell1-novncproxy-0" Jan 26 17:11:31 crc kubenswrapper[4754]: I0126 17:11:31.860394 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j5n8m\" (UniqueName: \"kubernetes.io/projected/fbd0b5ca-aa69-4a3a-a81d-5b912f5db6a1-kube-api-access-j5n8m\") pod \"nova-cell1-novncproxy-0\" (UID: \"fbd0b5ca-aa69-4a3a-a81d-5b912f5db6a1\") " pod="openstack/nova-cell1-novncproxy-0" Jan 26 17:11:31 crc kubenswrapper[4754]: I0126 17:11:31.939636 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Jan 26 17:11:31 crc kubenswrapper[4754]: I0126 17:11:31.940389 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2cff5231-483a-4d14-abf2-5ae85efc4edf" containerName="ceilometer-central-agent" containerID="cri-o://bbb85d0bafc9230cf8a2edca595fa74b484d83abcf59bfc91946a805f5f6685a" gracePeriod=30 Jan 26 17:11:31 crc kubenswrapper[4754]: I0126 17:11:31.940499 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2cff5231-483a-4d14-abf2-5ae85efc4edf" containerName="proxy-httpd" containerID="cri-o://856f0015d714ff8b75734961028e9b2aa5c3cf587b1a971f3716bcc5f5112156" gracePeriod=30 Jan 26 17:11:31 crc kubenswrapper[4754]: I0126 17:11:31.940499 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2cff5231-483a-4d14-abf2-5ae85efc4edf" containerName="ceilometer-notification-agent" containerID="cri-o://e03fe24cc4619f6479549af2e568a0b9eb6dfb1b18e2f3a703e6d81495aa1029" gracePeriod=30 Jan 26 17:11:31 crc kubenswrapper[4754]: I0126 17:11:31.940505 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2cff5231-483a-4d14-abf2-5ae85efc4edf" containerName="sg-core" containerID="cri-o://c702ed2bd4c2c40a50755ee03ac9ea27984331034e2ade38f8e5bca2f359f2be" gracePeriod=30 Jan 26 17:11:31 crc kubenswrapper[4754]: I0126 17:11:31.956007 4754 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="2cff5231-483a-4d14-abf2-5ae85efc4edf" containerName="proxy-httpd" probeResult="failure" output="Get \"https://10.217.0.203:3000/\": EOF" Jan 26 17:11:31 crc kubenswrapper[4754]: I0126 17:11:31.961907 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fbd0b5ca-aa69-4a3a-a81d-5b912f5db6a1-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"fbd0b5ca-aa69-4a3a-a81d-5b912f5db6a1\") " pod="openstack/nova-cell1-novncproxy-0" Jan 26 17:11:31 crc kubenswrapper[4754]: I0126 17:11:31.961979 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fbd0b5ca-aa69-4a3a-a81d-5b912f5db6a1-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"fbd0b5ca-aa69-4a3a-a81d-5b912f5db6a1\") " pod="openstack/nova-cell1-novncproxy-0" Jan 26 17:11:31 crc kubenswrapper[4754]: I0126 17:11:31.962011 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/fbd0b5ca-aa69-4a3a-a81d-5b912f5db6a1-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"fbd0b5ca-aa69-4a3a-a81d-5b912f5db6a1\") " pod="openstack/nova-cell1-novncproxy-0" Jan 26 17:11:31 crc kubenswrapper[4754]: I0126 17:11:31.962047 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/fbd0b5ca-aa69-4a3a-a81d-5b912f5db6a1-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"fbd0b5ca-aa69-4a3a-a81d-5b912f5db6a1\") " pod="openstack/nova-cell1-novncproxy-0" Jan 26 17:11:31 crc kubenswrapper[4754]: I0126 17:11:31.962175 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j5n8m\" (UniqueName: \"kubernetes.io/projected/fbd0b5ca-aa69-4a3a-a81d-5b912f5db6a1-kube-api-access-j5n8m\") pod \"nova-cell1-novncproxy-0\" (UID: \"fbd0b5ca-aa69-4a3a-a81d-5b912f5db6a1\") " pod="openstack/nova-cell1-novncproxy-0" Jan 26 17:11:31 crc kubenswrapper[4754]: I0126 17:11:31.967713 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fbd0b5ca-aa69-4a3a-a81d-5b912f5db6a1-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"fbd0b5ca-aa69-4a3a-a81d-5b912f5db6a1\") " pod="openstack/nova-cell1-novncproxy-0" Jan 26 17:11:31 crc kubenswrapper[4754]: I0126 17:11:31.968231 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/fbd0b5ca-aa69-4a3a-a81d-5b912f5db6a1-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"fbd0b5ca-aa69-4a3a-a81d-5b912f5db6a1\") " pod="openstack/nova-cell1-novncproxy-0" Jan 26 17:11:31 crc kubenswrapper[4754]: I0126 17:11:31.969048 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/fbd0b5ca-aa69-4a3a-a81d-5b912f5db6a1-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"fbd0b5ca-aa69-4a3a-a81d-5b912f5db6a1\") " pod="openstack/nova-cell1-novncproxy-0" Jan 26 17:11:31 crc kubenswrapper[4754]: I0126 17:11:31.973769 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fbd0b5ca-aa69-4a3a-a81d-5b912f5db6a1-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"fbd0b5ca-aa69-4a3a-a81d-5b912f5db6a1\") " pod="openstack/nova-cell1-novncproxy-0" Jan 26 17:11:31 crc kubenswrapper[4754]: I0126 17:11:31.997276 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j5n8m\" (UniqueName: \"kubernetes.io/projected/fbd0b5ca-aa69-4a3a-a81d-5b912f5db6a1-kube-api-access-j5n8m\") pod \"nova-cell1-novncproxy-0\" (UID: \"fbd0b5ca-aa69-4a3a-a81d-5b912f5db6a1\") " pod="openstack/nova-cell1-novncproxy-0" Jan 26 17:11:32 crc kubenswrapper[4754]: I0126 17:11:32.031499 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Jan 26 17:11:32 crc kubenswrapper[4754]: I0126 17:11:32.396364 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Jan 26 17:11:32 crc kubenswrapper[4754]: I0126 17:11:32.450832 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-89c5cd4d5-8wrps" event={"ID":"443a6550-c50d-453e-9f10-3875fcb41c18","Type":"ContainerStarted","Data":"da940fcb0040028dfd4aebd2f8c5ab637c5ac7300ce1a8c53fa14feb47b58d3b"} Jan 26 17:11:32 crc kubenswrapper[4754]: I0126 17:11:32.450973 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-89c5cd4d5-8wrps" Jan 26 17:11:32 crc kubenswrapper[4754]: I0126 17:11:32.453921 4754 generic.go:334] "Generic (PLEG): container finished" podID="2cff5231-483a-4d14-abf2-5ae85efc4edf" containerID="856f0015d714ff8b75734961028e9b2aa5c3cf587b1a971f3716bcc5f5112156" exitCode=0 Jan 26 17:11:32 crc kubenswrapper[4754]: I0126 17:11:32.453954 4754 generic.go:334] "Generic (PLEG): container finished" podID="2cff5231-483a-4d14-abf2-5ae85efc4edf" containerID="c702ed2bd4c2c40a50755ee03ac9ea27984331034e2ade38f8e5bca2f359f2be" exitCode=2 Jan 26 17:11:32 crc kubenswrapper[4754]: I0126 17:11:32.453968 4754 generic.go:334] "Generic (PLEG): container finished" podID="2cff5231-483a-4d14-abf2-5ae85efc4edf" containerID="bbb85d0bafc9230cf8a2edca595fa74b484d83abcf59bfc91946a805f5f6685a" exitCode=0 Jan 26 17:11:32 crc kubenswrapper[4754]: I0126 17:11:32.453988 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2cff5231-483a-4d14-abf2-5ae85efc4edf","Type":"ContainerDied","Data":"856f0015d714ff8b75734961028e9b2aa5c3cf587b1a971f3716bcc5f5112156"} Jan 26 17:11:32 crc kubenswrapper[4754]: I0126 17:11:32.454037 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2cff5231-483a-4d14-abf2-5ae85efc4edf","Type":"ContainerDied","Data":"c702ed2bd4c2c40a50755ee03ac9ea27984331034e2ade38f8e5bca2f359f2be"} Jan 26 17:11:32 crc kubenswrapper[4754]: I0126 17:11:32.454051 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2cff5231-483a-4d14-abf2-5ae85efc4edf","Type":"ContainerDied","Data":"bbb85d0bafc9230cf8a2edca595fa74b484d83abcf59bfc91946a805f5f6685a"} Jan 26 17:11:32 crc kubenswrapper[4754]: I0126 17:11:32.454136 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="d2701786-b1e9-4a86-b751-63e1dd98eeba" containerName="nova-api-log" containerID="cri-o://3fac24e179c3d5fb3303205cbc6c04bd843a7190a169d7f411fe4d9ede9cd718" gracePeriod=30 Jan 26 17:11:32 crc kubenswrapper[4754]: I0126 17:11:32.454195 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="d2701786-b1e9-4a86-b751-63e1dd98eeba" containerName="nova-api-api" containerID="cri-o://86fe60aad1ad15e9ae3a4c78134c0e5a58adf5afaa53f605d1679062f022d2e7" gracePeriod=30 Jan 26 17:11:32 crc kubenswrapper[4754]: I0126 17:11:32.557420 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-89c5cd4d5-8wrps" podStartSLOduration=3.557399198 podStartE2EDuration="3.557399198s" podCreationTimestamp="2026-01-26 17:11:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 17:11:32.477207109 +0000 UTC m=+1459.001387563" watchObservedRunningTime="2026-01-26 17:11:32.557399198 +0000 UTC m=+1459.081579632" Jan 26 17:11:32 crc kubenswrapper[4754]: I0126 17:11:32.558083 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Jan 26 17:11:33 crc kubenswrapper[4754]: I0126 17:11:33.463219 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"fbd0b5ca-aa69-4a3a-a81d-5b912f5db6a1","Type":"ContainerStarted","Data":"c60be909a10986bf150efff58852b2de137784ea1ab7cdbc90439bce54a772ff"} Jan 26 17:11:33 crc kubenswrapper[4754]: I0126 17:11:33.463519 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"fbd0b5ca-aa69-4a3a-a81d-5b912f5db6a1","Type":"ContainerStarted","Data":"0ef1727e795e36761bac42603e6c63e9073d64cc914fd55c89bc25c8f3a1c9f0"} Jan 26 17:11:33 crc kubenswrapper[4754]: I0126 17:11:33.465490 4754 generic.go:334] "Generic (PLEG): container finished" podID="d2701786-b1e9-4a86-b751-63e1dd98eeba" containerID="3fac24e179c3d5fb3303205cbc6c04bd843a7190a169d7f411fe4d9ede9cd718" exitCode=143 Jan 26 17:11:33 crc kubenswrapper[4754]: I0126 17:11:33.465543 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d2701786-b1e9-4a86-b751-63e1dd98eeba","Type":"ContainerDied","Data":"3fac24e179c3d5fb3303205cbc6c04bd843a7190a169d7f411fe4d9ede9cd718"} Jan 26 17:11:33 crc kubenswrapper[4754]: I0126 17:11:33.478184 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.478168204 podStartE2EDuration="2.478168204s" podCreationTimestamp="2026-01-26 17:11:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 17:11:33.476494888 +0000 UTC m=+1460.000675332" watchObservedRunningTime="2026-01-26 17:11:33.478168204 +0000 UTC m=+1460.002348638" Jan 26 17:11:34 crc kubenswrapper[4754]: I0126 17:11:34.176681 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 26 17:11:34 crc kubenswrapper[4754]: I0126 17:11:34.364888 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2cff5231-483a-4d14-abf2-5ae85efc4edf-config-data\") pod \"2cff5231-483a-4d14-abf2-5ae85efc4edf\" (UID: \"2cff5231-483a-4d14-abf2-5ae85efc4edf\") " Jan 26 17:11:34 crc kubenswrapper[4754]: I0126 17:11:34.365219 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2cff5231-483a-4d14-abf2-5ae85efc4edf-run-httpd\") pod \"2cff5231-483a-4d14-abf2-5ae85efc4edf\" (UID: \"2cff5231-483a-4d14-abf2-5ae85efc4edf\") " Jan 26 17:11:34 crc kubenswrapper[4754]: I0126 17:11:34.365269 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2cff5231-483a-4d14-abf2-5ae85efc4edf-log-httpd\") pod \"2cff5231-483a-4d14-abf2-5ae85efc4edf\" (UID: \"2cff5231-483a-4d14-abf2-5ae85efc4edf\") " Jan 26 17:11:34 crc kubenswrapper[4754]: I0126 17:11:34.365304 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-frlx6\" (UniqueName: \"kubernetes.io/projected/2cff5231-483a-4d14-abf2-5ae85efc4edf-kube-api-access-frlx6\") pod \"2cff5231-483a-4d14-abf2-5ae85efc4edf\" (UID: \"2cff5231-483a-4d14-abf2-5ae85efc4edf\") " Jan 26 17:11:34 crc kubenswrapper[4754]: I0126 17:11:34.365383 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2cff5231-483a-4d14-abf2-5ae85efc4edf-combined-ca-bundle\") pod \"2cff5231-483a-4d14-abf2-5ae85efc4edf\" (UID: \"2cff5231-483a-4d14-abf2-5ae85efc4edf\") " Jan 26 17:11:34 crc kubenswrapper[4754]: I0126 17:11:34.365715 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2cff5231-483a-4d14-abf2-5ae85efc4edf-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "2cff5231-483a-4d14-abf2-5ae85efc4edf" (UID: "2cff5231-483a-4d14-abf2-5ae85efc4edf"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:11:34 crc kubenswrapper[4754]: I0126 17:11:34.365812 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2cff5231-483a-4d14-abf2-5ae85efc4edf-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "2cff5231-483a-4d14-abf2-5ae85efc4edf" (UID: "2cff5231-483a-4d14-abf2-5ae85efc4edf"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:11:34 crc kubenswrapper[4754]: I0126 17:11:34.366042 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2cff5231-483a-4d14-abf2-5ae85efc4edf-scripts\") pod \"2cff5231-483a-4d14-abf2-5ae85efc4edf\" (UID: \"2cff5231-483a-4d14-abf2-5ae85efc4edf\") " Jan 26 17:11:34 crc kubenswrapper[4754]: I0126 17:11:34.366103 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2cff5231-483a-4d14-abf2-5ae85efc4edf-sg-core-conf-yaml\") pod \"2cff5231-483a-4d14-abf2-5ae85efc4edf\" (UID: \"2cff5231-483a-4d14-abf2-5ae85efc4edf\") " Jan 26 17:11:34 crc kubenswrapper[4754]: I0126 17:11:34.366136 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/2cff5231-483a-4d14-abf2-5ae85efc4edf-ceilometer-tls-certs\") pod \"2cff5231-483a-4d14-abf2-5ae85efc4edf\" (UID: \"2cff5231-483a-4d14-abf2-5ae85efc4edf\") " Jan 26 17:11:34 crc kubenswrapper[4754]: I0126 17:11:34.366723 4754 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2cff5231-483a-4d14-abf2-5ae85efc4edf-run-httpd\") on node \"crc\" DevicePath \"\"" Jan 26 17:11:34 crc kubenswrapper[4754]: I0126 17:11:34.366762 4754 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2cff5231-483a-4d14-abf2-5ae85efc4edf-log-httpd\") on node \"crc\" DevicePath \"\"" Jan 26 17:11:34 crc kubenswrapper[4754]: I0126 17:11:34.372797 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2cff5231-483a-4d14-abf2-5ae85efc4edf-kube-api-access-frlx6" (OuterVolumeSpecName: "kube-api-access-frlx6") pod "2cff5231-483a-4d14-abf2-5ae85efc4edf" (UID: "2cff5231-483a-4d14-abf2-5ae85efc4edf"). InnerVolumeSpecName "kube-api-access-frlx6". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:11:34 crc kubenswrapper[4754]: I0126 17:11:34.375805 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2cff5231-483a-4d14-abf2-5ae85efc4edf-scripts" (OuterVolumeSpecName: "scripts") pod "2cff5231-483a-4d14-abf2-5ae85efc4edf" (UID: "2cff5231-483a-4d14-abf2-5ae85efc4edf"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:11:34 crc kubenswrapper[4754]: I0126 17:11:34.398958 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2cff5231-483a-4d14-abf2-5ae85efc4edf-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "2cff5231-483a-4d14-abf2-5ae85efc4edf" (UID: "2cff5231-483a-4d14-abf2-5ae85efc4edf"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:11:34 crc kubenswrapper[4754]: I0126 17:11:34.436561 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2cff5231-483a-4d14-abf2-5ae85efc4edf-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "2cff5231-483a-4d14-abf2-5ae85efc4edf" (UID: "2cff5231-483a-4d14-abf2-5ae85efc4edf"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:11:34 crc kubenswrapper[4754]: I0126 17:11:34.465462 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2cff5231-483a-4d14-abf2-5ae85efc4edf-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2cff5231-483a-4d14-abf2-5ae85efc4edf" (UID: "2cff5231-483a-4d14-abf2-5ae85efc4edf"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:11:34 crc kubenswrapper[4754]: I0126 17:11:34.468014 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-frlx6\" (UniqueName: \"kubernetes.io/projected/2cff5231-483a-4d14-abf2-5ae85efc4edf-kube-api-access-frlx6\") on node \"crc\" DevicePath \"\"" Jan 26 17:11:34 crc kubenswrapper[4754]: I0126 17:11:34.468046 4754 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2cff5231-483a-4d14-abf2-5ae85efc4edf-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 26 17:11:34 crc kubenswrapper[4754]: I0126 17:11:34.468057 4754 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2cff5231-483a-4d14-abf2-5ae85efc4edf-scripts\") on node \"crc\" DevicePath \"\"" Jan 26 17:11:34 crc kubenswrapper[4754]: I0126 17:11:34.468066 4754 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2cff5231-483a-4d14-abf2-5ae85efc4edf-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Jan 26 17:11:34 crc kubenswrapper[4754]: I0126 17:11:34.468074 4754 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/2cff5231-483a-4d14-abf2-5ae85efc4edf-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Jan 26 17:11:34 crc kubenswrapper[4754]: I0126 17:11:34.477936 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2cff5231-483a-4d14-abf2-5ae85efc4edf-config-data" (OuterVolumeSpecName: "config-data") pod "2cff5231-483a-4d14-abf2-5ae85efc4edf" (UID: "2cff5231-483a-4d14-abf2-5ae85efc4edf"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:11:34 crc kubenswrapper[4754]: I0126 17:11:34.482511 4754 generic.go:334] "Generic (PLEG): container finished" podID="2cff5231-483a-4d14-abf2-5ae85efc4edf" containerID="e03fe24cc4619f6479549af2e568a0b9eb6dfb1b18e2f3a703e6d81495aa1029" exitCode=0 Jan 26 17:11:34 crc kubenswrapper[4754]: I0126 17:11:34.482571 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2cff5231-483a-4d14-abf2-5ae85efc4edf","Type":"ContainerDied","Data":"e03fe24cc4619f6479549af2e568a0b9eb6dfb1b18e2f3a703e6d81495aa1029"} Jan 26 17:11:34 crc kubenswrapper[4754]: I0126 17:11:34.482888 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2cff5231-483a-4d14-abf2-5ae85efc4edf","Type":"ContainerDied","Data":"89811eb1ead413cb52a39b5cd2b62eee48456da05d7ecc6dda828260057c51b0"} Jan 26 17:11:34 crc kubenswrapper[4754]: I0126 17:11:34.482628 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 26 17:11:34 crc kubenswrapper[4754]: I0126 17:11:34.482922 4754 scope.go:117] "RemoveContainer" containerID="856f0015d714ff8b75734961028e9b2aa5c3cf587b1a971f3716bcc5f5112156" Jan 26 17:11:34 crc kubenswrapper[4754]: I0126 17:11:34.540097 4754 scope.go:117] "RemoveContainer" containerID="c702ed2bd4c2c40a50755ee03ac9ea27984331034e2ade38f8e5bca2f359f2be" Jan 26 17:11:34 crc kubenswrapper[4754]: I0126 17:11:34.570157 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Jan 26 17:11:34 crc kubenswrapper[4754]: I0126 17:11:34.572281 4754 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2cff5231-483a-4d14-abf2-5ae85efc4edf-config-data\") on node \"crc\" DevicePath \"\"" Jan 26 17:11:34 crc kubenswrapper[4754]: I0126 17:11:34.582251 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Jan 26 17:11:34 crc kubenswrapper[4754]: I0126 17:11:34.587332 4754 scope.go:117] "RemoveContainer" containerID="e03fe24cc4619f6479549af2e568a0b9eb6dfb1b18e2f3a703e6d81495aa1029" Jan 26 17:11:34 crc kubenswrapper[4754]: I0126 17:11:34.594198 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Jan 26 17:11:34 crc kubenswrapper[4754]: E0126 17:11:34.595379 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2cff5231-483a-4d14-abf2-5ae85efc4edf" containerName="ceilometer-central-agent" Jan 26 17:11:34 crc kubenswrapper[4754]: I0126 17:11:34.595530 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="2cff5231-483a-4d14-abf2-5ae85efc4edf" containerName="ceilometer-central-agent" Jan 26 17:11:34 crc kubenswrapper[4754]: E0126 17:11:34.595699 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2cff5231-483a-4d14-abf2-5ae85efc4edf" containerName="ceilometer-notification-agent" Jan 26 17:11:34 crc kubenswrapper[4754]: I0126 17:11:34.595790 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="2cff5231-483a-4d14-abf2-5ae85efc4edf" containerName="ceilometer-notification-agent" Jan 26 17:11:34 crc kubenswrapper[4754]: E0126 17:11:34.595883 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2cff5231-483a-4d14-abf2-5ae85efc4edf" containerName="proxy-httpd" Jan 26 17:11:34 crc kubenswrapper[4754]: I0126 17:11:34.595970 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="2cff5231-483a-4d14-abf2-5ae85efc4edf" containerName="proxy-httpd" Jan 26 17:11:34 crc kubenswrapper[4754]: E0126 17:11:34.596057 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2cff5231-483a-4d14-abf2-5ae85efc4edf" containerName="sg-core" Jan 26 17:11:34 crc kubenswrapper[4754]: I0126 17:11:34.596136 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="2cff5231-483a-4d14-abf2-5ae85efc4edf" containerName="sg-core" Jan 26 17:11:34 crc kubenswrapper[4754]: I0126 17:11:34.596446 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="2cff5231-483a-4d14-abf2-5ae85efc4edf" containerName="sg-core" Jan 26 17:11:34 crc kubenswrapper[4754]: I0126 17:11:34.596547 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="2cff5231-483a-4d14-abf2-5ae85efc4edf" containerName="proxy-httpd" Jan 26 17:11:34 crc kubenswrapper[4754]: I0126 17:11:34.596651 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="2cff5231-483a-4d14-abf2-5ae85efc4edf" containerName="ceilometer-central-agent" Jan 26 17:11:34 crc kubenswrapper[4754]: I0126 17:11:34.596770 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="2cff5231-483a-4d14-abf2-5ae85efc4edf" containerName="ceilometer-notification-agent" Jan 26 17:11:34 crc kubenswrapper[4754]: I0126 17:11:34.599790 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 26 17:11:34 crc kubenswrapper[4754]: I0126 17:11:34.603366 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Jan 26 17:11:34 crc kubenswrapper[4754]: I0126 17:11:34.603803 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Jan 26 17:11:34 crc kubenswrapper[4754]: I0126 17:11:34.605814 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Jan 26 17:11:34 crc kubenswrapper[4754]: I0126 17:11:34.614261 4754 scope.go:117] "RemoveContainer" containerID="bbb85d0bafc9230cf8a2edca595fa74b484d83abcf59bfc91946a805f5f6685a" Jan 26 17:11:34 crc kubenswrapper[4754]: I0126 17:11:34.625914 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Jan 26 17:11:34 crc kubenswrapper[4754]: I0126 17:11:34.653123 4754 scope.go:117] "RemoveContainer" containerID="856f0015d714ff8b75734961028e9b2aa5c3cf587b1a971f3716bcc5f5112156" Jan 26 17:11:34 crc kubenswrapper[4754]: E0126 17:11:34.653425 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"856f0015d714ff8b75734961028e9b2aa5c3cf587b1a971f3716bcc5f5112156\": container with ID starting with 856f0015d714ff8b75734961028e9b2aa5c3cf587b1a971f3716bcc5f5112156 not found: ID does not exist" containerID="856f0015d714ff8b75734961028e9b2aa5c3cf587b1a971f3716bcc5f5112156" Jan 26 17:11:34 crc kubenswrapper[4754]: I0126 17:11:34.653453 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"856f0015d714ff8b75734961028e9b2aa5c3cf587b1a971f3716bcc5f5112156"} err="failed to get container status \"856f0015d714ff8b75734961028e9b2aa5c3cf587b1a971f3716bcc5f5112156\": rpc error: code = NotFound desc = could not find container \"856f0015d714ff8b75734961028e9b2aa5c3cf587b1a971f3716bcc5f5112156\": container with ID starting with 856f0015d714ff8b75734961028e9b2aa5c3cf587b1a971f3716bcc5f5112156 not found: ID does not exist" Jan 26 17:11:34 crc kubenswrapper[4754]: I0126 17:11:34.653475 4754 scope.go:117] "RemoveContainer" containerID="c702ed2bd4c2c40a50755ee03ac9ea27984331034e2ade38f8e5bca2f359f2be" Jan 26 17:11:34 crc kubenswrapper[4754]: E0126 17:11:34.653702 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c702ed2bd4c2c40a50755ee03ac9ea27984331034e2ade38f8e5bca2f359f2be\": container with ID starting with c702ed2bd4c2c40a50755ee03ac9ea27984331034e2ade38f8e5bca2f359f2be not found: ID does not exist" containerID="c702ed2bd4c2c40a50755ee03ac9ea27984331034e2ade38f8e5bca2f359f2be" Jan 26 17:11:34 crc kubenswrapper[4754]: I0126 17:11:34.653723 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c702ed2bd4c2c40a50755ee03ac9ea27984331034e2ade38f8e5bca2f359f2be"} err="failed to get container status \"c702ed2bd4c2c40a50755ee03ac9ea27984331034e2ade38f8e5bca2f359f2be\": rpc error: code = NotFound desc = could not find container \"c702ed2bd4c2c40a50755ee03ac9ea27984331034e2ade38f8e5bca2f359f2be\": container with ID starting with c702ed2bd4c2c40a50755ee03ac9ea27984331034e2ade38f8e5bca2f359f2be not found: ID does not exist" Jan 26 17:11:34 crc kubenswrapper[4754]: I0126 17:11:34.653744 4754 scope.go:117] "RemoveContainer" containerID="e03fe24cc4619f6479549af2e568a0b9eb6dfb1b18e2f3a703e6d81495aa1029" Jan 26 17:11:34 crc kubenswrapper[4754]: E0126 17:11:34.654024 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e03fe24cc4619f6479549af2e568a0b9eb6dfb1b18e2f3a703e6d81495aa1029\": container with ID starting with e03fe24cc4619f6479549af2e568a0b9eb6dfb1b18e2f3a703e6d81495aa1029 not found: ID does not exist" containerID="e03fe24cc4619f6479549af2e568a0b9eb6dfb1b18e2f3a703e6d81495aa1029" Jan 26 17:11:34 crc kubenswrapper[4754]: I0126 17:11:34.654048 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e03fe24cc4619f6479549af2e568a0b9eb6dfb1b18e2f3a703e6d81495aa1029"} err="failed to get container status \"e03fe24cc4619f6479549af2e568a0b9eb6dfb1b18e2f3a703e6d81495aa1029\": rpc error: code = NotFound desc = could not find container \"e03fe24cc4619f6479549af2e568a0b9eb6dfb1b18e2f3a703e6d81495aa1029\": container with ID starting with e03fe24cc4619f6479549af2e568a0b9eb6dfb1b18e2f3a703e6d81495aa1029 not found: ID does not exist" Jan 26 17:11:34 crc kubenswrapper[4754]: I0126 17:11:34.654062 4754 scope.go:117] "RemoveContainer" containerID="bbb85d0bafc9230cf8a2edca595fa74b484d83abcf59bfc91946a805f5f6685a" Jan 26 17:11:34 crc kubenswrapper[4754]: E0126 17:11:34.654274 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bbb85d0bafc9230cf8a2edca595fa74b484d83abcf59bfc91946a805f5f6685a\": container with ID starting with bbb85d0bafc9230cf8a2edca595fa74b484d83abcf59bfc91946a805f5f6685a not found: ID does not exist" containerID="bbb85d0bafc9230cf8a2edca595fa74b484d83abcf59bfc91946a805f5f6685a" Jan 26 17:11:34 crc kubenswrapper[4754]: I0126 17:11:34.654293 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bbb85d0bafc9230cf8a2edca595fa74b484d83abcf59bfc91946a805f5f6685a"} err="failed to get container status \"bbb85d0bafc9230cf8a2edca595fa74b484d83abcf59bfc91946a805f5f6685a\": rpc error: code = NotFound desc = could not find container \"bbb85d0bafc9230cf8a2edca595fa74b484d83abcf59bfc91946a805f5f6685a\": container with ID starting with bbb85d0bafc9230cf8a2edca595fa74b484d83abcf59bfc91946a805f5f6685a not found: ID does not exist" Jan 26 17:11:34 crc kubenswrapper[4754]: I0126 17:11:34.776149 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b41f5e0e-1829-41f9-926c-ce15178e16c8-log-httpd\") pod \"ceilometer-0\" (UID: \"b41f5e0e-1829-41f9-926c-ce15178e16c8\") " pod="openstack/ceilometer-0" Jan 26 17:11:34 crc kubenswrapper[4754]: I0126 17:11:34.776239 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/b41f5e0e-1829-41f9-926c-ce15178e16c8-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"b41f5e0e-1829-41f9-926c-ce15178e16c8\") " pod="openstack/ceilometer-0" Jan 26 17:11:34 crc kubenswrapper[4754]: I0126 17:11:34.776274 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b41f5e0e-1829-41f9-926c-ce15178e16c8-scripts\") pod \"ceilometer-0\" (UID: \"b41f5e0e-1829-41f9-926c-ce15178e16c8\") " pod="openstack/ceilometer-0" Jan 26 17:11:34 crc kubenswrapper[4754]: I0126 17:11:34.776919 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b41f5e0e-1829-41f9-926c-ce15178e16c8-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b41f5e0e-1829-41f9-926c-ce15178e16c8\") " pod="openstack/ceilometer-0" Jan 26 17:11:34 crc kubenswrapper[4754]: I0126 17:11:34.777176 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b41f5e0e-1829-41f9-926c-ce15178e16c8-config-data\") pod \"ceilometer-0\" (UID: \"b41f5e0e-1829-41f9-926c-ce15178e16c8\") " pod="openstack/ceilometer-0" Jan 26 17:11:34 crc kubenswrapper[4754]: I0126 17:11:34.778177 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b41f5e0e-1829-41f9-926c-ce15178e16c8-run-httpd\") pod \"ceilometer-0\" (UID: \"b41f5e0e-1829-41f9-926c-ce15178e16c8\") " pod="openstack/ceilometer-0" Jan 26 17:11:34 crc kubenswrapper[4754]: I0126 17:11:34.778289 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b41f5e0e-1829-41f9-926c-ce15178e16c8-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b41f5e0e-1829-41f9-926c-ce15178e16c8\") " pod="openstack/ceilometer-0" Jan 26 17:11:34 crc kubenswrapper[4754]: I0126 17:11:34.778539 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5mtp7\" (UniqueName: \"kubernetes.io/projected/b41f5e0e-1829-41f9-926c-ce15178e16c8-kube-api-access-5mtp7\") pod \"ceilometer-0\" (UID: \"b41f5e0e-1829-41f9-926c-ce15178e16c8\") " pod="openstack/ceilometer-0" Jan 26 17:11:34 crc kubenswrapper[4754]: I0126 17:11:34.879908 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b41f5e0e-1829-41f9-926c-ce15178e16c8-run-httpd\") pod \"ceilometer-0\" (UID: \"b41f5e0e-1829-41f9-926c-ce15178e16c8\") " pod="openstack/ceilometer-0" Jan 26 17:11:34 crc kubenswrapper[4754]: I0126 17:11:34.879979 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b41f5e0e-1829-41f9-926c-ce15178e16c8-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b41f5e0e-1829-41f9-926c-ce15178e16c8\") " pod="openstack/ceilometer-0" Jan 26 17:11:34 crc kubenswrapper[4754]: I0126 17:11:34.880063 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5mtp7\" (UniqueName: \"kubernetes.io/projected/b41f5e0e-1829-41f9-926c-ce15178e16c8-kube-api-access-5mtp7\") pod \"ceilometer-0\" (UID: \"b41f5e0e-1829-41f9-926c-ce15178e16c8\") " pod="openstack/ceilometer-0" Jan 26 17:11:34 crc kubenswrapper[4754]: I0126 17:11:34.880153 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b41f5e0e-1829-41f9-926c-ce15178e16c8-log-httpd\") pod \"ceilometer-0\" (UID: \"b41f5e0e-1829-41f9-926c-ce15178e16c8\") " pod="openstack/ceilometer-0" Jan 26 17:11:34 crc kubenswrapper[4754]: I0126 17:11:34.880236 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/b41f5e0e-1829-41f9-926c-ce15178e16c8-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"b41f5e0e-1829-41f9-926c-ce15178e16c8\") " pod="openstack/ceilometer-0" Jan 26 17:11:34 crc kubenswrapper[4754]: I0126 17:11:34.880311 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b41f5e0e-1829-41f9-926c-ce15178e16c8-scripts\") pod \"ceilometer-0\" (UID: \"b41f5e0e-1829-41f9-926c-ce15178e16c8\") " pod="openstack/ceilometer-0" Jan 26 17:11:34 crc kubenswrapper[4754]: I0126 17:11:34.880360 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b41f5e0e-1829-41f9-926c-ce15178e16c8-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b41f5e0e-1829-41f9-926c-ce15178e16c8\") " pod="openstack/ceilometer-0" Jan 26 17:11:34 crc kubenswrapper[4754]: I0126 17:11:34.880426 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b41f5e0e-1829-41f9-926c-ce15178e16c8-config-data\") pod \"ceilometer-0\" (UID: \"b41f5e0e-1829-41f9-926c-ce15178e16c8\") " pod="openstack/ceilometer-0" Jan 26 17:11:34 crc kubenswrapper[4754]: I0126 17:11:34.880960 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b41f5e0e-1829-41f9-926c-ce15178e16c8-log-httpd\") pod \"ceilometer-0\" (UID: \"b41f5e0e-1829-41f9-926c-ce15178e16c8\") " pod="openstack/ceilometer-0" Jan 26 17:11:34 crc kubenswrapper[4754]: I0126 17:11:34.881053 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b41f5e0e-1829-41f9-926c-ce15178e16c8-run-httpd\") pod \"ceilometer-0\" (UID: \"b41f5e0e-1829-41f9-926c-ce15178e16c8\") " pod="openstack/ceilometer-0" Jan 26 17:11:34 crc kubenswrapper[4754]: I0126 17:11:34.884177 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b41f5e0e-1829-41f9-926c-ce15178e16c8-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b41f5e0e-1829-41f9-926c-ce15178e16c8\") " pod="openstack/ceilometer-0" Jan 26 17:11:34 crc kubenswrapper[4754]: I0126 17:11:34.884785 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b41f5e0e-1829-41f9-926c-ce15178e16c8-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b41f5e0e-1829-41f9-926c-ce15178e16c8\") " pod="openstack/ceilometer-0" Jan 26 17:11:34 crc kubenswrapper[4754]: I0126 17:11:34.885248 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/b41f5e0e-1829-41f9-926c-ce15178e16c8-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"b41f5e0e-1829-41f9-926c-ce15178e16c8\") " pod="openstack/ceilometer-0" Jan 26 17:11:34 crc kubenswrapper[4754]: I0126 17:11:34.885428 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b41f5e0e-1829-41f9-926c-ce15178e16c8-config-data\") pod \"ceilometer-0\" (UID: \"b41f5e0e-1829-41f9-926c-ce15178e16c8\") " pod="openstack/ceilometer-0" Jan 26 17:11:34 crc kubenswrapper[4754]: I0126 17:11:34.885657 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b41f5e0e-1829-41f9-926c-ce15178e16c8-scripts\") pod \"ceilometer-0\" (UID: \"b41f5e0e-1829-41f9-926c-ce15178e16c8\") " pod="openstack/ceilometer-0" Jan 26 17:11:34 crc kubenswrapper[4754]: I0126 17:11:34.898230 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5mtp7\" (UniqueName: \"kubernetes.io/projected/b41f5e0e-1829-41f9-926c-ce15178e16c8-kube-api-access-5mtp7\") pod \"ceilometer-0\" (UID: \"b41f5e0e-1829-41f9-926c-ce15178e16c8\") " pod="openstack/ceilometer-0" Jan 26 17:11:34 crc kubenswrapper[4754]: I0126 17:11:34.918925 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 26 17:11:35 crc kubenswrapper[4754]: I0126 17:11:35.359318 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Jan 26 17:11:35 crc kubenswrapper[4754]: W0126 17:11:35.365448 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb41f5e0e_1829_41f9_926c_ce15178e16c8.slice/crio-4da04a8a275c70521db7a28ecdde67bcf00885fb21591798e5f111ea361e2d8d WatchSource:0}: Error finding container 4da04a8a275c70521db7a28ecdde67bcf00885fb21591798e5f111ea361e2d8d: Status 404 returned error can't find the container with id 4da04a8a275c70521db7a28ecdde67bcf00885fb21591798e5f111ea361e2d8d Jan 26 17:11:35 crc kubenswrapper[4754]: I0126 17:11:35.492125 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b41f5e0e-1829-41f9-926c-ce15178e16c8","Type":"ContainerStarted","Data":"4da04a8a275c70521db7a28ecdde67bcf00885fb21591798e5f111ea361e2d8d"} Jan 26 17:11:35 crc kubenswrapper[4754]: I0126 17:11:35.781514 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2cff5231-483a-4d14-abf2-5ae85efc4edf" path="/var/lib/kubelet/pods/2cff5231-483a-4d14-abf2-5ae85efc4edf/volumes" Jan 26 17:11:36 crc kubenswrapper[4754]: I0126 17:11:36.162546 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Jan 26 17:11:36 crc kubenswrapper[4754]: I0126 17:11:36.314794 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h547r\" (UniqueName: \"kubernetes.io/projected/d2701786-b1e9-4a86-b751-63e1dd98eeba-kube-api-access-h547r\") pod \"d2701786-b1e9-4a86-b751-63e1dd98eeba\" (UID: \"d2701786-b1e9-4a86-b751-63e1dd98eeba\") " Jan 26 17:11:36 crc kubenswrapper[4754]: I0126 17:11:36.314850 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d2701786-b1e9-4a86-b751-63e1dd98eeba-logs\") pod \"d2701786-b1e9-4a86-b751-63e1dd98eeba\" (UID: \"d2701786-b1e9-4a86-b751-63e1dd98eeba\") " Jan 26 17:11:36 crc kubenswrapper[4754]: I0126 17:11:36.315062 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2701786-b1e9-4a86-b751-63e1dd98eeba-combined-ca-bundle\") pod \"d2701786-b1e9-4a86-b751-63e1dd98eeba\" (UID: \"d2701786-b1e9-4a86-b751-63e1dd98eeba\") " Jan 26 17:11:36 crc kubenswrapper[4754]: I0126 17:11:36.315122 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d2701786-b1e9-4a86-b751-63e1dd98eeba-config-data\") pod \"d2701786-b1e9-4a86-b751-63e1dd98eeba\" (UID: \"d2701786-b1e9-4a86-b751-63e1dd98eeba\") " Jan 26 17:11:36 crc kubenswrapper[4754]: I0126 17:11:36.315968 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d2701786-b1e9-4a86-b751-63e1dd98eeba-logs" (OuterVolumeSpecName: "logs") pod "d2701786-b1e9-4a86-b751-63e1dd98eeba" (UID: "d2701786-b1e9-4a86-b751-63e1dd98eeba"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:11:36 crc kubenswrapper[4754]: I0126 17:11:36.322898 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d2701786-b1e9-4a86-b751-63e1dd98eeba-kube-api-access-h547r" (OuterVolumeSpecName: "kube-api-access-h547r") pod "d2701786-b1e9-4a86-b751-63e1dd98eeba" (UID: "d2701786-b1e9-4a86-b751-63e1dd98eeba"). InnerVolumeSpecName "kube-api-access-h547r". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:11:36 crc kubenswrapper[4754]: I0126 17:11:36.346059 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d2701786-b1e9-4a86-b751-63e1dd98eeba-config-data" (OuterVolumeSpecName: "config-data") pod "d2701786-b1e9-4a86-b751-63e1dd98eeba" (UID: "d2701786-b1e9-4a86-b751-63e1dd98eeba"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:11:36 crc kubenswrapper[4754]: I0126 17:11:36.357601 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d2701786-b1e9-4a86-b751-63e1dd98eeba-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d2701786-b1e9-4a86-b751-63e1dd98eeba" (UID: "d2701786-b1e9-4a86-b751-63e1dd98eeba"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:11:36 crc kubenswrapper[4754]: I0126 17:11:36.418334 4754 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2701786-b1e9-4a86-b751-63e1dd98eeba-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 26 17:11:36 crc kubenswrapper[4754]: I0126 17:11:36.418370 4754 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d2701786-b1e9-4a86-b751-63e1dd98eeba-config-data\") on node \"crc\" DevicePath \"\"" Jan 26 17:11:36 crc kubenswrapper[4754]: I0126 17:11:36.418382 4754 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d2701786-b1e9-4a86-b751-63e1dd98eeba-logs\") on node \"crc\" DevicePath \"\"" Jan 26 17:11:36 crc kubenswrapper[4754]: I0126 17:11:36.418395 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h547r\" (UniqueName: \"kubernetes.io/projected/d2701786-b1e9-4a86-b751-63e1dd98eeba-kube-api-access-h547r\") on node \"crc\" DevicePath \"\"" Jan 26 17:11:36 crc kubenswrapper[4754]: I0126 17:11:36.504516 4754 generic.go:334] "Generic (PLEG): container finished" podID="d2701786-b1e9-4a86-b751-63e1dd98eeba" containerID="86fe60aad1ad15e9ae3a4c78134c0e5a58adf5afaa53f605d1679062f022d2e7" exitCode=0 Jan 26 17:11:36 crc kubenswrapper[4754]: I0126 17:11:36.504572 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Jan 26 17:11:36 crc kubenswrapper[4754]: I0126 17:11:36.504601 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d2701786-b1e9-4a86-b751-63e1dd98eeba","Type":"ContainerDied","Data":"86fe60aad1ad15e9ae3a4c78134c0e5a58adf5afaa53f605d1679062f022d2e7"} Jan 26 17:11:36 crc kubenswrapper[4754]: I0126 17:11:36.504632 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d2701786-b1e9-4a86-b751-63e1dd98eeba","Type":"ContainerDied","Data":"fdc2672d5c94d24c79accfc673effaff68e21c4e594087180df6a01190672fab"} Jan 26 17:11:36 crc kubenswrapper[4754]: I0126 17:11:36.504650 4754 scope.go:117] "RemoveContainer" containerID="86fe60aad1ad15e9ae3a4c78134c0e5a58adf5afaa53f605d1679062f022d2e7" Jan 26 17:11:36 crc kubenswrapper[4754]: I0126 17:11:36.506427 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b41f5e0e-1829-41f9-926c-ce15178e16c8","Type":"ContainerStarted","Data":"7172d14b3d1371fed007b4ce70bd1449f170414f4145af577bb936c3cd72ae3a"} Jan 26 17:11:36 crc kubenswrapper[4754]: I0126 17:11:36.548250 4754 scope.go:117] "RemoveContainer" containerID="3fac24e179c3d5fb3303205cbc6c04bd843a7190a169d7f411fe4d9ede9cd718" Jan 26 17:11:36 crc kubenswrapper[4754]: I0126 17:11:36.559639 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Jan 26 17:11:36 crc kubenswrapper[4754]: I0126 17:11:36.576530 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Jan 26 17:11:36 crc kubenswrapper[4754]: I0126 17:11:36.596492 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Jan 26 17:11:36 crc kubenswrapper[4754]: E0126 17:11:36.596919 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2701786-b1e9-4a86-b751-63e1dd98eeba" containerName="nova-api-api" Jan 26 17:11:36 crc kubenswrapper[4754]: I0126 17:11:36.596937 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2701786-b1e9-4a86-b751-63e1dd98eeba" containerName="nova-api-api" Jan 26 17:11:36 crc kubenswrapper[4754]: E0126 17:11:36.596964 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2701786-b1e9-4a86-b751-63e1dd98eeba" containerName="nova-api-log" Jan 26 17:11:36 crc kubenswrapper[4754]: I0126 17:11:36.596971 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2701786-b1e9-4a86-b751-63e1dd98eeba" containerName="nova-api-log" Jan 26 17:11:36 crc kubenswrapper[4754]: I0126 17:11:36.597131 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="d2701786-b1e9-4a86-b751-63e1dd98eeba" containerName="nova-api-api" Jan 26 17:11:36 crc kubenswrapper[4754]: I0126 17:11:36.597163 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="d2701786-b1e9-4a86-b751-63e1dd98eeba" containerName="nova-api-log" Jan 26 17:11:36 crc kubenswrapper[4754]: I0126 17:11:36.598160 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Jan 26 17:11:36 crc kubenswrapper[4754]: I0126 17:11:36.601182 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Jan 26 17:11:36 crc kubenswrapper[4754]: I0126 17:11:36.601384 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Jan 26 17:11:36 crc kubenswrapper[4754]: I0126 17:11:36.601941 4754 scope.go:117] "RemoveContainer" containerID="86fe60aad1ad15e9ae3a4c78134c0e5a58adf5afaa53f605d1679062f022d2e7" Jan 26 17:11:36 crc kubenswrapper[4754]: E0126 17:11:36.602972 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"86fe60aad1ad15e9ae3a4c78134c0e5a58adf5afaa53f605d1679062f022d2e7\": container with ID starting with 86fe60aad1ad15e9ae3a4c78134c0e5a58adf5afaa53f605d1679062f022d2e7 not found: ID does not exist" containerID="86fe60aad1ad15e9ae3a4c78134c0e5a58adf5afaa53f605d1679062f022d2e7" Jan 26 17:11:36 crc kubenswrapper[4754]: I0126 17:11:36.603025 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"86fe60aad1ad15e9ae3a4c78134c0e5a58adf5afaa53f605d1679062f022d2e7"} err="failed to get container status \"86fe60aad1ad15e9ae3a4c78134c0e5a58adf5afaa53f605d1679062f022d2e7\": rpc error: code = NotFound desc = could not find container \"86fe60aad1ad15e9ae3a4c78134c0e5a58adf5afaa53f605d1679062f022d2e7\": container with ID starting with 86fe60aad1ad15e9ae3a4c78134c0e5a58adf5afaa53f605d1679062f022d2e7 not found: ID does not exist" Jan 26 17:11:36 crc kubenswrapper[4754]: I0126 17:11:36.603046 4754 scope.go:117] "RemoveContainer" containerID="3fac24e179c3d5fb3303205cbc6c04bd843a7190a169d7f411fe4d9ede9cd718" Jan 26 17:11:36 crc kubenswrapper[4754]: E0126 17:11:36.603373 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3fac24e179c3d5fb3303205cbc6c04bd843a7190a169d7f411fe4d9ede9cd718\": container with ID starting with 3fac24e179c3d5fb3303205cbc6c04bd843a7190a169d7f411fe4d9ede9cd718 not found: ID does not exist" containerID="3fac24e179c3d5fb3303205cbc6c04bd843a7190a169d7f411fe4d9ede9cd718" Jan 26 17:11:36 crc kubenswrapper[4754]: I0126 17:11:36.603420 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3fac24e179c3d5fb3303205cbc6c04bd843a7190a169d7f411fe4d9ede9cd718"} err="failed to get container status \"3fac24e179c3d5fb3303205cbc6c04bd843a7190a169d7f411fe4d9ede9cd718\": rpc error: code = NotFound desc = could not find container \"3fac24e179c3d5fb3303205cbc6c04bd843a7190a169d7f411fe4d9ede9cd718\": container with ID starting with 3fac24e179c3d5fb3303205cbc6c04bd843a7190a169d7f411fe4d9ede9cd718 not found: ID does not exist" Jan 26 17:11:36 crc kubenswrapper[4754]: I0126 17:11:36.604901 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Jan 26 17:11:36 crc kubenswrapper[4754]: I0126 17:11:36.607427 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Jan 26 17:11:36 crc kubenswrapper[4754]: I0126 17:11:36.727387 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/19c7e3a1-e40c-401b-bd01-d2a50f14daf0-public-tls-certs\") pod \"nova-api-0\" (UID: \"19c7e3a1-e40c-401b-bd01-d2a50f14daf0\") " pod="openstack/nova-api-0" Jan 26 17:11:36 crc kubenswrapper[4754]: I0126 17:11:36.727929 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19c7e3a1-e40c-401b-bd01-d2a50f14daf0-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"19c7e3a1-e40c-401b-bd01-d2a50f14daf0\") " pod="openstack/nova-api-0" Jan 26 17:11:36 crc kubenswrapper[4754]: I0126 17:11:36.728033 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/19c7e3a1-e40c-401b-bd01-d2a50f14daf0-config-data\") pod \"nova-api-0\" (UID: \"19c7e3a1-e40c-401b-bd01-d2a50f14daf0\") " pod="openstack/nova-api-0" Jan 26 17:11:36 crc kubenswrapper[4754]: I0126 17:11:36.728093 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/19c7e3a1-e40c-401b-bd01-d2a50f14daf0-logs\") pod \"nova-api-0\" (UID: \"19c7e3a1-e40c-401b-bd01-d2a50f14daf0\") " pod="openstack/nova-api-0" Jan 26 17:11:36 crc kubenswrapper[4754]: I0126 17:11:36.728740 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/19c7e3a1-e40c-401b-bd01-d2a50f14daf0-internal-tls-certs\") pod \"nova-api-0\" (UID: \"19c7e3a1-e40c-401b-bd01-d2a50f14daf0\") " pod="openstack/nova-api-0" Jan 26 17:11:36 crc kubenswrapper[4754]: I0126 17:11:36.728801 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qgkcc\" (UniqueName: \"kubernetes.io/projected/19c7e3a1-e40c-401b-bd01-d2a50f14daf0-kube-api-access-qgkcc\") pod \"nova-api-0\" (UID: \"19c7e3a1-e40c-401b-bd01-d2a50f14daf0\") " pod="openstack/nova-api-0" Jan 26 17:11:36 crc kubenswrapper[4754]: I0126 17:11:36.831296 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/19c7e3a1-e40c-401b-bd01-d2a50f14daf0-config-data\") pod \"nova-api-0\" (UID: \"19c7e3a1-e40c-401b-bd01-d2a50f14daf0\") " pod="openstack/nova-api-0" Jan 26 17:11:36 crc kubenswrapper[4754]: I0126 17:11:36.831939 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/19c7e3a1-e40c-401b-bd01-d2a50f14daf0-logs\") pod \"nova-api-0\" (UID: \"19c7e3a1-e40c-401b-bd01-d2a50f14daf0\") " pod="openstack/nova-api-0" Jan 26 17:11:36 crc kubenswrapper[4754]: I0126 17:11:36.831982 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/19c7e3a1-e40c-401b-bd01-d2a50f14daf0-internal-tls-certs\") pod \"nova-api-0\" (UID: \"19c7e3a1-e40c-401b-bd01-d2a50f14daf0\") " pod="openstack/nova-api-0" Jan 26 17:11:36 crc kubenswrapper[4754]: I0126 17:11:36.832024 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qgkcc\" (UniqueName: \"kubernetes.io/projected/19c7e3a1-e40c-401b-bd01-d2a50f14daf0-kube-api-access-qgkcc\") pod \"nova-api-0\" (UID: \"19c7e3a1-e40c-401b-bd01-d2a50f14daf0\") " pod="openstack/nova-api-0" Jan 26 17:11:36 crc kubenswrapper[4754]: I0126 17:11:36.832247 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/19c7e3a1-e40c-401b-bd01-d2a50f14daf0-public-tls-certs\") pod \"nova-api-0\" (UID: \"19c7e3a1-e40c-401b-bd01-d2a50f14daf0\") " pod="openstack/nova-api-0" Jan 26 17:11:36 crc kubenswrapper[4754]: I0126 17:11:36.832285 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19c7e3a1-e40c-401b-bd01-d2a50f14daf0-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"19c7e3a1-e40c-401b-bd01-d2a50f14daf0\") " pod="openstack/nova-api-0" Jan 26 17:11:36 crc kubenswrapper[4754]: I0126 17:11:36.832419 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/19c7e3a1-e40c-401b-bd01-d2a50f14daf0-logs\") pod \"nova-api-0\" (UID: \"19c7e3a1-e40c-401b-bd01-d2a50f14daf0\") " pod="openstack/nova-api-0" Jan 26 17:11:36 crc kubenswrapper[4754]: I0126 17:11:36.837022 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/19c7e3a1-e40c-401b-bd01-d2a50f14daf0-internal-tls-certs\") pod \"nova-api-0\" (UID: \"19c7e3a1-e40c-401b-bd01-d2a50f14daf0\") " pod="openstack/nova-api-0" Jan 26 17:11:36 crc kubenswrapper[4754]: I0126 17:11:36.839076 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/19c7e3a1-e40c-401b-bd01-d2a50f14daf0-public-tls-certs\") pod \"nova-api-0\" (UID: \"19c7e3a1-e40c-401b-bd01-d2a50f14daf0\") " pod="openstack/nova-api-0" Jan 26 17:11:36 crc kubenswrapper[4754]: I0126 17:11:36.839185 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19c7e3a1-e40c-401b-bd01-d2a50f14daf0-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"19c7e3a1-e40c-401b-bd01-d2a50f14daf0\") " pod="openstack/nova-api-0" Jan 26 17:11:36 crc kubenswrapper[4754]: I0126 17:11:36.841435 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/19c7e3a1-e40c-401b-bd01-d2a50f14daf0-config-data\") pod \"nova-api-0\" (UID: \"19c7e3a1-e40c-401b-bd01-d2a50f14daf0\") " pod="openstack/nova-api-0" Jan 26 17:11:36 crc kubenswrapper[4754]: I0126 17:11:36.853295 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qgkcc\" (UniqueName: \"kubernetes.io/projected/19c7e3a1-e40c-401b-bd01-d2a50f14daf0-kube-api-access-qgkcc\") pod \"nova-api-0\" (UID: \"19c7e3a1-e40c-401b-bd01-d2a50f14daf0\") " pod="openstack/nova-api-0" Jan 26 17:11:36 crc kubenswrapper[4754]: I0126 17:11:36.931640 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Jan 26 17:11:37 crc kubenswrapper[4754]: I0126 17:11:37.031990 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Jan 26 17:11:37 crc kubenswrapper[4754]: I0126 17:11:37.434277 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Jan 26 17:11:37 crc kubenswrapper[4754]: W0126 17:11:37.434780 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod19c7e3a1_e40c_401b_bd01_d2a50f14daf0.slice/crio-d13a156e3bdf571dfb66b2fb4994bc6066b523c0a6c088c7e31dd118c5f11a32 WatchSource:0}: Error finding container d13a156e3bdf571dfb66b2fb4994bc6066b523c0a6c088c7e31dd118c5f11a32: Status 404 returned error can't find the container with id d13a156e3bdf571dfb66b2fb4994bc6066b523c0a6c088c7e31dd118c5f11a32 Jan 26 17:11:37 crc kubenswrapper[4754]: I0126 17:11:37.523047 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b41f5e0e-1829-41f9-926c-ce15178e16c8","Type":"ContainerStarted","Data":"41dafc99f47a366f03fb8da19f8ebd637175b3586f705f62bae5896fe72438b5"} Jan 26 17:11:37 crc kubenswrapper[4754]: I0126 17:11:37.523099 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b41f5e0e-1829-41f9-926c-ce15178e16c8","Type":"ContainerStarted","Data":"b6893a66925f042e232fe76a4e6cee76ec9fa4bdfada3894930c9b8586362c4f"} Jan 26 17:11:37 crc kubenswrapper[4754]: I0126 17:11:37.524351 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"19c7e3a1-e40c-401b-bd01-d2a50f14daf0","Type":"ContainerStarted","Data":"d13a156e3bdf571dfb66b2fb4994bc6066b523c0a6c088c7e31dd118c5f11a32"} Jan 26 17:11:37 crc kubenswrapper[4754]: I0126 17:11:37.778260 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d2701786-b1e9-4a86-b751-63e1dd98eeba" path="/var/lib/kubelet/pods/d2701786-b1e9-4a86-b751-63e1dd98eeba/volumes" Jan 26 17:11:38 crc kubenswrapper[4754]: I0126 17:11:38.533604 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"19c7e3a1-e40c-401b-bd01-d2a50f14daf0","Type":"ContainerStarted","Data":"2afb5680b5deea7b4857e1eaf4bdb0cb81e95fbdaa9a07d49daff55e3720c563"} Jan 26 17:11:38 crc kubenswrapper[4754]: I0126 17:11:38.534197 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"19c7e3a1-e40c-401b-bd01-d2a50f14daf0","Type":"ContainerStarted","Data":"d6c3849288f4a670bee53c01dd166a87f4ed85c177794e0645c152549b809bd1"} Jan 26 17:11:38 crc kubenswrapper[4754]: I0126 17:11:38.570602 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.570576245 podStartE2EDuration="2.570576245s" podCreationTimestamp="2026-01-26 17:11:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 17:11:38.556150318 +0000 UTC m=+1465.080330762" watchObservedRunningTime="2026-01-26 17:11:38.570576245 +0000 UTC m=+1465.094756679" Jan 26 17:11:39 crc kubenswrapper[4754]: I0126 17:11:39.550242 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b41f5e0e-1829-41f9-926c-ce15178e16c8","Type":"ContainerStarted","Data":"2e5e3bce41ed7d42b7ffdbb4b456785914669761c7aab4ca120efb1372d00aab"} Jan 26 17:11:39 crc kubenswrapper[4754]: I0126 17:11:39.550940 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Jan 26 17:11:39 crc kubenswrapper[4754]: I0126 17:11:39.580587 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.072289435 podStartE2EDuration="5.580569169s" podCreationTimestamp="2026-01-26 17:11:34 +0000 UTC" firstStartedPulling="2026-01-26 17:11:35.367465931 +0000 UTC m=+1461.891646365" lastFinishedPulling="2026-01-26 17:11:38.875745665 +0000 UTC m=+1465.399926099" observedRunningTime="2026-01-26 17:11:39.569463293 +0000 UTC m=+1466.093643737" watchObservedRunningTime="2026-01-26 17:11:39.580569169 +0000 UTC m=+1466.104749603" Jan 26 17:11:39 crc kubenswrapper[4754]: I0126 17:11:39.836963 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-mtczv"] Jan 26 17:11:39 crc kubenswrapper[4754]: I0126 17:11:39.839499 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mtczv" Jan 26 17:11:39 crc kubenswrapper[4754]: I0126 17:11:39.866199 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-mtczv"] Jan 26 17:11:39 crc kubenswrapper[4754]: I0126 17:11:39.990706 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e904e81b-a294-404a-8899-b4c86e043347-utilities\") pod \"community-operators-mtczv\" (UID: \"e904e81b-a294-404a-8899-b4c86e043347\") " pod="openshift-marketplace/community-operators-mtczv" Jan 26 17:11:39 crc kubenswrapper[4754]: I0126 17:11:39.990753 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e904e81b-a294-404a-8899-b4c86e043347-catalog-content\") pod \"community-operators-mtczv\" (UID: \"e904e81b-a294-404a-8899-b4c86e043347\") " pod="openshift-marketplace/community-operators-mtczv" Jan 26 17:11:39 crc kubenswrapper[4754]: I0126 17:11:39.990892 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hxr6p\" (UniqueName: \"kubernetes.io/projected/e904e81b-a294-404a-8899-b4c86e043347-kube-api-access-hxr6p\") pod \"community-operators-mtczv\" (UID: \"e904e81b-a294-404a-8899-b4c86e043347\") " pod="openshift-marketplace/community-operators-mtczv" Jan 26 17:11:40 crc kubenswrapper[4754]: I0126 17:11:40.092359 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hxr6p\" (UniqueName: \"kubernetes.io/projected/e904e81b-a294-404a-8899-b4c86e043347-kube-api-access-hxr6p\") pod \"community-operators-mtczv\" (UID: \"e904e81b-a294-404a-8899-b4c86e043347\") " pod="openshift-marketplace/community-operators-mtczv" Jan 26 17:11:40 crc kubenswrapper[4754]: I0126 17:11:40.092498 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e904e81b-a294-404a-8899-b4c86e043347-utilities\") pod \"community-operators-mtczv\" (UID: \"e904e81b-a294-404a-8899-b4c86e043347\") " pod="openshift-marketplace/community-operators-mtczv" Jan 26 17:11:40 crc kubenswrapper[4754]: I0126 17:11:40.092520 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e904e81b-a294-404a-8899-b4c86e043347-catalog-content\") pod \"community-operators-mtczv\" (UID: \"e904e81b-a294-404a-8899-b4c86e043347\") " pod="openshift-marketplace/community-operators-mtczv" Jan 26 17:11:40 crc kubenswrapper[4754]: I0126 17:11:40.093107 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e904e81b-a294-404a-8899-b4c86e043347-catalog-content\") pod \"community-operators-mtczv\" (UID: \"e904e81b-a294-404a-8899-b4c86e043347\") " pod="openshift-marketplace/community-operators-mtczv" Jan 26 17:11:40 crc kubenswrapper[4754]: I0126 17:11:40.093241 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e904e81b-a294-404a-8899-b4c86e043347-utilities\") pod \"community-operators-mtczv\" (UID: \"e904e81b-a294-404a-8899-b4c86e043347\") " pod="openshift-marketplace/community-operators-mtczv" Jan 26 17:11:40 crc kubenswrapper[4754]: I0126 17:11:40.117190 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hxr6p\" (UniqueName: \"kubernetes.io/projected/e904e81b-a294-404a-8899-b4c86e043347-kube-api-access-hxr6p\") pod \"community-operators-mtczv\" (UID: \"e904e81b-a294-404a-8899-b4c86e043347\") " pod="openshift-marketplace/community-operators-mtczv" Jan 26 17:11:40 crc kubenswrapper[4754]: I0126 17:11:40.179842 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mtczv" Jan 26 17:11:40 crc kubenswrapper[4754]: I0126 17:11:40.315272 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-89c5cd4d5-8wrps" Jan 26 17:11:40 crc kubenswrapper[4754]: I0126 17:11:40.419303 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-gr868"] Jan 26 17:11:40 crc kubenswrapper[4754]: I0126 17:11:40.419579 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-757b4f8459-gr868" podUID="3dbb5ed9-e2d4-45c9-90c2-3d56a4b4a6ba" containerName="dnsmasq-dns" containerID="cri-o://b7fce4148dc1c9bff90d8acb44402b775254d76641bf9c8076c267852e207723" gracePeriod=10 Jan 26 17:11:40 crc kubenswrapper[4754]: I0126 17:11:40.821360 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-mtczv"] Jan 26 17:11:40 crc kubenswrapper[4754]: I0126 17:11:40.980623 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-757b4f8459-gr868" Jan 26 17:11:41 crc kubenswrapper[4754]: I0126 17:11:41.122183 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3dbb5ed9-e2d4-45c9-90c2-3d56a4b4a6ba-dns-swift-storage-0\") pod \"3dbb5ed9-e2d4-45c9-90c2-3d56a4b4a6ba\" (UID: \"3dbb5ed9-e2d4-45c9-90c2-3d56a4b4a6ba\") " Jan 26 17:11:41 crc kubenswrapper[4754]: I0126 17:11:41.123093 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3dbb5ed9-e2d4-45c9-90c2-3d56a4b4a6ba-ovsdbserver-nb\") pod \"3dbb5ed9-e2d4-45c9-90c2-3d56a4b4a6ba\" (UID: \"3dbb5ed9-e2d4-45c9-90c2-3d56a4b4a6ba\") " Jan 26 17:11:41 crc kubenswrapper[4754]: I0126 17:11:41.123122 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h5psp\" (UniqueName: \"kubernetes.io/projected/3dbb5ed9-e2d4-45c9-90c2-3d56a4b4a6ba-kube-api-access-h5psp\") pod \"3dbb5ed9-e2d4-45c9-90c2-3d56a4b4a6ba\" (UID: \"3dbb5ed9-e2d4-45c9-90c2-3d56a4b4a6ba\") " Jan 26 17:11:41 crc kubenswrapper[4754]: I0126 17:11:41.123171 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3dbb5ed9-e2d4-45c9-90c2-3d56a4b4a6ba-config\") pod \"3dbb5ed9-e2d4-45c9-90c2-3d56a4b4a6ba\" (UID: \"3dbb5ed9-e2d4-45c9-90c2-3d56a4b4a6ba\") " Jan 26 17:11:41 crc kubenswrapper[4754]: I0126 17:11:41.123194 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3dbb5ed9-e2d4-45c9-90c2-3d56a4b4a6ba-ovsdbserver-sb\") pod \"3dbb5ed9-e2d4-45c9-90c2-3d56a4b4a6ba\" (UID: \"3dbb5ed9-e2d4-45c9-90c2-3d56a4b4a6ba\") " Jan 26 17:11:41 crc kubenswrapper[4754]: I0126 17:11:41.123289 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3dbb5ed9-e2d4-45c9-90c2-3d56a4b4a6ba-dns-svc\") pod \"3dbb5ed9-e2d4-45c9-90c2-3d56a4b4a6ba\" (UID: \"3dbb5ed9-e2d4-45c9-90c2-3d56a4b4a6ba\") " Jan 26 17:11:41 crc kubenswrapper[4754]: I0126 17:11:41.139879 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3dbb5ed9-e2d4-45c9-90c2-3d56a4b4a6ba-kube-api-access-h5psp" (OuterVolumeSpecName: "kube-api-access-h5psp") pod "3dbb5ed9-e2d4-45c9-90c2-3d56a4b4a6ba" (UID: "3dbb5ed9-e2d4-45c9-90c2-3d56a4b4a6ba"). InnerVolumeSpecName "kube-api-access-h5psp". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:11:41 crc kubenswrapper[4754]: I0126 17:11:41.195703 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3dbb5ed9-e2d4-45c9-90c2-3d56a4b4a6ba-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "3dbb5ed9-e2d4-45c9-90c2-3d56a4b4a6ba" (UID: "3dbb5ed9-e2d4-45c9-90c2-3d56a4b4a6ba"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:11:41 crc kubenswrapper[4754]: I0126 17:11:41.225910 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h5psp\" (UniqueName: \"kubernetes.io/projected/3dbb5ed9-e2d4-45c9-90c2-3d56a4b4a6ba-kube-api-access-h5psp\") on node \"crc\" DevicePath \"\"" Jan 26 17:11:41 crc kubenswrapper[4754]: I0126 17:11:41.225946 4754 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3dbb5ed9-e2d4-45c9-90c2-3d56a4b4a6ba-dns-svc\") on node \"crc\" DevicePath \"\"" Jan 26 17:11:41 crc kubenswrapper[4754]: I0126 17:11:41.233113 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3dbb5ed9-e2d4-45c9-90c2-3d56a4b4a6ba-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "3dbb5ed9-e2d4-45c9-90c2-3d56a4b4a6ba" (UID: "3dbb5ed9-e2d4-45c9-90c2-3d56a4b4a6ba"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:11:41 crc kubenswrapper[4754]: I0126 17:11:41.234243 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3dbb5ed9-e2d4-45c9-90c2-3d56a4b4a6ba-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "3dbb5ed9-e2d4-45c9-90c2-3d56a4b4a6ba" (UID: "3dbb5ed9-e2d4-45c9-90c2-3d56a4b4a6ba"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:11:41 crc kubenswrapper[4754]: I0126 17:11:41.242207 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3dbb5ed9-e2d4-45c9-90c2-3d56a4b4a6ba-config" (OuterVolumeSpecName: "config") pod "3dbb5ed9-e2d4-45c9-90c2-3d56a4b4a6ba" (UID: "3dbb5ed9-e2d4-45c9-90c2-3d56a4b4a6ba"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:11:41 crc kubenswrapper[4754]: I0126 17:11:41.255564 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3dbb5ed9-e2d4-45c9-90c2-3d56a4b4a6ba-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "3dbb5ed9-e2d4-45c9-90c2-3d56a4b4a6ba" (UID: "3dbb5ed9-e2d4-45c9-90c2-3d56a4b4a6ba"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:11:41 crc kubenswrapper[4754]: I0126 17:11:41.327227 4754 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3dbb5ed9-e2d4-45c9-90c2-3d56a4b4a6ba-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Jan 26 17:11:41 crc kubenswrapper[4754]: I0126 17:11:41.327272 4754 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3dbb5ed9-e2d4-45c9-90c2-3d56a4b4a6ba-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Jan 26 17:11:41 crc kubenswrapper[4754]: I0126 17:11:41.327288 4754 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3dbb5ed9-e2d4-45c9-90c2-3d56a4b4a6ba-config\") on node \"crc\" DevicePath \"\"" Jan 26 17:11:41 crc kubenswrapper[4754]: I0126 17:11:41.327299 4754 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3dbb5ed9-e2d4-45c9-90c2-3d56a4b4a6ba-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Jan 26 17:11:41 crc kubenswrapper[4754]: I0126 17:11:41.571709 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-757b4f8459-gr868" Jan 26 17:11:41 crc kubenswrapper[4754]: I0126 17:11:41.571709 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-757b4f8459-gr868" event={"ID":"3dbb5ed9-e2d4-45c9-90c2-3d56a4b4a6ba","Type":"ContainerDied","Data":"b7fce4148dc1c9bff90d8acb44402b775254d76641bf9c8076c267852e207723"} Jan 26 17:11:41 crc kubenswrapper[4754]: I0126 17:11:41.572829 4754 scope.go:117] "RemoveContainer" containerID="b7fce4148dc1c9bff90d8acb44402b775254d76641bf9c8076c267852e207723" Jan 26 17:11:41 crc kubenswrapper[4754]: I0126 17:11:41.571649 4754 generic.go:334] "Generic (PLEG): container finished" podID="3dbb5ed9-e2d4-45c9-90c2-3d56a4b4a6ba" containerID="b7fce4148dc1c9bff90d8acb44402b775254d76641bf9c8076c267852e207723" exitCode=0 Jan 26 17:11:41 crc kubenswrapper[4754]: I0126 17:11:41.573007 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-757b4f8459-gr868" event={"ID":"3dbb5ed9-e2d4-45c9-90c2-3d56a4b4a6ba","Type":"ContainerDied","Data":"0b11e0d40d1a2ab4a954951a3251aa13984038b682c61e4503e1a16ee5d2c1e9"} Jan 26 17:11:41 crc kubenswrapper[4754]: I0126 17:11:41.605034 4754 generic.go:334] "Generic (PLEG): container finished" podID="e904e81b-a294-404a-8899-b4c86e043347" containerID="3d7d3c37c0853417594c89f5a5c074cdbc45350775c240b0c099690aa3ef9b4a" exitCode=0 Jan 26 17:11:41 crc kubenswrapper[4754]: I0126 17:11:41.605288 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mtczv" event={"ID":"e904e81b-a294-404a-8899-b4c86e043347","Type":"ContainerDied","Data":"3d7d3c37c0853417594c89f5a5c074cdbc45350775c240b0c099690aa3ef9b4a"} Jan 26 17:11:41 crc kubenswrapper[4754]: I0126 17:11:41.605378 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mtczv" event={"ID":"e904e81b-a294-404a-8899-b4c86e043347","Type":"ContainerStarted","Data":"45d63869b71a35a067613fd98630f6afc9684c25730f6a1111e4c4c7d801ac36"} Jan 26 17:11:41 crc kubenswrapper[4754]: I0126 17:11:41.643375 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-gr868"] Jan 26 17:11:41 crc kubenswrapper[4754]: I0126 17:11:41.676735 4754 scope.go:117] "RemoveContainer" containerID="ef702a387578e08af58300cf968ed63cc88a4db4cfedf1c3f865d2be42ec5add" Jan 26 17:11:41 crc kubenswrapper[4754]: I0126 17:11:41.680993 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-gr868"] Jan 26 17:11:41 crc kubenswrapper[4754]: I0126 17:11:41.709161 4754 scope.go:117] "RemoveContainer" containerID="b7fce4148dc1c9bff90d8acb44402b775254d76641bf9c8076c267852e207723" Jan 26 17:11:41 crc kubenswrapper[4754]: E0126 17:11:41.710052 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b7fce4148dc1c9bff90d8acb44402b775254d76641bf9c8076c267852e207723\": container with ID starting with b7fce4148dc1c9bff90d8acb44402b775254d76641bf9c8076c267852e207723 not found: ID does not exist" containerID="b7fce4148dc1c9bff90d8acb44402b775254d76641bf9c8076c267852e207723" Jan 26 17:11:41 crc kubenswrapper[4754]: I0126 17:11:41.710120 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b7fce4148dc1c9bff90d8acb44402b775254d76641bf9c8076c267852e207723"} err="failed to get container status \"b7fce4148dc1c9bff90d8acb44402b775254d76641bf9c8076c267852e207723\": rpc error: code = NotFound desc = could not find container \"b7fce4148dc1c9bff90d8acb44402b775254d76641bf9c8076c267852e207723\": container with ID starting with b7fce4148dc1c9bff90d8acb44402b775254d76641bf9c8076c267852e207723 not found: ID does not exist" Jan 26 17:11:41 crc kubenswrapper[4754]: I0126 17:11:41.710148 4754 scope.go:117] "RemoveContainer" containerID="ef702a387578e08af58300cf968ed63cc88a4db4cfedf1c3f865d2be42ec5add" Jan 26 17:11:41 crc kubenswrapper[4754]: E0126 17:11:41.710624 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ef702a387578e08af58300cf968ed63cc88a4db4cfedf1c3f865d2be42ec5add\": container with ID starting with ef702a387578e08af58300cf968ed63cc88a4db4cfedf1c3f865d2be42ec5add not found: ID does not exist" containerID="ef702a387578e08af58300cf968ed63cc88a4db4cfedf1c3f865d2be42ec5add" Jan 26 17:11:41 crc kubenswrapper[4754]: I0126 17:11:41.710643 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ef702a387578e08af58300cf968ed63cc88a4db4cfedf1c3f865d2be42ec5add"} err="failed to get container status \"ef702a387578e08af58300cf968ed63cc88a4db4cfedf1c3f865d2be42ec5add\": rpc error: code = NotFound desc = could not find container \"ef702a387578e08af58300cf968ed63cc88a4db4cfedf1c3f865d2be42ec5add\": container with ID starting with ef702a387578e08af58300cf968ed63cc88a4db4cfedf1c3f865d2be42ec5add not found: ID does not exist" Jan 26 17:11:41 crc kubenswrapper[4754]: I0126 17:11:41.778303 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3dbb5ed9-e2d4-45c9-90c2-3d56a4b4a6ba" path="/var/lib/kubelet/pods/3dbb5ed9-e2d4-45c9-90c2-3d56a4b4a6ba/volumes" Jan 26 17:11:42 crc kubenswrapper[4754]: I0126 17:11:42.032485 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Jan 26 17:11:42 crc kubenswrapper[4754]: I0126 17:11:42.054413 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Jan 26 17:11:42 crc kubenswrapper[4754]: I0126 17:11:42.617192 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mtczv" event={"ID":"e904e81b-a294-404a-8899-b4c86e043347","Type":"ContainerStarted","Data":"74569f9b27b99ea46c4b3e679ef1384fac663804f457351c3f1c504c5e850b90"} Jan 26 17:11:42 crc kubenswrapper[4754]: I0126 17:11:42.633724 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Jan 26 17:11:42 crc kubenswrapper[4754]: I0126 17:11:42.812332 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-bfp9x"] Jan 26 17:11:42 crc kubenswrapper[4754]: E0126 17:11:42.812846 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3dbb5ed9-e2d4-45c9-90c2-3d56a4b4a6ba" containerName="dnsmasq-dns" Jan 26 17:11:42 crc kubenswrapper[4754]: I0126 17:11:42.812870 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="3dbb5ed9-e2d4-45c9-90c2-3d56a4b4a6ba" containerName="dnsmasq-dns" Jan 26 17:11:42 crc kubenswrapper[4754]: E0126 17:11:42.812888 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3dbb5ed9-e2d4-45c9-90c2-3d56a4b4a6ba" containerName="init" Jan 26 17:11:42 crc kubenswrapper[4754]: I0126 17:11:42.812896 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="3dbb5ed9-e2d4-45c9-90c2-3d56a4b4a6ba" containerName="init" Jan 26 17:11:42 crc kubenswrapper[4754]: I0126 17:11:42.813133 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="3dbb5ed9-e2d4-45c9-90c2-3d56a4b4a6ba" containerName="dnsmasq-dns" Jan 26 17:11:42 crc kubenswrapper[4754]: I0126 17:11:42.813844 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-bfp9x" Jan 26 17:11:42 crc kubenswrapper[4754]: I0126 17:11:42.817879 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Jan 26 17:11:42 crc kubenswrapper[4754]: I0126 17:11:42.818376 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Jan 26 17:11:42 crc kubenswrapper[4754]: I0126 17:11:42.823000 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-bfp9x"] Jan 26 17:11:42 crc kubenswrapper[4754]: I0126 17:11:42.854564 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f22d2aa-d1fd-4a87-b0bd-9e3fb661d254-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-bfp9x\" (UID: \"7f22d2aa-d1fd-4a87-b0bd-9e3fb661d254\") " pod="openstack/nova-cell1-cell-mapping-bfp9x" Jan 26 17:11:42 crc kubenswrapper[4754]: I0126 17:11:42.854704 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7f22d2aa-d1fd-4a87-b0bd-9e3fb661d254-scripts\") pod \"nova-cell1-cell-mapping-bfp9x\" (UID: \"7f22d2aa-d1fd-4a87-b0bd-9e3fb661d254\") " pod="openstack/nova-cell1-cell-mapping-bfp9x" Jan 26 17:11:42 crc kubenswrapper[4754]: I0126 17:11:42.854751 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7f22d2aa-d1fd-4a87-b0bd-9e3fb661d254-config-data\") pod \"nova-cell1-cell-mapping-bfp9x\" (UID: \"7f22d2aa-d1fd-4a87-b0bd-9e3fb661d254\") " pod="openstack/nova-cell1-cell-mapping-bfp9x" Jan 26 17:11:42 crc kubenswrapper[4754]: I0126 17:11:42.854878 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9d8lf\" (UniqueName: \"kubernetes.io/projected/7f22d2aa-d1fd-4a87-b0bd-9e3fb661d254-kube-api-access-9d8lf\") pod \"nova-cell1-cell-mapping-bfp9x\" (UID: \"7f22d2aa-d1fd-4a87-b0bd-9e3fb661d254\") " pod="openstack/nova-cell1-cell-mapping-bfp9x" Jan 26 17:11:42 crc kubenswrapper[4754]: I0126 17:11:42.956031 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9d8lf\" (UniqueName: \"kubernetes.io/projected/7f22d2aa-d1fd-4a87-b0bd-9e3fb661d254-kube-api-access-9d8lf\") pod \"nova-cell1-cell-mapping-bfp9x\" (UID: \"7f22d2aa-d1fd-4a87-b0bd-9e3fb661d254\") " pod="openstack/nova-cell1-cell-mapping-bfp9x" Jan 26 17:11:42 crc kubenswrapper[4754]: I0126 17:11:42.956100 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f22d2aa-d1fd-4a87-b0bd-9e3fb661d254-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-bfp9x\" (UID: \"7f22d2aa-d1fd-4a87-b0bd-9e3fb661d254\") " pod="openstack/nova-cell1-cell-mapping-bfp9x" Jan 26 17:11:42 crc kubenswrapper[4754]: I0126 17:11:42.956178 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7f22d2aa-d1fd-4a87-b0bd-9e3fb661d254-scripts\") pod \"nova-cell1-cell-mapping-bfp9x\" (UID: \"7f22d2aa-d1fd-4a87-b0bd-9e3fb661d254\") " pod="openstack/nova-cell1-cell-mapping-bfp9x" Jan 26 17:11:42 crc kubenswrapper[4754]: I0126 17:11:42.956206 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7f22d2aa-d1fd-4a87-b0bd-9e3fb661d254-config-data\") pod \"nova-cell1-cell-mapping-bfp9x\" (UID: \"7f22d2aa-d1fd-4a87-b0bd-9e3fb661d254\") " pod="openstack/nova-cell1-cell-mapping-bfp9x" Jan 26 17:11:42 crc kubenswrapper[4754]: I0126 17:11:42.962361 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7f22d2aa-d1fd-4a87-b0bd-9e3fb661d254-scripts\") pod \"nova-cell1-cell-mapping-bfp9x\" (UID: \"7f22d2aa-d1fd-4a87-b0bd-9e3fb661d254\") " pod="openstack/nova-cell1-cell-mapping-bfp9x" Jan 26 17:11:42 crc kubenswrapper[4754]: I0126 17:11:42.962361 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f22d2aa-d1fd-4a87-b0bd-9e3fb661d254-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-bfp9x\" (UID: \"7f22d2aa-d1fd-4a87-b0bd-9e3fb661d254\") " pod="openstack/nova-cell1-cell-mapping-bfp9x" Jan 26 17:11:42 crc kubenswrapper[4754]: I0126 17:11:42.962607 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7f22d2aa-d1fd-4a87-b0bd-9e3fb661d254-config-data\") pod \"nova-cell1-cell-mapping-bfp9x\" (UID: \"7f22d2aa-d1fd-4a87-b0bd-9e3fb661d254\") " pod="openstack/nova-cell1-cell-mapping-bfp9x" Jan 26 17:11:42 crc kubenswrapper[4754]: I0126 17:11:42.985097 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9d8lf\" (UniqueName: \"kubernetes.io/projected/7f22d2aa-d1fd-4a87-b0bd-9e3fb661d254-kube-api-access-9d8lf\") pod \"nova-cell1-cell-mapping-bfp9x\" (UID: \"7f22d2aa-d1fd-4a87-b0bd-9e3fb661d254\") " pod="openstack/nova-cell1-cell-mapping-bfp9x" Jan 26 17:11:43 crc kubenswrapper[4754]: I0126 17:11:43.128984 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-bfp9x" Jan 26 17:11:43 crc kubenswrapper[4754]: W0126 17:11:43.601999 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7f22d2aa_d1fd_4a87_b0bd_9e3fb661d254.slice/crio-c15d095cc951805c41419891c5f22df32be120c964fd151c4872035fd4116869 WatchSource:0}: Error finding container c15d095cc951805c41419891c5f22df32be120c964fd151c4872035fd4116869: Status 404 returned error can't find the container with id c15d095cc951805c41419891c5f22df32be120c964fd151c4872035fd4116869 Jan 26 17:11:43 crc kubenswrapper[4754]: I0126 17:11:43.609410 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-bfp9x"] Jan 26 17:11:43 crc kubenswrapper[4754]: I0126 17:11:43.632845 4754 generic.go:334] "Generic (PLEG): container finished" podID="e904e81b-a294-404a-8899-b4c86e043347" containerID="74569f9b27b99ea46c4b3e679ef1384fac663804f457351c3f1c504c5e850b90" exitCode=0 Jan 26 17:11:43 crc kubenswrapper[4754]: I0126 17:11:43.632922 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mtczv" event={"ID":"e904e81b-a294-404a-8899-b4c86e043347","Type":"ContainerDied","Data":"74569f9b27b99ea46c4b3e679ef1384fac663804f457351c3f1c504c5e850b90"} Jan 26 17:11:43 crc kubenswrapper[4754]: I0126 17:11:43.640783 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-bfp9x" event={"ID":"7f22d2aa-d1fd-4a87-b0bd-9e3fb661d254","Type":"ContainerStarted","Data":"c15d095cc951805c41419891c5f22df32be120c964fd151c4872035fd4116869"} Jan 26 17:11:44 crc kubenswrapper[4754]: I0126 17:11:44.657239 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-bfp9x" event={"ID":"7f22d2aa-d1fd-4a87-b0bd-9e3fb661d254","Type":"ContainerStarted","Data":"1dc2a31a2a2541e5978333a908c08012c296165c761ff33581b1b2a9ea38bb40"} Jan 26 17:11:44 crc kubenswrapper[4754]: I0126 17:11:44.662255 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mtczv" event={"ID":"e904e81b-a294-404a-8899-b4c86e043347","Type":"ContainerStarted","Data":"37a0be58857466d6f638ee9509d8e6884ba390e469cd941e9b8b451309666fd2"} Jan 26 17:11:44 crc kubenswrapper[4754]: I0126 17:11:44.681584 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-bfp9x" podStartSLOduration=2.681563717 podStartE2EDuration="2.681563717s" podCreationTimestamp="2026-01-26 17:11:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 17:11:44.675447548 +0000 UTC m=+1471.199627982" watchObservedRunningTime="2026-01-26 17:11:44.681563717 +0000 UTC m=+1471.205744151" Jan 26 17:11:44 crc kubenswrapper[4754]: I0126 17:11:44.704100 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-mtczv" podStartSLOduration=3.244908965 podStartE2EDuration="5.704079797s" podCreationTimestamp="2026-01-26 17:11:39 +0000 UTC" firstStartedPulling="2026-01-26 17:11:41.608576179 +0000 UTC m=+1468.132756613" lastFinishedPulling="2026-01-26 17:11:44.067747011 +0000 UTC m=+1470.591927445" observedRunningTime="2026-01-26 17:11:44.696933981 +0000 UTC m=+1471.221114415" watchObservedRunningTime="2026-01-26 17:11:44.704079797 +0000 UTC m=+1471.228260231" Jan 26 17:11:46 crc kubenswrapper[4754]: I0126 17:11:46.933726 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Jan 26 17:11:46 crc kubenswrapper[4754]: I0126 17:11:46.934285 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Jan 26 17:11:47 crc kubenswrapper[4754]: I0126 17:11:47.941845 4754 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="19c7e3a1-e40c-401b-bd01-d2a50f14daf0" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.209:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Jan 26 17:11:47 crc kubenswrapper[4754]: I0126 17:11:47.941844 4754 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="19c7e3a1-e40c-401b-bd01-d2a50f14daf0" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.209:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Jan 26 17:11:49 crc kubenswrapper[4754]: I0126 17:11:49.705231 4754 generic.go:334] "Generic (PLEG): container finished" podID="7f22d2aa-d1fd-4a87-b0bd-9e3fb661d254" containerID="1dc2a31a2a2541e5978333a908c08012c296165c761ff33581b1b2a9ea38bb40" exitCode=0 Jan 26 17:11:49 crc kubenswrapper[4754]: I0126 17:11:49.705511 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-bfp9x" event={"ID":"7f22d2aa-d1fd-4a87-b0bd-9e3fb661d254","Type":"ContainerDied","Data":"1dc2a31a2a2541e5978333a908c08012c296165c761ff33581b1b2a9ea38bb40"} Jan 26 17:11:50 crc kubenswrapper[4754]: I0126 17:11:50.180179 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-mtczv" Jan 26 17:11:50 crc kubenswrapper[4754]: I0126 17:11:50.180261 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-mtczv" Jan 26 17:11:50 crc kubenswrapper[4754]: I0126 17:11:50.226871 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-mtczv" Jan 26 17:11:50 crc kubenswrapper[4754]: I0126 17:11:50.771069 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-mtczv" Jan 26 17:11:50 crc kubenswrapper[4754]: I0126 17:11:50.821474 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-mtczv"] Jan 26 17:11:51 crc kubenswrapper[4754]: I0126 17:11:51.068057 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-bfp9x" Jan 26 17:11:51 crc kubenswrapper[4754]: I0126 17:11:51.220007 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f22d2aa-d1fd-4a87-b0bd-9e3fb661d254-combined-ca-bundle\") pod \"7f22d2aa-d1fd-4a87-b0bd-9e3fb661d254\" (UID: \"7f22d2aa-d1fd-4a87-b0bd-9e3fb661d254\") " Jan 26 17:11:51 crc kubenswrapper[4754]: I0126 17:11:51.220117 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7f22d2aa-d1fd-4a87-b0bd-9e3fb661d254-config-data\") pod \"7f22d2aa-d1fd-4a87-b0bd-9e3fb661d254\" (UID: \"7f22d2aa-d1fd-4a87-b0bd-9e3fb661d254\") " Jan 26 17:11:51 crc kubenswrapper[4754]: I0126 17:11:51.220183 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7f22d2aa-d1fd-4a87-b0bd-9e3fb661d254-scripts\") pod \"7f22d2aa-d1fd-4a87-b0bd-9e3fb661d254\" (UID: \"7f22d2aa-d1fd-4a87-b0bd-9e3fb661d254\") " Jan 26 17:11:51 crc kubenswrapper[4754]: I0126 17:11:51.220215 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9d8lf\" (UniqueName: \"kubernetes.io/projected/7f22d2aa-d1fd-4a87-b0bd-9e3fb661d254-kube-api-access-9d8lf\") pod \"7f22d2aa-d1fd-4a87-b0bd-9e3fb661d254\" (UID: \"7f22d2aa-d1fd-4a87-b0bd-9e3fb661d254\") " Jan 26 17:11:51 crc kubenswrapper[4754]: I0126 17:11:51.227579 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7f22d2aa-d1fd-4a87-b0bd-9e3fb661d254-scripts" (OuterVolumeSpecName: "scripts") pod "7f22d2aa-d1fd-4a87-b0bd-9e3fb661d254" (UID: "7f22d2aa-d1fd-4a87-b0bd-9e3fb661d254"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:11:51 crc kubenswrapper[4754]: I0126 17:11:51.227646 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7f22d2aa-d1fd-4a87-b0bd-9e3fb661d254-kube-api-access-9d8lf" (OuterVolumeSpecName: "kube-api-access-9d8lf") pod "7f22d2aa-d1fd-4a87-b0bd-9e3fb661d254" (UID: "7f22d2aa-d1fd-4a87-b0bd-9e3fb661d254"). InnerVolumeSpecName "kube-api-access-9d8lf". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:11:51 crc kubenswrapper[4754]: I0126 17:11:51.255684 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7f22d2aa-d1fd-4a87-b0bd-9e3fb661d254-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7f22d2aa-d1fd-4a87-b0bd-9e3fb661d254" (UID: "7f22d2aa-d1fd-4a87-b0bd-9e3fb661d254"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:11:51 crc kubenswrapper[4754]: I0126 17:11:51.255806 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7f22d2aa-d1fd-4a87-b0bd-9e3fb661d254-config-data" (OuterVolumeSpecName: "config-data") pod "7f22d2aa-d1fd-4a87-b0bd-9e3fb661d254" (UID: "7f22d2aa-d1fd-4a87-b0bd-9e3fb661d254"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:11:51 crc kubenswrapper[4754]: I0126 17:11:51.322128 4754 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f22d2aa-d1fd-4a87-b0bd-9e3fb661d254-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 26 17:11:51 crc kubenswrapper[4754]: I0126 17:11:51.322159 4754 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7f22d2aa-d1fd-4a87-b0bd-9e3fb661d254-config-data\") on node \"crc\" DevicePath \"\"" Jan 26 17:11:51 crc kubenswrapper[4754]: I0126 17:11:51.322170 4754 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7f22d2aa-d1fd-4a87-b0bd-9e3fb661d254-scripts\") on node \"crc\" DevicePath \"\"" Jan 26 17:11:51 crc kubenswrapper[4754]: I0126 17:11:51.322179 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9d8lf\" (UniqueName: \"kubernetes.io/projected/7f22d2aa-d1fd-4a87-b0bd-9e3fb661d254-kube-api-access-9d8lf\") on node \"crc\" DevicePath \"\"" Jan 26 17:11:51 crc kubenswrapper[4754]: I0126 17:11:51.723911 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-bfp9x" event={"ID":"7f22d2aa-d1fd-4a87-b0bd-9e3fb661d254","Type":"ContainerDied","Data":"c15d095cc951805c41419891c5f22df32be120c964fd151c4872035fd4116869"} Jan 26 17:11:51 crc kubenswrapper[4754]: I0126 17:11:51.723952 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-bfp9x" Jan 26 17:11:51 crc kubenswrapper[4754]: I0126 17:11:51.723952 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c15d095cc951805c41419891c5f22df32be120c964fd151c4872035fd4116869" Jan 26 17:11:51 crc kubenswrapper[4754]: I0126 17:11:51.904859 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Jan 26 17:11:51 crc kubenswrapper[4754]: I0126 17:11:51.905063 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="ef3b4399-44ec-476a-a0d5-e727b76c624e" containerName="nova-scheduler-scheduler" containerID="cri-o://5f0b0d2d8d721b498cb7253818255e0158226a13e1da30cdda99a766305b2734" gracePeriod=30 Jan 26 17:11:51 crc kubenswrapper[4754]: I0126 17:11:51.912995 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Jan 26 17:11:51 crc kubenswrapper[4754]: I0126 17:11:51.913211 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="19c7e3a1-e40c-401b-bd01-d2a50f14daf0" containerName="nova-api-log" containerID="cri-o://d6c3849288f4a670bee53c01dd166a87f4ed85c177794e0645c152549b809bd1" gracePeriod=30 Jan 26 17:11:51 crc kubenswrapper[4754]: I0126 17:11:51.913441 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="19c7e3a1-e40c-401b-bd01-d2a50f14daf0" containerName="nova-api-api" containerID="cri-o://2afb5680b5deea7b4857e1eaf4bdb0cb81e95fbdaa9a07d49daff55e3720c563" gracePeriod=30 Jan 26 17:11:51 crc kubenswrapper[4754]: I0126 17:11:51.935421 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Jan 26 17:11:51 crc kubenswrapper[4754]: I0126 17:11:51.935711 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="37f5b6db-8ef3-4e4f-9763-b150188c1466" containerName="nova-metadata-log" containerID="cri-o://911ce027833e9fd70123c130dde99ab0e73b4f39eb0c6df15227bfe3a7a4e4cf" gracePeriod=30 Jan 26 17:11:51 crc kubenswrapper[4754]: I0126 17:11:51.935855 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="37f5b6db-8ef3-4e4f-9763-b150188c1466" containerName="nova-metadata-metadata" containerID="cri-o://b8bcc64a29d1de162bd574631509ab962e70a5eba797922712a474b28c35c163" gracePeriod=30 Jan 26 17:11:52 crc kubenswrapper[4754]: I0126 17:11:52.736966 4754 generic.go:334] "Generic (PLEG): container finished" podID="37f5b6db-8ef3-4e4f-9763-b150188c1466" containerID="911ce027833e9fd70123c130dde99ab0e73b4f39eb0c6df15227bfe3a7a4e4cf" exitCode=143 Jan 26 17:11:52 crc kubenswrapper[4754]: I0126 17:11:52.737065 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"37f5b6db-8ef3-4e4f-9763-b150188c1466","Type":"ContainerDied","Data":"911ce027833e9fd70123c130dde99ab0e73b4f39eb0c6df15227bfe3a7a4e4cf"} Jan 26 17:11:52 crc kubenswrapper[4754]: I0126 17:11:52.739840 4754 generic.go:334] "Generic (PLEG): container finished" podID="19c7e3a1-e40c-401b-bd01-d2a50f14daf0" containerID="d6c3849288f4a670bee53c01dd166a87f4ed85c177794e0645c152549b809bd1" exitCode=143 Jan 26 17:11:52 crc kubenswrapper[4754]: I0126 17:11:52.739940 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"19c7e3a1-e40c-401b-bd01-d2a50f14daf0","Type":"ContainerDied","Data":"d6c3849288f4a670bee53c01dd166a87f4ed85c177794e0645c152549b809bd1"} Jan 26 17:11:52 crc kubenswrapper[4754]: I0126 17:11:52.740187 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-mtczv" podUID="e904e81b-a294-404a-8899-b4c86e043347" containerName="registry-server" containerID="cri-o://37a0be58857466d6f638ee9509d8e6884ba390e469cd941e9b8b451309666fd2" gracePeriod=2 Jan 26 17:11:53 crc kubenswrapper[4754]: I0126 17:11:53.219758 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mtczv" Jan 26 17:11:53 crc kubenswrapper[4754]: I0126 17:11:53.368614 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hxr6p\" (UniqueName: \"kubernetes.io/projected/e904e81b-a294-404a-8899-b4c86e043347-kube-api-access-hxr6p\") pod \"e904e81b-a294-404a-8899-b4c86e043347\" (UID: \"e904e81b-a294-404a-8899-b4c86e043347\") " Jan 26 17:11:53 crc kubenswrapper[4754]: I0126 17:11:53.368834 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e904e81b-a294-404a-8899-b4c86e043347-utilities\") pod \"e904e81b-a294-404a-8899-b4c86e043347\" (UID: \"e904e81b-a294-404a-8899-b4c86e043347\") " Jan 26 17:11:53 crc kubenswrapper[4754]: I0126 17:11:53.368866 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e904e81b-a294-404a-8899-b4c86e043347-catalog-content\") pod \"e904e81b-a294-404a-8899-b4c86e043347\" (UID: \"e904e81b-a294-404a-8899-b4c86e043347\") " Jan 26 17:11:53 crc kubenswrapper[4754]: I0126 17:11:53.369707 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e904e81b-a294-404a-8899-b4c86e043347-utilities" (OuterVolumeSpecName: "utilities") pod "e904e81b-a294-404a-8899-b4c86e043347" (UID: "e904e81b-a294-404a-8899-b4c86e043347"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:11:53 crc kubenswrapper[4754]: I0126 17:11:53.373386 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e904e81b-a294-404a-8899-b4c86e043347-kube-api-access-hxr6p" (OuterVolumeSpecName: "kube-api-access-hxr6p") pod "e904e81b-a294-404a-8899-b4c86e043347" (UID: "e904e81b-a294-404a-8899-b4c86e043347"). InnerVolumeSpecName "kube-api-access-hxr6p". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:11:53 crc kubenswrapper[4754]: I0126 17:11:53.423217 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e904e81b-a294-404a-8899-b4c86e043347-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e904e81b-a294-404a-8899-b4c86e043347" (UID: "e904e81b-a294-404a-8899-b4c86e043347"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:11:53 crc kubenswrapper[4754]: I0126 17:11:53.470892 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hxr6p\" (UniqueName: \"kubernetes.io/projected/e904e81b-a294-404a-8899-b4c86e043347-kube-api-access-hxr6p\") on node \"crc\" DevicePath \"\"" Jan 26 17:11:53 crc kubenswrapper[4754]: I0126 17:11:53.470925 4754 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e904e81b-a294-404a-8899-b4c86e043347-utilities\") on node \"crc\" DevicePath \"\"" Jan 26 17:11:53 crc kubenswrapper[4754]: I0126 17:11:53.470935 4754 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e904e81b-a294-404a-8899-b4c86e043347-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 26 17:11:53 crc kubenswrapper[4754]: I0126 17:11:53.771125 4754 generic.go:334] "Generic (PLEG): container finished" podID="e904e81b-a294-404a-8899-b4c86e043347" containerID="37a0be58857466d6f638ee9509d8e6884ba390e469cd941e9b8b451309666fd2" exitCode=0 Jan 26 17:11:53 crc kubenswrapper[4754]: I0126 17:11:53.784309 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mtczv" Jan 26 17:11:53 crc kubenswrapper[4754]: I0126 17:11:53.786984 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mtczv" event={"ID":"e904e81b-a294-404a-8899-b4c86e043347","Type":"ContainerDied","Data":"37a0be58857466d6f638ee9509d8e6884ba390e469cd941e9b8b451309666fd2"} Jan 26 17:11:53 crc kubenswrapper[4754]: I0126 17:11:53.787031 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mtczv" event={"ID":"e904e81b-a294-404a-8899-b4c86e043347","Type":"ContainerDied","Data":"45d63869b71a35a067613fd98630f6afc9684c25730f6a1111e4c4c7d801ac36"} Jan 26 17:11:53 crc kubenswrapper[4754]: I0126 17:11:53.787054 4754 scope.go:117] "RemoveContainer" containerID="37a0be58857466d6f638ee9509d8e6884ba390e469cd941e9b8b451309666fd2" Jan 26 17:11:53 crc kubenswrapper[4754]: I0126 17:11:53.830453 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-mtczv"] Jan 26 17:11:53 crc kubenswrapper[4754]: I0126 17:11:53.831272 4754 scope.go:117] "RemoveContainer" containerID="74569f9b27b99ea46c4b3e679ef1384fac663804f457351c3f1c504c5e850b90" Jan 26 17:11:53 crc kubenswrapper[4754]: I0126 17:11:53.841145 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-mtczv"] Jan 26 17:11:53 crc kubenswrapper[4754]: I0126 17:11:53.865383 4754 scope.go:117] "RemoveContainer" containerID="3d7d3c37c0853417594c89f5a5c074cdbc45350775c240b0c099690aa3ef9b4a" Jan 26 17:11:53 crc kubenswrapper[4754]: I0126 17:11:53.906135 4754 scope.go:117] "RemoveContainer" containerID="37a0be58857466d6f638ee9509d8e6884ba390e469cd941e9b8b451309666fd2" Jan 26 17:11:53 crc kubenswrapper[4754]: E0126 17:11:53.906751 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"37a0be58857466d6f638ee9509d8e6884ba390e469cd941e9b8b451309666fd2\": container with ID starting with 37a0be58857466d6f638ee9509d8e6884ba390e469cd941e9b8b451309666fd2 not found: ID does not exist" containerID="37a0be58857466d6f638ee9509d8e6884ba390e469cd941e9b8b451309666fd2" Jan 26 17:11:53 crc kubenswrapper[4754]: I0126 17:11:53.906817 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"37a0be58857466d6f638ee9509d8e6884ba390e469cd941e9b8b451309666fd2"} err="failed to get container status \"37a0be58857466d6f638ee9509d8e6884ba390e469cd941e9b8b451309666fd2\": rpc error: code = NotFound desc = could not find container \"37a0be58857466d6f638ee9509d8e6884ba390e469cd941e9b8b451309666fd2\": container with ID starting with 37a0be58857466d6f638ee9509d8e6884ba390e469cd941e9b8b451309666fd2 not found: ID does not exist" Jan 26 17:11:53 crc kubenswrapper[4754]: I0126 17:11:53.906846 4754 scope.go:117] "RemoveContainer" containerID="74569f9b27b99ea46c4b3e679ef1384fac663804f457351c3f1c504c5e850b90" Jan 26 17:11:53 crc kubenswrapper[4754]: E0126 17:11:53.908321 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"74569f9b27b99ea46c4b3e679ef1384fac663804f457351c3f1c504c5e850b90\": container with ID starting with 74569f9b27b99ea46c4b3e679ef1384fac663804f457351c3f1c504c5e850b90 not found: ID does not exist" containerID="74569f9b27b99ea46c4b3e679ef1384fac663804f457351c3f1c504c5e850b90" Jan 26 17:11:53 crc kubenswrapper[4754]: I0126 17:11:53.908380 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"74569f9b27b99ea46c4b3e679ef1384fac663804f457351c3f1c504c5e850b90"} err="failed to get container status \"74569f9b27b99ea46c4b3e679ef1384fac663804f457351c3f1c504c5e850b90\": rpc error: code = NotFound desc = could not find container \"74569f9b27b99ea46c4b3e679ef1384fac663804f457351c3f1c504c5e850b90\": container with ID starting with 74569f9b27b99ea46c4b3e679ef1384fac663804f457351c3f1c504c5e850b90 not found: ID does not exist" Jan 26 17:11:53 crc kubenswrapper[4754]: I0126 17:11:53.908413 4754 scope.go:117] "RemoveContainer" containerID="3d7d3c37c0853417594c89f5a5c074cdbc45350775c240b0c099690aa3ef9b4a" Jan 26 17:11:53 crc kubenswrapper[4754]: E0126 17:11:53.908802 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3d7d3c37c0853417594c89f5a5c074cdbc45350775c240b0c099690aa3ef9b4a\": container with ID starting with 3d7d3c37c0853417594c89f5a5c074cdbc45350775c240b0c099690aa3ef9b4a not found: ID does not exist" containerID="3d7d3c37c0853417594c89f5a5c074cdbc45350775c240b0c099690aa3ef9b4a" Jan 26 17:11:53 crc kubenswrapper[4754]: I0126 17:11:53.908851 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3d7d3c37c0853417594c89f5a5c074cdbc45350775c240b0c099690aa3ef9b4a"} err="failed to get container status \"3d7d3c37c0853417594c89f5a5c074cdbc45350775c240b0c099690aa3ef9b4a\": rpc error: code = NotFound desc = could not find container \"3d7d3c37c0853417594c89f5a5c074cdbc45350775c240b0c099690aa3ef9b4a\": container with ID starting with 3d7d3c37c0853417594c89f5a5c074cdbc45350775c240b0c099690aa3ef9b4a not found: ID does not exist" Jan 26 17:11:55 crc kubenswrapper[4754]: I0126 17:11:54.427977 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Jan 26 17:11:55 crc kubenswrapper[4754]: I0126 17:11:54.590078 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef3b4399-44ec-476a-a0d5-e727b76c624e-combined-ca-bundle\") pod \"ef3b4399-44ec-476a-a0d5-e727b76c624e\" (UID: \"ef3b4399-44ec-476a-a0d5-e727b76c624e\") " Jan 26 17:11:55 crc kubenswrapper[4754]: I0126 17:11:54.590459 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wj9p7\" (UniqueName: \"kubernetes.io/projected/ef3b4399-44ec-476a-a0d5-e727b76c624e-kube-api-access-wj9p7\") pod \"ef3b4399-44ec-476a-a0d5-e727b76c624e\" (UID: \"ef3b4399-44ec-476a-a0d5-e727b76c624e\") " Jan 26 17:11:55 crc kubenswrapper[4754]: I0126 17:11:54.590486 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ef3b4399-44ec-476a-a0d5-e727b76c624e-config-data\") pod \"ef3b4399-44ec-476a-a0d5-e727b76c624e\" (UID: \"ef3b4399-44ec-476a-a0d5-e727b76c624e\") " Jan 26 17:11:55 crc kubenswrapper[4754]: I0126 17:11:54.594626 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ef3b4399-44ec-476a-a0d5-e727b76c624e-kube-api-access-wj9p7" (OuterVolumeSpecName: "kube-api-access-wj9p7") pod "ef3b4399-44ec-476a-a0d5-e727b76c624e" (UID: "ef3b4399-44ec-476a-a0d5-e727b76c624e"). InnerVolumeSpecName "kube-api-access-wj9p7". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:11:55 crc kubenswrapper[4754]: I0126 17:11:54.623316 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ef3b4399-44ec-476a-a0d5-e727b76c624e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ef3b4399-44ec-476a-a0d5-e727b76c624e" (UID: "ef3b4399-44ec-476a-a0d5-e727b76c624e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:11:55 crc kubenswrapper[4754]: I0126 17:11:54.629279 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ef3b4399-44ec-476a-a0d5-e727b76c624e-config-data" (OuterVolumeSpecName: "config-data") pod "ef3b4399-44ec-476a-a0d5-e727b76c624e" (UID: "ef3b4399-44ec-476a-a0d5-e727b76c624e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:11:55 crc kubenswrapper[4754]: I0126 17:11:54.692519 4754 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef3b4399-44ec-476a-a0d5-e727b76c624e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 26 17:11:55 crc kubenswrapper[4754]: I0126 17:11:54.692548 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wj9p7\" (UniqueName: \"kubernetes.io/projected/ef3b4399-44ec-476a-a0d5-e727b76c624e-kube-api-access-wj9p7\") on node \"crc\" DevicePath \"\"" Jan 26 17:11:55 crc kubenswrapper[4754]: I0126 17:11:54.692564 4754 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ef3b4399-44ec-476a-a0d5-e727b76c624e-config-data\") on node \"crc\" DevicePath \"\"" Jan 26 17:11:55 crc kubenswrapper[4754]: I0126 17:11:54.781283 4754 generic.go:334] "Generic (PLEG): container finished" podID="ef3b4399-44ec-476a-a0d5-e727b76c624e" containerID="5f0b0d2d8d721b498cb7253818255e0158226a13e1da30cdda99a766305b2734" exitCode=0 Jan 26 17:11:55 crc kubenswrapper[4754]: I0126 17:11:54.781332 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Jan 26 17:11:55 crc kubenswrapper[4754]: I0126 17:11:54.781368 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"ef3b4399-44ec-476a-a0d5-e727b76c624e","Type":"ContainerDied","Data":"5f0b0d2d8d721b498cb7253818255e0158226a13e1da30cdda99a766305b2734"} Jan 26 17:11:55 crc kubenswrapper[4754]: I0126 17:11:54.781397 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"ef3b4399-44ec-476a-a0d5-e727b76c624e","Type":"ContainerDied","Data":"fca9d9cf754082914a4799e437c7f2999dfdb540035e937df1644ce078be05b6"} Jan 26 17:11:55 crc kubenswrapper[4754]: I0126 17:11:54.781417 4754 scope.go:117] "RemoveContainer" containerID="5f0b0d2d8d721b498cb7253818255e0158226a13e1da30cdda99a766305b2734" Jan 26 17:11:55 crc kubenswrapper[4754]: I0126 17:11:54.812630 4754 scope.go:117] "RemoveContainer" containerID="5f0b0d2d8d721b498cb7253818255e0158226a13e1da30cdda99a766305b2734" Jan 26 17:11:55 crc kubenswrapper[4754]: E0126 17:11:54.813411 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5f0b0d2d8d721b498cb7253818255e0158226a13e1da30cdda99a766305b2734\": container with ID starting with 5f0b0d2d8d721b498cb7253818255e0158226a13e1da30cdda99a766305b2734 not found: ID does not exist" containerID="5f0b0d2d8d721b498cb7253818255e0158226a13e1da30cdda99a766305b2734" Jan 26 17:11:55 crc kubenswrapper[4754]: I0126 17:11:54.813471 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5f0b0d2d8d721b498cb7253818255e0158226a13e1da30cdda99a766305b2734"} err="failed to get container status \"5f0b0d2d8d721b498cb7253818255e0158226a13e1da30cdda99a766305b2734\": rpc error: code = NotFound desc = could not find container \"5f0b0d2d8d721b498cb7253818255e0158226a13e1da30cdda99a766305b2734\": container with ID starting with 5f0b0d2d8d721b498cb7253818255e0158226a13e1da30cdda99a766305b2734 not found: ID does not exist" Jan 26 17:11:55 crc kubenswrapper[4754]: I0126 17:11:54.823729 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Jan 26 17:11:55 crc kubenswrapper[4754]: I0126 17:11:54.839523 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Jan 26 17:11:55 crc kubenswrapper[4754]: I0126 17:11:54.850919 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Jan 26 17:11:55 crc kubenswrapper[4754]: E0126 17:11:54.851460 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e904e81b-a294-404a-8899-b4c86e043347" containerName="extract-utilities" Jan 26 17:11:55 crc kubenswrapper[4754]: I0126 17:11:54.851473 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="e904e81b-a294-404a-8899-b4c86e043347" containerName="extract-utilities" Jan 26 17:11:55 crc kubenswrapper[4754]: E0126 17:11:54.851567 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef3b4399-44ec-476a-a0d5-e727b76c624e" containerName="nova-scheduler-scheduler" Jan 26 17:11:55 crc kubenswrapper[4754]: I0126 17:11:54.851575 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef3b4399-44ec-476a-a0d5-e727b76c624e" containerName="nova-scheduler-scheduler" Jan 26 17:11:55 crc kubenswrapper[4754]: E0126 17:11:54.851600 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e904e81b-a294-404a-8899-b4c86e043347" containerName="registry-server" Jan 26 17:11:55 crc kubenswrapper[4754]: I0126 17:11:54.851606 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="e904e81b-a294-404a-8899-b4c86e043347" containerName="registry-server" Jan 26 17:11:55 crc kubenswrapper[4754]: E0126 17:11:54.851613 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7f22d2aa-d1fd-4a87-b0bd-9e3fb661d254" containerName="nova-manage" Jan 26 17:11:55 crc kubenswrapper[4754]: I0126 17:11:54.851618 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="7f22d2aa-d1fd-4a87-b0bd-9e3fb661d254" containerName="nova-manage" Jan 26 17:11:55 crc kubenswrapper[4754]: E0126 17:11:54.851630 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e904e81b-a294-404a-8899-b4c86e043347" containerName="extract-content" Jan 26 17:11:55 crc kubenswrapper[4754]: I0126 17:11:54.851636 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="e904e81b-a294-404a-8899-b4c86e043347" containerName="extract-content" Jan 26 17:11:55 crc kubenswrapper[4754]: I0126 17:11:54.851864 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="7f22d2aa-d1fd-4a87-b0bd-9e3fb661d254" containerName="nova-manage" Jan 26 17:11:55 crc kubenswrapper[4754]: I0126 17:11:54.851878 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="e904e81b-a294-404a-8899-b4c86e043347" containerName="registry-server" Jan 26 17:11:55 crc kubenswrapper[4754]: I0126 17:11:54.851893 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="ef3b4399-44ec-476a-a0d5-e727b76c624e" containerName="nova-scheduler-scheduler" Jan 26 17:11:55 crc kubenswrapper[4754]: I0126 17:11:54.852898 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Jan 26 17:11:55 crc kubenswrapper[4754]: I0126 17:11:54.860738 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Jan 26 17:11:55 crc kubenswrapper[4754]: I0126 17:11:54.862857 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Jan 26 17:11:55 crc kubenswrapper[4754]: I0126 17:11:54.896180 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6f952bb4-186c-4e7c-bf01-67d8a6985319-config-data\") pod \"nova-scheduler-0\" (UID: \"6f952bb4-186c-4e7c-bf01-67d8a6985319\") " pod="openstack/nova-scheduler-0" Jan 26 17:11:55 crc kubenswrapper[4754]: I0126 17:11:54.896337 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f952bb4-186c-4e7c-bf01-67d8a6985319-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"6f952bb4-186c-4e7c-bf01-67d8a6985319\") " pod="openstack/nova-scheduler-0" Jan 26 17:11:55 crc kubenswrapper[4754]: I0126 17:11:54.896556 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sjfmk\" (UniqueName: \"kubernetes.io/projected/6f952bb4-186c-4e7c-bf01-67d8a6985319-kube-api-access-sjfmk\") pod \"nova-scheduler-0\" (UID: \"6f952bb4-186c-4e7c-bf01-67d8a6985319\") " pod="openstack/nova-scheduler-0" Jan 26 17:11:55 crc kubenswrapper[4754]: I0126 17:11:54.997659 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sjfmk\" (UniqueName: \"kubernetes.io/projected/6f952bb4-186c-4e7c-bf01-67d8a6985319-kube-api-access-sjfmk\") pod \"nova-scheduler-0\" (UID: \"6f952bb4-186c-4e7c-bf01-67d8a6985319\") " pod="openstack/nova-scheduler-0" Jan 26 17:11:55 crc kubenswrapper[4754]: I0126 17:11:54.997786 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6f952bb4-186c-4e7c-bf01-67d8a6985319-config-data\") pod \"nova-scheduler-0\" (UID: \"6f952bb4-186c-4e7c-bf01-67d8a6985319\") " pod="openstack/nova-scheduler-0" Jan 26 17:11:55 crc kubenswrapper[4754]: I0126 17:11:54.997914 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f952bb4-186c-4e7c-bf01-67d8a6985319-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"6f952bb4-186c-4e7c-bf01-67d8a6985319\") " pod="openstack/nova-scheduler-0" Jan 26 17:11:55 crc kubenswrapper[4754]: I0126 17:11:55.002036 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6f952bb4-186c-4e7c-bf01-67d8a6985319-config-data\") pod \"nova-scheduler-0\" (UID: \"6f952bb4-186c-4e7c-bf01-67d8a6985319\") " pod="openstack/nova-scheduler-0" Jan 26 17:11:55 crc kubenswrapper[4754]: I0126 17:11:55.002210 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f952bb4-186c-4e7c-bf01-67d8a6985319-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"6f952bb4-186c-4e7c-bf01-67d8a6985319\") " pod="openstack/nova-scheduler-0" Jan 26 17:11:55 crc kubenswrapper[4754]: I0126 17:11:55.017252 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sjfmk\" (UniqueName: \"kubernetes.io/projected/6f952bb4-186c-4e7c-bf01-67d8a6985319-kube-api-access-sjfmk\") pod \"nova-scheduler-0\" (UID: \"6f952bb4-186c-4e7c-bf01-67d8a6985319\") " pod="openstack/nova-scheduler-0" Jan 26 17:11:55 crc kubenswrapper[4754]: I0126 17:11:55.068397 4754 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="37f5b6db-8ef3-4e4f-9763-b150188c1466" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.201:8775/\": read tcp 10.217.0.2:55072->10.217.0.201:8775: read: connection reset by peer" Jan 26 17:11:55 crc kubenswrapper[4754]: I0126 17:11:55.068404 4754 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="37f5b6db-8ef3-4e4f-9763-b150188c1466" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.201:8775/\": read tcp 10.217.0.2:55086->10.217.0.201:8775: read: connection reset by peer" Jan 26 17:11:55 crc kubenswrapper[4754]: I0126 17:11:55.187239 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Jan 26 17:11:55 crc kubenswrapper[4754]: I0126 17:11:55.684496 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Jan 26 17:11:55 crc kubenswrapper[4754]: I0126 17:11:55.694071 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Jan 26 17:11:55 crc kubenswrapper[4754]: W0126 17:11:55.774639 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6f952bb4_186c_4e7c_bf01_67d8a6985319.slice/crio-72130c8bd4a3dc36614bf554b39838881fd61069e26218502e769a68eee4583a WatchSource:0}: Error finding container 72130c8bd4a3dc36614bf554b39838881fd61069e26218502e769a68eee4583a: Status 404 returned error can't find the container with id 72130c8bd4a3dc36614bf554b39838881fd61069e26218502e769a68eee4583a Jan 26 17:11:55 crc kubenswrapper[4754]: I0126 17:11:55.790456 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e904e81b-a294-404a-8899-b4c86e043347" path="/var/lib/kubelet/pods/e904e81b-a294-404a-8899-b4c86e043347/volumes" Jan 26 17:11:55 crc kubenswrapper[4754]: I0126 17:11:55.792741 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ef3b4399-44ec-476a-a0d5-e727b76c624e" path="/var/lib/kubelet/pods/ef3b4399-44ec-476a-a0d5-e727b76c624e/volumes" Jan 26 17:11:55 crc kubenswrapper[4754]: I0126 17:11:55.796555 4754 generic.go:334] "Generic (PLEG): container finished" podID="19c7e3a1-e40c-401b-bd01-d2a50f14daf0" containerID="2afb5680b5deea7b4857e1eaf4bdb0cb81e95fbdaa9a07d49daff55e3720c563" exitCode=0 Jan 26 17:11:55 crc kubenswrapper[4754]: I0126 17:11:55.796724 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Jan 26 17:11:55 crc kubenswrapper[4754]: I0126 17:11:55.800897 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Jan 26 17:11:55 crc kubenswrapper[4754]: I0126 17:11:55.800932 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"19c7e3a1-e40c-401b-bd01-d2a50f14daf0","Type":"ContainerDied","Data":"2afb5680b5deea7b4857e1eaf4bdb0cb81e95fbdaa9a07d49daff55e3720c563"} Jan 26 17:11:55 crc kubenswrapper[4754]: I0126 17:11:55.800955 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"19c7e3a1-e40c-401b-bd01-d2a50f14daf0","Type":"ContainerDied","Data":"d13a156e3bdf571dfb66b2fb4994bc6066b523c0a6c088c7e31dd118c5f11a32"} Jan 26 17:11:55 crc kubenswrapper[4754]: I0126 17:11:55.800968 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"6f952bb4-186c-4e7c-bf01-67d8a6985319","Type":"ContainerStarted","Data":"72130c8bd4a3dc36614bf554b39838881fd61069e26218502e769a68eee4583a"} Jan 26 17:11:55 crc kubenswrapper[4754]: I0126 17:11:55.800989 4754 scope.go:117] "RemoveContainer" containerID="2afb5680b5deea7b4857e1eaf4bdb0cb81e95fbdaa9a07d49daff55e3720c563" Jan 26 17:11:55 crc kubenswrapper[4754]: I0126 17:11:55.802445 4754 generic.go:334] "Generic (PLEG): container finished" podID="37f5b6db-8ef3-4e4f-9763-b150188c1466" containerID="b8bcc64a29d1de162bd574631509ab962e70a5eba797922712a474b28c35c163" exitCode=0 Jan 26 17:11:55 crc kubenswrapper[4754]: I0126 17:11:55.802475 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"37f5b6db-8ef3-4e4f-9763-b150188c1466","Type":"ContainerDied","Data":"b8bcc64a29d1de162bd574631509ab962e70a5eba797922712a474b28c35c163"} Jan 26 17:11:55 crc kubenswrapper[4754]: I0126 17:11:55.802493 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"37f5b6db-8ef3-4e4f-9763-b150188c1466","Type":"ContainerDied","Data":"40200c2ad928b0e5d0cf5e04a79d89f21e722b73f5f6bee5ec83eb7ebff6d363"} Jan 26 17:11:55 crc kubenswrapper[4754]: I0126 17:11:55.802552 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Jan 26 17:11:55 crc kubenswrapper[4754]: I0126 17:11:55.810565 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37f5b6db-8ef3-4e4f-9763-b150188c1466-config-data\") pod \"37f5b6db-8ef3-4e4f-9763-b150188c1466\" (UID: \"37f5b6db-8ef3-4e4f-9763-b150188c1466\") " Jan 26 17:11:55 crc kubenswrapper[4754]: I0126 17:11:55.810632 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/19c7e3a1-e40c-401b-bd01-d2a50f14daf0-config-data\") pod \"19c7e3a1-e40c-401b-bd01-d2a50f14daf0\" (UID: \"19c7e3a1-e40c-401b-bd01-d2a50f14daf0\") " Jan 26 17:11:55 crc kubenswrapper[4754]: I0126 17:11:55.810684 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qgkcc\" (UniqueName: \"kubernetes.io/projected/19c7e3a1-e40c-401b-bd01-d2a50f14daf0-kube-api-access-qgkcc\") pod \"19c7e3a1-e40c-401b-bd01-d2a50f14daf0\" (UID: \"19c7e3a1-e40c-401b-bd01-d2a50f14daf0\") " Jan 26 17:11:55 crc kubenswrapper[4754]: I0126 17:11:55.810755 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/37f5b6db-8ef3-4e4f-9763-b150188c1466-logs\") pod \"37f5b6db-8ef3-4e4f-9763-b150188c1466\" (UID: \"37f5b6db-8ef3-4e4f-9763-b150188c1466\") " Jan 26 17:11:55 crc kubenswrapper[4754]: I0126 17:11:55.810807 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/37f5b6db-8ef3-4e4f-9763-b150188c1466-nova-metadata-tls-certs\") pod \"37f5b6db-8ef3-4e4f-9763-b150188c1466\" (UID: \"37f5b6db-8ef3-4e4f-9763-b150188c1466\") " Jan 26 17:11:55 crc kubenswrapper[4754]: I0126 17:11:55.810884 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/19c7e3a1-e40c-401b-bd01-d2a50f14daf0-public-tls-certs\") pod \"19c7e3a1-e40c-401b-bd01-d2a50f14daf0\" (UID: \"19c7e3a1-e40c-401b-bd01-d2a50f14daf0\") " Jan 26 17:11:55 crc kubenswrapper[4754]: I0126 17:11:55.810942 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/19c7e3a1-e40c-401b-bd01-d2a50f14daf0-logs\") pod \"19c7e3a1-e40c-401b-bd01-d2a50f14daf0\" (UID: \"19c7e3a1-e40c-401b-bd01-d2a50f14daf0\") " Jan 26 17:11:55 crc kubenswrapper[4754]: I0126 17:11:55.810960 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19c7e3a1-e40c-401b-bd01-d2a50f14daf0-combined-ca-bundle\") pod \"19c7e3a1-e40c-401b-bd01-d2a50f14daf0\" (UID: \"19c7e3a1-e40c-401b-bd01-d2a50f14daf0\") " Jan 26 17:11:55 crc kubenswrapper[4754]: I0126 17:11:55.810980 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/19c7e3a1-e40c-401b-bd01-d2a50f14daf0-internal-tls-certs\") pod \"19c7e3a1-e40c-401b-bd01-d2a50f14daf0\" (UID: \"19c7e3a1-e40c-401b-bd01-d2a50f14daf0\") " Jan 26 17:11:55 crc kubenswrapper[4754]: I0126 17:11:55.811018 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hf9q7\" (UniqueName: \"kubernetes.io/projected/37f5b6db-8ef3-4e4f-9763-b150188c1466-kube-api-access-hf9q7\") pod \"37f5b6db-8ef3-4e4f-9763-b150188c1466\" (UID: \"37f5b6db-8ef3-4e4f-9763-b150188c1466\") " Jan 26 17:11:55 crc kubenswrapper[4754]: I0126 17:11:55.811046 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37f5b6db-8ef3-4e4f-9763-b150188c1466-combined-ca-bundle\") pod \"37f5b6db-8ef3-4e4f-9763-b150188c1466\" (UID: \"37f5b6db-8ef3-4e4f-9763-b150188c1466\") " Jan 26 17:11:55 crc kubenswrapper[4754]: I0126 17:11:55.817305 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/37f5b6db-8ef3-4e4f-9763-b150188c1466-kube-api-access-hf9q7" (OuterVolumeSpecName: "kube-api-access-hf9q7") pod "37f5b6db-8ef3-4e4f-9763-b150188c1466" (UID: "37f5b6db-8ef3-4e4f-9763-b150188c1466"). InnerVolumeSpecName "kube-api-access-hf9q7". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:11:55 crc kubenswrapper[4754]: I0126 17:11:55.817739 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/37f5b6db-8ef3-4e4f-9763-b150188c1466-logs" (OuterVolumeSpecName: "logs") pod "37f5b6db-8ef3-4e4f-9763-b150188c1466" (UID: "37f5b6db-8ef3-4e4f-9763-b150188c1466"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:11:55 crc kubenswrapper[4754]: I0126 17:11:55.818924 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/19c7e3a1-e40c-401b-bd01-d2a50f14daf0-logs" (OuterVolumeSpecName: "logs") pod "19c7e3a1-e40c-401b-bd01-d2a50f14daf0" (UID: "19c7e3a1-e40c-401b-bd01-d2a50f14daf0"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:11:55 crc kubenswrapper[4754]: I0126 17:11:55.822247 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/19c7e3a1-e40c-401b-bd01-d2a50f14daf0-kube-api-access-qgkcc" (OuterVolumeSpecName: "kube-api-access-qgkcc") pod "19c7e3a1-e40c-401b-bd01-d2a50f14daf0" (UID: "19c7e3a1-e40c-401b-bd01-d2a50f14daf0"). InnerVolumeSpecName "kube-api-access-qgkcc". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:11:55 crc kubenswrapper[4754]: I0126 17:11:55.843166 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/37f5b6db-8ef3-4e4f-9763-b150188c1466-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "37f5b6db-8ef3-4e4f-9763-b150188c1466" (UID: "37f5b6db-8ef3-4e4f-9763-b150188c1466"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:11:55 crc kubenswrapper[4754]: I0126 17:11:55.857523 4754 scope.go:117] "RemoveContainer" containerID="d6c3849288f4a670bee53c01dd166a87f4ed85c177794e0645c152549b809bd1" Jan 26 17:11:55 crc kubenswrapper[4754]: I0126 17:11:55.861070 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/37f5b6db-8ef3-4e4f-9763-b150188c1466-config-data" (OuterVolumeSpecName: "config-data") pod "37f5b6db-8ef3-4e4f-9763-b150188c1466" (UID: "37f5b6db-8ef3-4e4f-9763-b150188c1466"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:11:55 crc kubenswrapper[4754]: I0126 17:11:55.861311 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/19c7e3a1-e40c-401b-bd01-d2a50f14daf0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "19c7e3a1-e40c-401b-bd01-d2a50f14daf0" (UID: "19c7e3a1-e40c-401b-bd01-d2a50f14daf0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:11:55 crc kubenswrapper[4754]: I0126 17:11:55.868057 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/19c7e3a1-e40c-401b-bd01-d2a50f14daf0-config-data" (OuterVolumeSpecName: "config-data") pod "19c7e3a1-e40c-401b-bd01-d2a50f14daf0" (UID: "19c7e3a1-e40c-401b-bd01-d2a50f14daf0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:11:55 crc kubenswrapper[4754]: I0126 17:11:55.882407 4754 scope.go:117] "RemoveContainer" containerID="2afb5680b5deea7b4857e1eaf4bdb0cb81e95fbdaa9a07d49daff55e3720c563" Jan 26 17:11:55 crc kubenswrapper[4754]: E0126 17:11:55.882956 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2afb5680b5deea7b4857e1eaf4bdb0cb81e95fbdaa9a07d49daff55e3720c563\": container with ID starting with 2afb5680b5deea7b4857e1eaf4bdb0cb81e95fbdaa9a07d49daff55e3720c563 not found: ID does not exist" containerID="2afb5680b5deea7b4857e1eaf4bdb0cb81e95fbdaa9a07d49daff55e3720c563" Jan 26 17:11:55 crc kubenswrapper[4754]: I0126 17:11:55.882994 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2afb5680b5deea7b4857e1eaf4bdb0cb81e95fbdaa9a07d49daff55e3720c563"} err="failed to get container status \"2afb5680b5deea7b4857e1eaf4bdb0cb81e95fbdaa9a07d49daff55e3720c563\": rpc error: code = NotFound desc = could not find container \"2afb5680b5deea7b4857e1eaf4bdb0cb81e95fbdaa9a07d49daff55e3720c563\": container with ID starting with 2afb5680b5deea7b4857e1eaf4bdb0cb81e95fbdaa9a07d49daff55e3720c563 not found: ID does not exist" Jan 26 17:11:55 crc kubenswrapper[4754]: I0126 17:11:55.883015 4754 scope.go:117] "RemoveContainer" containerID="d6c3849288f4a670bee53c01dd166a87f4ed85c177794e0645c152549b809bd1" Jan 26 17:11:55 crc kubenswrapper[4754]: E0126 17:11:55.883325 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d6c3849288f4a670bee53c01dd166a87f4ed85c177794e0645c152549b809bd1\": container with ID starting with d6c3849288f4a670bee53c01dd166a87f4ed85c177794e0645c152549b809bd1 not found: ID does not exist" containerID="d6c3849288f4a670bee53c01dd166a87f4ed85c177794e0645c152549b809bd1" Jan 26 17:11:55 crc kubenswrapper[4754]: I0126 17:11:55.883427 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d6c3849288f4a670bee53c01dd166a87f4ed85c177794e0645c152549b809bd1"} err="failed to get container status \"d6c3849288f4a670bee53c01dd166a87f4ed85c177794e0645c152549b809bd1\": rpc error: code = NotFound desc = could not find container \"d6c3849288f4a670bee53c01dd166a87f4ed85c177794e0645c152549b809bd1\": container with ID starting with d6c3849288f4a670bee53c01dd166a87f4ed85c177794e0645c152549b809bd1 not found: ID does not exist" Jan 26 17:11:55 crc kubenswrapper[4754]: I0126 17:11:55.883521 4754 scope.go:117] "RemoveContainer" containerID="b8bcc64a29d1de162bd574631509ab962e70a5eba797922712a474b28c35c163" Jan 26 17:11:55 crc kubenswrapper[4754]: I0126 17:11:55.889635 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/19c7e3a1-e40c-401b-bd01-d2a50f14daf0-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "19c7e3a1-e40c-401b-bd01-d2a50f14daf0" (UID: "19c7e3a1-e40c-401b-bd01-d2a50f14daf0"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:11:55 crc kubenswrapper[4754]: I0126 17:11:55.900094 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/37f5b6db-8ef3-4e4f-9763-b150188c1466-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "37f5b6db-8ef3-4e4f-9763-b150188c1466" (UID: "37f5b6db-8ef3-4e4f-9763-b150188c1466"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:11:55 crc kubenswrapper[4754]: I0126 17:11:55.906948 4754 scope.go:117] "RemoveContainer" containerID="911ce027833e9fd70123c130dde99ab0e73b4f39eb0c6df15227bfe3a7a4e4cf" Jan 26 17:11:55 crc kubenswrapper[4754]: I0126 17:11:55.907874 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/19c7e3a1-e40c-401b-bd01-d2a50f14daf0-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "19c7e3a1-e40c-401b-bd01-d2a50f14daf0" (UID: "19c7e3a1-e40c-401b-bd01-d2a50f14daf0"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:11:55 crc kubenswrapper[4754]: I0126 17:11:55.914647 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hf9q7\" (UniqueName: \"kubernetes.io/projected/37f5b6db-8ef3-4e4f-9763-b150188c1466-kube-api-access-hf9q7\") on node \"crc\" DevicePath \"\"" Jan 26 17:11:55 crc kubenswrapper[4754]: I0126 17:11:55.914747 4754 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37f5b6db-8ef3-4e4f-9763-b150188c1466-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 26 17:11:55 crc kubenswrapper[4754]: I0126 17:11:55.914759 4754 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37f5b6db-8ef3-4e4f-9763-b150188c1466-config-data\") on node \"crc\" DevicePath \"\"" Jan 26 17:11:55 crc kubenswrapper[4754]: I0126 17:11:55.914768 4754 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/19c7e3a1-e40c-401b-bd01-d2a50f14daf0-config-data\") on node \"crc\" DevicePath \"\"" Jan 26 17:11:55 crc kubenswrapper[4754]: I0126 17:11:55.914777 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qgkcc\" (UniqueName: \"kubernetes.io/projected/19c7e3a1-e40c-401b-bd01-d2a50f14daf0-kube-api-access-qgkcc\") on node \"crc\" DevicePath \"\"" Jan 26 17:11:55 crc kubenswrapper[4754]: I0126 17:11:55.914786 4754 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/37f5b6db-8ef3-4e4f-9763-b150188c1466-logs\") on node \"crc\" DevicePath \"\"" Jan 26 17:11:55 crc kubenswrapper[4754]: I0126 17:11:55.914795 4754 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/37f5b6db-8ef3-4e4f-9763-b150188c1466-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Jan 26 17:11:55 crc kubenswrapper[4754]: I0126 17:11:55.914803 4754 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/19c7e3a1-e40c-401b-bd01-d2a50f14daf0-public-tls-certs\") on node \"crc\" DevicePath \"\"" Jan 26 17:11:55 crc kubenswrapper[4754]: I0126 17:11:55.914811 4754 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/19c7e3a1-e40c-401b-bd01-d2a50f14daf0-logs\") on node \"crc\" DevicePath \"\"" Jan 26 17:11:55 crc kubenswrapper[4754]: I0126 17:11:55.914819 4754 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19c7e3a1-e40c-401b-bd01-d2a50f14daf0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 26 17:11:55 crc kubenswrapper[4754]: I0126 17:11:55.914827 4754 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/19c7e3a1-e40c-401b-bd01-d2a50f14daf0-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Jan 26 17:11:55 crc kubenswrapper[4754]: I0126 17:11:55.928149 4754 scope.go:117] "RemoveContainer" containerID="b8bcc64a29d1de162bd574631509ab962e70a5eba797922712a474b28c35c163" Jan 26 17:11:55 crc kubenswrapper[4754]: E0126 17:11:55.928619 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b8bcc64a29d1de162bd574631509ab962e70a5eba797922712a474b28c35c163\": container with ID starting with b8bcc64a29d1de162bd574631509ab962e70a5eba797922712a474b28c35c163 not found: ID does not exist" containerID="b8bcc64a29d1de162bd574631509ab962e70a5eba797922712a474b28c35c163" Jan 26 17:11:55 crc kubenswrapper[4754]: I0126 17:11:55.928654 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b8bcc64a29d1de162bd574631509ab962e70a5eba797922712a474b28c35c163"} err="failed to get container status \"b8bcc64a29d1de162bd574631509ab962e70a5eba797922712a474b28c35c163\": rpc error: code = NotFound desc = could not find container \"b8bcc64a29d1de162bd574631509ab962e70a5eba797922712a474b28c35c163\": container with ID starting with b8bcc64a29d1de162bd574631509ab962e70a5eba797922712a474b28c35c163 not found: ID does not exist" Jan 26 17:11:55 crc kubenswrapper[4754]: I0126 17:11:55.928694 4754 scope.go:117] "RemoveContainer" containerID="911ce027833e9fd70123c130dde99ab0e73b4f39eb0c6df15227bfe3a7a4e4cf" Jan 26 17:11:55 crc kubenswrapper[4754]: E0126 17:11:55.929058 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"911ce027833e9fd70123c130dde99ab0e73b4f39eb0c6df15227bfe3a7a4e4cf\": container with ID starting with 911ce027833e9fd70123c130dde99ab0e73b4f39eb0c6df15227bfe3a7a4e4cf not found: ID does not exist" containerID="911ce027833e9fd70123c130dde99ab0e73b4f39eb0c6df15227bfe3a7a4e4cf" Jan 26 17:11:55 crc kubenswrapper[4754]: I0126 17:11:55.929089 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"911ce027833e9fd70123c130dde99ab0e73b4f39eb0c6df15227bfe3a7a4e4cf"} err="failed to get container status \"911ce027833e9fd70123c130dde99ab0e73b4f39eb0c6df15227bfe3a7a4e4cf\": rpc error: code = NotFound desc = could not find container \"911ce027833e9fd70123c130dde99ab0e73b4f39eb0c6df15227bfe3a7a4e4cf\": container with ID starting with 911ce027833e9fd70123c130dde99ab0e73b4f39eb0c6df15227bfe3a7a4e4cf not found: ID does not exist" Jan 26 17:11:56 crc kubenswrapper[4754]: I0126 17:11:56.136474 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Jan 26 17:11:56 crc kubenswrapper[4754]: I0126 17:11:56.190632 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Jan 26 17:11:56 crc kubenswrapper[4754]: I0126 17:11:56.204031 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Jan 26 17:11:56 crc kubenswrapper[4754]: I0126 17:11:56.217657 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Jan 26 17:11:56 crc kubenswrapper[4754]: I0126 17:11:56.228142 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Jan 26 17:11:56 crc kubenswrapper[4754]: E0126 17:11:56.228527 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37f5b6db-8ef3-4e4f-9763-b150188c1466" containerName="nova-metadata-metadata" Jan 26 17:11:56 crc kubenswrapper[4754]: I0126 17:11:56.228554 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="37f5b6db-8ef3-4e4f-9763-b150188c1466" containerName="nova-metadata-metadata" Jan 26 17:11:56 crc kubenswrapper[4754]: E0126 17:11:56.228578 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37f5b6db-8ef3-4e4f-9763-b150188c1466" containerName="nova-metadata-log" Jan 26 17:11:56 crc kubenswrapper[4754]: I0126 17:11:56.228587 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="37f5b6db-8ef3-4e4f-9763-b150188c1466" containerName="nova-metadata-log" Jan 26 17:11:56 crc kubenswrapper[4754]: E0126 17:11:56.228609 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19c7e3a1-e40c-401b-bd01-d2a50f14daf0" containerName="nova-api-log" Jan 26 17:11:56 crc kubenswrapper[4754]: I0126 17:11:56.228620 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="19c7e3a1-e40c-401b-bd01-d2a50f14daf0" containerName="nova-api-log" Jan 26 17:11:56 crc kubenswrapper[4754]: E0126 17:11:56.228635 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19c7e3a1-e40c-401b-bd01-d2a50f14daf0" containerName="nova-api-api" Jan 26 17:11:56 crc kubenswrapper[4754]: I0126 17:11:56.228643 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="19c7e3a1-e40c-401b-bd01-d2a50f14daf0" containerName="nova-api-api" Jan 26 17:11:56 crc kubenswrapper[4754]: I0126 17:11:56.228903 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="19c7e3a1-e40c-401b-bd01-d2a50f14daf0" containerName="nova-api-api" Jan 26 17:11:56 crc kubenswrapper[4754]: I0126 17:11:56.228938 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="37f5b6db-8ef3-4e4f-9763-b150188c1466" containerName="nova-metadata-metadata" Jan 26 17:11:56 crc kubenswrapper[4754]: I0126 17:11:56.228948 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="19c7e3a1-e40c-401b-bd01-d2a50f14daf0" containerName="nova-api-log" Jan 26 17:11:56 crc kubenswrapper[4754]: I0126 17:11:56.228968 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="37f5b6db-8ef3-4e4f-9763-b150188c1466" containerName="nova-metadata-log" Jan 26 17:11:56 crc kubenswrapper[4754]: I0126 17:11:56.230107 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Jan 26 17:11:56 crc kubenswrapper[4754]: I0126 17:11:56.233613 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Jan 26 17:11:56 crc kubenswrapper[4754]: I0126 17:11:56.238156 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Jan 26 17:11:56 crc kubenswrapper[4754]: I0126 17:11:56.255634 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Jan 26 17:11:56 crc kubenswrapper[4754]: I0126 17:11:56.265507 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Jan 26 17:11:56 crc kubenswrapper[4754]: I0126 17:11:56.267127 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Jan 26 17:11:56 crc kubenswrapper[4754]: I0126 17:11:56.269625 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Jan 26 17:11:56 crc kubenswrapper[4754]: I0126 17:11:56.270271 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Jan 26 17:11:56 crc kubenswrapper[4754]: I0126 17:11:56.270621 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Jan 26 17:11:56 crc kubenswrapper[4754]: I0126 17:11:56.275615 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Jan 26 17:11:56 crc kubenswrapper[4754]: I0126 17:11:56.323919 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e7eefb7-e66a-4c65-bda9-7997970991f3-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"4e7eefb7-e66a-4c65-bda9-7997970991f3\") " pod="openstack/nova-metadata-0" Jan 26 17:11:56 crc kubenswrapper[4754]: I0126 17:11:56.324018 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0bb1f389-c4a6-47dd-8445-c7125779ef38-public-tls-certs\") pod \"nova-api-0\" (UID: \"0bb1f389-c4a6-47dd-8445-c7125779ef38\") " pod="openstack/nova-api-0" Jan 26 17:11:56 crc kubenswrapper[4754]: I0126 17:11:56.324048 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0bb1f389-c4a6-47dd-8445-c7125779ef38-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"0bb1f389-c4a6-47dd-8445-c7125779ef38\") " pod="openstack/nova-api-0" Jan 26 17:11:56 crc kubenswrapper[4754]: I0126 17:11:56.324169 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4e7eefb7-e66a-4c65-bda9-7997970991f3-config-data\") pod \"nova-metadata-0\" (UID: \"4e7eefb7-e66a-4c65-bda9-7997970991f3\") " pod="openstack/nova-metadata-0" Jan 26 17:11:56 crc kubenswrapper[4754]: I0126 17:11:56.324293 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0bb1f389-c4a6-47dd-8445-c7125779ef38-internal-tls-certs\") pod \"nova-api-0\" (UID: \"0bb1f389-c4a6-47dd-8445-c7125779ef38\") " pod="openstack/nova-api-0" Jan 26 17:11:56 crc kubenswrapper[4754]: I0126 17:11:56.324739 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/4e7eefb7-e66a-4c65-bda9-7997970991f3-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"4e7eefb7-e66a-4c65-bda9-7997970991f3\") " pod="openstack/nova-metadata-0" Jan 26 17:11:56 crc kubenswrapper[4754]: I0126 17:11:56.325107 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zsrvl\" (UniqueName: \"kubernetes.io/projected/4e7eefb7-e66a-4c65-bda9-7997970991f3-kube-api-access-zsrvl\") pod \"nova-metadata-0\" (UID: \"4e7eefb7-e66a-4c65-bda9-7997970991f3\") " pod="openstack/nova-metadata-0" Jan 26 17:11:56 crc kubenswrapper[4754]: I0126 17:11:56.325189 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0bb1f389-c4a6-47dd-8445-c7125779ef38-logs\") pod \"nova-api-0\" (UID: \"0bb1f389-c4a6-47dd-8445-c7125779ef38\") " pod="openstack/nova-api-0" Jan 26 17:11:56 crc kubenswrapper[4754]: I0126 17:11:56.325306 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4e7eefb7-e66a-4c65-bda9-7997970991f3-logs\") pod \"nova-metadata-0\" (UID: \"4e7eefb7-e66a-4c65-bda9-7997970991f3\") " pod="openstack/nova-metadata-0" Jan 26 17:11:56 crc kubenswrapper[4754]: I0126 17:11:56.325430 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6b8k8\" (UniqueName: \"kubernetes.io/projected/0bb1f389-c4a6-47dd-8445-c7125779ef38-kube-api-access-6b8k8\") pod \"nova-api-0\" (UID: \"0bb1f389-c4a6-47dd-8445-c7125779ef38\") " pod="openstack/nova-api-0" Jan 26 17:11:56 crc kubenswrapper[4754]: I0126 17:11:56.325499 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0bb1f389-c4a6-47dd-8445-c7125779ef38-config-data\") pod \"nova-api-0\" (UID: \"0bb1f389-c4a6-47dd-8445-c7125779ef38\") " pod="openstack/nova-api-0" Jan 26 17:11:56 crc kubenswrapper[4754]: I0126 17:11:56.426579 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zsrvl\" (UniqueName: \"kubernetes.io/projected/4e7eefb7-e66a-4c65-bda9-7997970991f3-kube-api-access-zsrvl\") pod \"nova-metadata-0\" (UID: \"4e7eefb7-e66a-4c65-bda9-7997970991f3\") " pod="openstack/nova-metadata-0" Jan 26 17:11:56 crc kubenswrapper[4754]: I0126 17:11:56.426870 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0bb1f389-c4a6-47dd-8445-c7125779ef38-logs\") pod \"nova-api-0\" (UID: \"0bb1f389-c4a6-47dd-8445-c7125779ef38\") " pod="openstack/nova-api-0" Jan 26 17:11:56 crc kubenswrapper[4754]: I0126 17:11:56.426894 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4e7eefb7-e66a-4c65-bda9-7997970991f3-logs\") pod \"nova-metadata-0\" (UID: \"4e7eefb7-e66a-4c65-bda9-7997970991f3\") " pod="openstack/nova-metadata-0" Jan 26 17:11:56 crc kubenswrapper[4754]: I0126 17:11:56.426927 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6b8k8\" (UniqueName: \"kubernetes.io/projected/0bb1f389-c4a6-47dd-8445-c7125779ef38-kube-api-access-6b8k8\") pod \"nova-api-0\" (UID: \"0bb1f389-c4a6-47dd-8445-c7125779ef38\") " pod="openstack/nova-api-0" Jan 26 17:11:56 crc kubenswrapper[4754]: I0126 17:11:56.426947 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0bb1f389-c4a6-47dd-8445-c7125779ef38-config-data\") pod \"nova-api-0\" (UID: \"0bb1f389-c4a6-47dd-8445-c7125779ef38\") " pod="openstack/nova-api-0" Jan 26 17:11:56 crc kubenswrapper[4754]: I0126 17:11:56.426986 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e7eefb7-e66a-4c65-bda9-7997970991f3-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"4e7eefb7-e66a-4c65-bda9-7997970991f3\") " pod="openstack/nova-metadata-0" Jan 26 17:11:56 crc kubenswrapper[4754]: I0126 17:11:56.427026 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0bb1f389-c4a6-47dd-8445-c7125779ef38-public-tls-certs\") pod \"nova-api-0\" (UID: \"0bb1f389-c4a6-47dd-8445-c7125779ef38\") " pod="openstack/nova-api-0" Jan 26 17:11:56 crc kubenswrapper[4754]: I0126 17:11:56.427042 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0bb1f389-c4a6-47dd-8445-c7125779ef38-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"0bb1f389-c4a6-47dd-8445-c7125779ef38\") " pod="openstack/nova-api-0" Jan 26 17:11:56 crc kubenswrapper[4754]: I0126 17:11:56.427064 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4e7eefb7-e66a-4c65-bda9-7997970991f3-config-data\") pod \"nova-metadata-0\" (UID: \"4e7eefb7-e66a-4c65-bda9-7997970991f3\") " pod="openstack/nova-metadata-0" Jan 26 17:11:56 crc kubenswrapper[4754]: I0126 17:11:56.427090 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0bb1f389-c4a6-47dd-8445-c7125779ef38-internal-tls-certs\") pod \"nova-api-0\" (UID: \"0bb1f389-c4a6-47dd-8445-c7125779ef38\") " pod="openstack/nova-api-0" Jan 26 17:11:56 crc kubenswrapper[4754]: I0126 17:11:56.427121 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/4e7eefb7-e66a-4c65-bda9-7997970991f3-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"4e7eefb7-e66a-4c65-bda9-7997970991f3\") " pod="openstack/nova-metadata-0" Jan 26 17:11:56 crc kubenswrapper[4754]: I0126 17:11:56.427285 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4e7eefb7-e66a-4c65-bda9-7997970991f3-logs\") pod \"nova-metadata-0\" (UID: \"4e7eefb7-e66a-4c65-bda9-7997970991f3\") " pod="openstack/nova-metadata-0" Jan 26 17:11:56 crc kubenswrapper[4754]: I0126 17:11:56.427348 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0bb1f389-c4a6-47dd-8445-c7125779ef38-logs\") pod \"nova-api-0\" (UID: \"0bb1f389-c4a6-47dd-8445-c7125779ef38\") " pod="openstack/nova-api-0" Jan 26 17:11:56 crc kubenswrapper[4754]: I0126 17:11:56.432505 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/4e7eefb7-e66a-4c65-bda9-7997970991f3-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"4e7eefb7-e66a-4c65-bda9-7997970991f3\") " pod="openstack/nova-metadata-0" Jan 26 17:11:56 crc kubenswrapper[4754]: I0126 17:11:56.432864 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0bb1f389-c4a6-47dd-8445-c7125779ef38-internal-tls-certs\") pod \"nova-api-0\" (UID: \"0bb1f389-c4a6-47dd-8445-c7125779ef38\") " pod="openstack/nova-api-0" Jan 26 17:11:56 crc kubenswrapper[4754]: I0126 17:11:56.433090 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4e7eefb7-e66a-4c65-bda9-7997970991f3-config-data\") pod \"nova-metadata-0\" (UID: \"4e7eefb7-e66a-4c65-bda9-7997970991f3\") " pod="openstack/nova-metadata-0" Jan 26 17:11:56 crc kubenswrapper[4754]: I0126 17:11:56.433582 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0bb1f389-c4a6-47dd-8445-c7125779ef38-public-tls-certs\") pod \"nova-api-0\" (UID: \"0bb1f389-c4a6-47dd-8445-c7125779ef38\") " pod="openstack/nova-api-0" Jan 26 17:11:56 crc kubenswrapper[4754]: I0126 17:11:56.434648 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0bb1f389-c4a6-47dd-8445-c7125779ef38-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"0bb1f389-c4a6-47dd-8445-c7125779ef38\") " pod="openstack/nova-api-0" Jan 26 17:11:56 crc kubenswrapper[4754]: I0126 17:11:56.443909 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e7eefb7-e66a-4c65-bda9-7997970991f3-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"4e7eefb7-e66a-4c65-bda9-7997970991f3\") " pod="openstack/nova-metadata-0" Jan 26 17:11:56 crc kubenswrapper[4754]: I0126 17:11:56.444533 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0bb1f389-c4a6-47dd-8445-c7125779ef38-config-data\") pod \"nova-api-0\" (UID: \"0bb1f389-c4a6-47dd-8445-c7125779ef38\") " pod="openstack/nova-api-0" Jan 26 17:11:56 crc kubenswrapper[4754]: I0126 17:11:56.447578 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zsrvl\" (UniqueName: \"kubernetes.io/projected/4e7eefb7-e66a-4c65-bda9-7997970991f3-kube-api-access-zsrvl\") pod \"nova-metadata-0\" (UID: \"4e7eefb7-e66a-4c65-bda9-7997970991f3\") " pod="openstack/nova-metadata-0" Jan 26 17:11:56 crc kubenswrapper[4754]: I0126 17:11:56.453971 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6b8k8\" (UniqueName: \"kubernetes.io/projected/0bb1f389-c4a6-47dd-8445-c7125779ef38-kube-api-access-6b8k8\") pod \"nova-api-0\" (UID: \"0bb1f389-c4a6-47dd-8445-c7125779ef38\") " pod="openstack/nova-api-0" Jan 26 17:11:56 crc kubenswrapper[4754]: I0126 17:11:56.581008 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Jan 26 17:11:56 crc kubenswrapper[4754]: I0126 17:11:56.594232 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Jan 26 17:11:56 crc kubenswrapper[4754]: I0126 17:11:56.814736 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"6f952bb4-186c-4e7c-bf01-67d8a6985319","Type":"ContainerStarted","Data":"8bba6647856ac85434f5117e6a15c08092962153437e8109735c4bd5e831da93"} Jan 26 17:11:56 crc kubenswrapper[4754]: I0126 17:11:56.832896 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.8328758240000003 podStartE2EDuration="2.832875824s" podCreationTimestamp="2026-01-26 17:11:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 17:11:56.829375808 +0000 UTC m=+1483.353556242" watchObservedRunningTime="2026-01-26 17:11:56.832875824 +0000 UTC m=+1483.357056268" Jan 26 17:11:57 crc kubenswrapper[4754]: I0126 17:11:57.151583 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Jan 26 17:11:57 crc kubenswrapper[4754]: I0126 17:11:57.161300 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Jan 26 17:11:57 crc kubenswrapper[4754]: W0126 17:11:57.164867 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4e7eefb7_e66a_4c65_bda9_7997970991f3.slice/crio-18c2349a5acd72eda72880147979cbd88cbc79977ed0590a414b92745038cfb6 WatchSource:0}: Error finding container 18c2349a5acd72eda72880147979cbd88cbc79977ed0590a414b92745038cfb6: Status 404 returned error can't find the container with id 18c2349a5acd72eda72880147979cbd88cbc79977ed0590a414b92745038cfb6 Jan 26 17:11:57 crc kubenswrapper[4754]: I0126 17:11:57.782896 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="19c7e3a1-e40c-401b-bd01-d2a50f14daf0" path="/var/lib/kubelet/pods/19c7e3a1-e40c-401b-bd01-d2a50f14daf0/volumes" Jan 26 17:11:57 crc kubenswrapper[4754]: I0126 17:11:57.783786 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="37f5b6db-8ef3-4e4f-9763-b150188c1466" path="/var/lib/kubelet/pods/37f5b6db-8ef3-4e4f-9763-b150188c1466/volumes" Jan 26 17:11:57 crc kubenswrapper[4754]: I0126 17:11:57.832150 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"4e7eefb7-e66a-4c65-bda9-7997970991f3","Type":"ContainerStarted","Data":"75af95c0ccb8d1322a40bf886a7c1c38cc48d1e6f5dacebdcb4ea871aabf5318"} Jan 26 17:11:57 crc kubenswrapper[4754]: I0126 17:11:57.832191 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"4e7eefb7-e66a-4c65-bda9-7997970991f3","Type":"ContainerStarted","Data":"a2044788fb936b475d67c803c3bc7214788b03a867d98fc6a3b94c93702ad4c7"} Jan 26 17:11:57 crc kubenswrapper[4754]: I0126 17:11:57.832202 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"4e7eefb7-e66a-4c65-bda9-7997970991f3","Type":"ContainerStarted","Data":"18c2349a5acd72eda72880147979cbd88cbc79977ed0590a414b92745038cfb6"} Jan 26 17:11:57 crc kubenswrapper[4754]: I0126 17:11:57.836271 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"0bb1f389-c4a6-47dd-8445-c7125779ef38","Type":"ContainerStarted","Data":"5ce2c0b2a2030273bd632699713b8d9c5aa3c9cccb9d1ab52deb2dbd7f37efa8"} Jan 26 17:11:57 crc kubenswrapper[4754]: I0126 17:11:57.836309 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"0bb1f389-c4a6-47dd-8445-c7125779ef38","Type":"ContainerStarted","Data":"053e5b70da7042188475ecd881e188efdc8551dd95e2f0869540ba359fcace84"} Jan 26 17:11:57 crc kubenswrapper[4754]: I0126 17:11:57.836321 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"0bb1f389-c4a6-47dd-8445-c7125779ef38","Type":"ContainerStarted","Data":"39b19216654df3ef8aff17116a08fcf3004e87dca5a6c960e0953ff049d00144"} Jan 26 17:11:57 crc kubenswrapper[4754]: I0126 17:11:57.860278 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=1.860262248 podStartE2EDuration="1.860262248s" podCreationTimestamp="2026-01-26 17:11:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 17:11:57.856276577 +0000 UTC m=+1484.380457011" watchObservedRunningTime="2026-01-26 17:11:57.860262248 +0000 UTC m=+1484.384442672" Jan 26 17:11:57 crc kubenswrapper[4754]: I0126 17:11:57.887062 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=1.887041415 podStartE2EDuration="1.887041415s" podCreationTimestamp="2026-01-26 17:11:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 17:11:57.881160813 +0000 UTC m=+1484.405341257" watchObservedRunningTime="2026-01-26 17:11:57.887041415 +0000 UTC m=+1484.411221859" Jan 26 17:12:00 crc kubenswrapper[4754]: I0126 17:12:00.189014 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Jan 26 17:12:01 crc kubenswrapper[4754]: I0126 17:12:01.581968 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Jan 26 17:12:01 crc kubenswrapper[4754]: I0126 17:12:01.582108 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Jan 26 17:12:04 crc kubenswrapper[4754]: I0126 17:12:04.928187 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Jan 26 17:12:05 crc kubenswrapper[4754]: I0126 17:12:05.188714 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Jan 26 17:12:05 crc kubenswrapper[4754]: I0126 17:12:05.215291 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Jan 26 17:12:05 crc kubenswrapper[4754]: I0126 17:12:05.938144 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Jan 26 17:12:06 crc kubenswrapper[4754]: I0126 17:12:06.582505 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Jan 26 17:12:06 crc kubenswrapper[4754]: I0126 17:12:06.582740 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Jan 26 17:12:06 crc kubenswrapper[4754]: I0126 17:12:06.595123 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Jan 26 17:12:06 crc kubenswrapper[4754]: I0126 17:12:06.595334 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Jan 26 17:12:07 crc kubenswrapper[4754]: I0126 17:12:07.129141 4754 patch_prober.go:28] interesting pod/machine-config-daemon-x65wv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 26 17:12:07 crc kubenswrapper[4754]: I0126 17:12:07.130107 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 26 17:12:07 crc kubenswrapper[4754]: I0126 17:12:07.594914 4754 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="4e7eefb7-e66a-4c65-bda9-7997970991f3" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.213:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Jan 26 17:12:07 crc kubenswrapper[4754]: I0126 17:12:07.594917 4754 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="4e7eefb7-e66a-4c65-bda9-7997970991f3" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.213:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Jan 26 17:12:07 crc kubenswrapper[4754]: I0126 17:12:07.605844 4754 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="0bb1f389-c4a6-47dd-8445-c7125779ef38" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.214:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Jan 26 17:12:07 crc kubenswrapper[4754]: I0126 17:12:07.605877 4754 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="0bb1f389-c4a6-47dd-8445-c7125779ef38" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.214:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Jan 26 17:12:16 crc kubenswrapper[4754]: I0126 17:12:16.603088 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Jan 26 17:12:16 crc kubenswrapper[4754]: I0126 17:12:16.604977 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Jan 26 17:12:16 crc kubenswrapper[4754]: I0126 17:12:16.606148 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Jan 26 17:12:16 crc kubenswrapper[4754]: I0126 17:12:16.606597 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Jan 26 17:12:16 crc kubenswrapper[4754]: I0126 17:12:16.610763 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Jan 26 17:12:16 crc kubenswrapper[4754]: I0126 17:12:16.612800 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Jan 26 17:12:16 crc kubenswrapper[4754]: I0126 17:12:16.614173 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Jan 26 17:12:16 crc kubenswrapper[4754]: I0126 17:12:16.996224 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Jan 26 17:12:17 crc kubenswrapper[4754]: I0126 17:12:17.002196 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Jan 26 17:12:17 crc kubenswrapper[4754]: I0126 17:12:17.004469 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Jan 26 17:12:25 crc kubenswrapper[4754]: I0126 17:12:25.329161 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Jan 26 17:12:26 crc kubenswrapper[4754]: I0126 17:12:26.765259 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Jan 26 17:12:29 crc kubenswrapper[4754]: I0126 17:12:29.381778 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="29be5081-5097-4cd1-b35b-192df0ce8faf" containerName="rabbitmq" containerID="cri-o://0b6e796b8cc0ee4cab5cb13a889864aa381f1854cef6c88cc1c854144900c02a" gracePeriod=604796 Jan 26 17:12:31 crc kubenswrapper[4754]: I0126 17:12:31.230730 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="48810c5e-3c40-4cdc-8bab-47efa97e76fa" containerName="rabbitmq" containerID="cri-o://c74f5c36b9d479df65eb065e767730beafd576f0ba47ddcc2bc9a3d7da87b416" gracePeriod=604796 Jan 26 17:12:35 crc kubenswrapper[4754]: I0126 17:12:35.952015 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Jan 26 17:12:36 crc kubenswrapper[4754]: I0126 17:12:36.039859 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/29be5081-5097-4cd1-b35b-192df0ce8faf-config-data\") pod \"29be5081-5097-4cd1-b35b-192df0ce8faf\" (UID: \"29be5081-5097-4cd1-b35b-192df0ce8faf\") " Jan 26 17:12:36 crc kubenswrapper[4754]: I0126 17:12:36.039975 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/29be5081-5097-4cd1-b35b-192df0ce8faf-plugins-conf\") pod \"29be5081-5097-4cd1-b35b-192df0ce8faf\" (UID: \"29be5081-5097-4cd1-b35b-192df0ce8faf\") " Jan 26 17:12:36 crc kubenswrapper[4754]: I0126 17:12:36.040003 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"29be5081-5097-4cd1-b35b-192df0ce8faf\" (UID: \"29be5081-5097-4cd1-b35b-192df0ce8faf\") " Jan 26 17:12:36 crc kubenswrapper[4754]: I0126 17:12:36.040096 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/29be5081-5097-4cd1-b35b-192df0ce8faf-rabbitmq-tls\") pod \"29be5081-5097-4cd1-b35b-192df0ce8faf\" (UID: \"29be5081-5097-4cd1-b35b-192df0ce8faf\") " Jan 26 17:12:36 crc kubenswrapper[4754]: I0126 17:12:36.040305 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/29be5081-5097-4cd1-b35b-192df0ce8faf-rabbitmq-erlang-cookie\") pod \"29be5081-5097-4cd1-b35b-192df0ce8faf\" (UID: \"29be5081-5097-4cd1-b35b-192df0ce8faf\") " Jan 26 17:12:36 crc kubenswrapper[4754]: I0126 17:12:36.040332 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/29be5081-5097-4cd1-b35b-192df0ce8faf-rabbitmq-confd\") pod \"29be5081-5097-4cd1-b35b-192df0ce8faf\" (UID: \"29be5081-5097-4cd1-b35b-192df0ce8faf\") " Jan 26 17:12:36 crc kubenswrapper[4754]: I0126 17:12:36.040358 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/29be5081-5097-4cd1-b35b-192df0ce8faf-pod-info\") pod \"29be5081-5097-4cd1-b35b-192df0ce8faf\" (UID: \"29be5081-5097-4cd1-b35b-192df0ce8faf\") " Jan 26 17:12:36 crc kubenswrapper[4754]: I0126 17:12:36.040399 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/29be5081-5097-4cd1-b35b-192df0ce8faf-server-conf\") pod \"29be5081-5097-4cd1-b35b-192df0ce8faf\" (UID: \"29be5081-5097-4cd1-b35b-192df0ce8faf\") " Jan 26 17:12:36 crc kubenswrapper[4754]: I0126 17:12:36.040473 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/29be5081-5097-4cd1-b35b-192df0ce8faf-rabbitmq-plugins\") pod \"29be5081-5097-4cd1-b35b-192df0ce8faf\" (UID: \"29be5081-5097-4cd1-b35b-192df0ce8faf\") " Jan 26 17:12:36 crc kubenswrapper[4754]: I0126 17:12:36.040524 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/29be5081-5097-4cd1-b35b-192df0ce8faf-erlang-cookie-secret\") pod \"29be5081-5097-4cd1-b35b-192df0ce8faf\" (UID: \"29be5081-5097-4cd1-b35b-192df0ce8faf\") " Jan 26 17:12:36 crc kubenswrapper[4754]: I0126 17:12:36.040639 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w6twt\" (UniqueName: \"kubernetes.io/projected/29be5081-5097-4cd1-b35b-192df0ce8faf-kube-api-access-w6twt\") pod \"29be5081-5097-4cd1-b35b-192df0ce8faf\" (UID: \"29be5081-5097-4cd1-b35b-192df0ce8faf\") " Jan 26 17:12:36 crc kubenswrapper[4754]: I0126 17:12:36.040793 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/29be5081-5097-4cd1-b35b-192df0ce8faf-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "29be5081-5097-4cd1-b35b-192df0ce8faf" (UID: "29be5081-5097-4cd1-b35b-192df0ce8faf"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:12:36 crc kubenswrapper[4754]: I0126 17:12:36.040823 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/29be5081-5097-4cd1-b35b-192df0ce8faf-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "29be5081-5097-4cd1-b35b-192df0ce8faf" (UID: "29be5081-5097-4cd1-b35b-192df0ce8faf"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:12:36 crc kubenswrapper[4754]: I0126 17:12:36.041215 4754 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/29be5081-5097-4cd1-b35b-192df0ce8faf-plugins-conf\") on node \"crc\" DevicePath \"\"" Jan 26 17:12:36 crc kubenswrapper[4754]: I0126 17:12:36.041240 4754 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/29be5081-5097-4cd1-b35b-192df0ce8faf-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Jan 26 17:12:36 crc kubenswrapper[4754]: I0126 17:12:36.041290 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/29be5081-5097-4cd1-b35b-192df0ce8faf-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "29be5081-5097-4cd1-b35b-192df0ce8faf" (UID: "29be5081-5097-4cd1-b35b-192df0ce8faf"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:12:36 crc kubenswrapper[4754]: I0126 17:12:36.048620 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/29be5081-5097-4cd1-b35b-192df0ce8faf-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "29be5081-5097-4cd1-b35b-192df0ce8faf" (UID: "29be5081-5097-4cd1-b35b-192df0ce8faf"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:12:36 crc kubenswrapper[4754]: I0126 17:12:36.050080 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/29be5081-5097-4cd1-b35b-192df0ce8faf-kube-api-access-w6twt" (OuterVolumeSpecName: "kube-api-access-w6twt") pod "29be5081-5097-4cd1-b35b-192df0ce8faf" (UID: "29be5081-5097-4cd1-b35b-192df0ce8faf"). InnerVolumeSpecName "kube-api-access-w6twt". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:12:36 crc kubenswrapper[4754]: I0126 17:12:36.051903 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage01-crc" (OuterVolumeSpecName: "persistence") pod "29be5081-5097-4cd1-b35b-192df0ce8faf" (UID: "29be5081-5097-4cd1-b35b-192df0ce8faf"). InnerVolumeSpecName "local-storage01-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Jan 26 17:12:36 crc kubenswrapper[4754]: I0126 17:12:36.054943 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/29be5081-5097-4cd1-b35b-192df0ce8faf-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "29be5081-5097-4cd1-b35b-192df0ce8faf" (UID: "29be5081-5097-4cd1-b35b-192df0ce8faf"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:12:36 crc kubenswrapper[4754]: I0126 17:12:36.056293 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/29be5081-5097-4cd1-b35b-192df0ce8faf-pod-info" (OuterVolumeSpecName: "pod-info") pod "29be5081-5097-4cd1-b35b-192df0ce8faf" (UID: "29be5081-5097-4cd1-b35b-192df0ce8faf"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Jan 26 17:12:36 crc kubenswrapper[4754]: I0126 17:12:36.146143 4754 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" " Jan 26 17:12:36 crc kubenswrapper[4754]: I0126 17:12:36.146185 4754 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/29be5081-5097-4cd1-b35b-192df0ce8faf-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Jan 26 17:12:36 crc kubenswrapper[4754]: I0126 17:12:36.146204 4754 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/29be5081-5097-4cd1-b35b-192df0ce8faf-pod-info\") on node \"crc\" DevicePath \"\"" Jan 26 17:12:36 crc kubenswrapper[4754]: I0126 17:12:36.146216 4754 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/29be5081-5097-4cd1-b35b-192df0ce8faf-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Jan 26 17:12:36 crc kubenswrapper[4754]: I0126 17:12:36.146228 4754 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/29be5081-5097-4cd1-b35b-192df0ce8faf-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Jan 26 17:12:36 crc kubenswrapper[4754]: I0126 17:12:36.146241 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w6twt\" (UniqueName: \"kubernetes.io/projected/29be5081-5097-4cd1-b35b-192df0ce8faf-kube-api-access-w6twt\") on node \"crc\" DevicePath \"\"" Jan 26 17:12:36 crc kubenswrapper[4754]: I0126 17:12:36.146339 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/29be5081-5097-4cd1-b35b-192df0ce8faf-config-data" (OuterVolumeSpecName: "config-data") pod "29be5081-5097-4cd1-b35b-192df0ce8faf" (UID: "29be5081-5097-4cd1-b35b-192df0ce8faf"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:12:36 crc kubenswrapper[4754]: I0126 17:12:36.177322 4754 generic.go:334] "Generic (PLEG): container finished" podID="29be5081-5097-4cd1-b35b-192df0ce8faf" containerID="0b6e796b8cc0ee4cab5cb13a889864aa381f1854cef6c88cc1c854144900c02a" exitCode=0 Jan 26 17:12:36 crc kubenswrapper[4754]: I0126 17:12:36.177375 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"29be5081-5097-4cd1-b35b-192df0ce8faf","Type":"ContainerDied","Data":"0b6e796b8cc0ee4cab5cb13a889864aa381f1854cef6c88cc1c854144900c02a"} Jan 26 17:12:36 crc kubenswrapper[4754]: I0126 17:12:36.177407 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"29be5081-5097-4cd1-b35b-192df0ce8faf","Type":"ContainerDied","Data":"24b4abcf05c2fd2797980e930d42b56f348c915f39e805757ce1375a89a43382"} Jan 26 17:12:36 crc kubenswrapper[4754]: I0126 17:12:36.177428 4754 scope.go:117] "RemoveContainer" containerID="0b6e796b8cc0ee4cab5cb13a889864aa381f1854cef6c88cc1c854144900c02a" Jan 26 17:12:36 crc kubenswrapper[4754]: I0126 17:12:36.177592 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Jan 26 17:12:36 crc kubenswrapper[4754]: I0126 17:12:36.190599 4754 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage01-crc" (UniqueName: "kubernetes.io/local-volume/local-storage01-crc") on node "crc" Jan 26 17:12:36 crc kubenswrapper[4754]: I0126 17:12:36.201226 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/29be5081-5097-4cd1-b35b-192df0ce8faf-server-conf" (OuterVolumeSpecName: "server-conf") pod "29be5081-5097-4cd1-b35b-192df0ce8faf" (UID: "29be5081-5097-4cd1-b35b-192df0ce8faf"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:12:36 crc kubenswrapper[4754]: I0126 17:12:36.208702 4754 scope.go:117] "RemoveContainer" containerID="3f67cb54c8ac7414b3b3fe5d6bc7d975fa28e160e0a72a1773a4721747eb4df3" Jan 26 17:12:36 crc kubenswrapper[4754]: I0126 17:12:36.241592 4754 scope.go:117] "RemoveContainer" containerID="0b6e796b8cc0ee4cab5cb13a889864aa381f1854cef6c88cc1c854144900c02a" Jan 26 17:12:36 crc kubenswrapper[4754]: E0126 17:12:36.242784 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0b6e796b8cc0ee4cab5cb13a889864aa381f1854cef6c88cc1c854144900c02a\": container with ID starting with 0b6e796b8cc0ee4cab5cb13a889864aa381f1854cef6c88cc1c854144900c02a not found: ID does not exist" containerID="0b6e796b8cc0ee4cab5cb13a889864aa381f1854cef6c88cc1c854144900c02a" Jan 26 17:12:36 crc kubenswrapper[4754]: I0126 17:12:36.242821 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0b6e796b8cc0ee4cab5cb13a889864aa381f1854cef6c88cc1c854144900c02a"} err="failed to get container status \"0b6e796b8cc0ee4cab5cb13a889864aa381f1854cef6c88cc1c854144900c02a\": rpc error: code = NotFound desc = could not find container \"0b6e796b8cc0ee4cab5cb13a889864aa381f1854cef6c88cc1c854144900c02a\": container with ID starting with 0b6e796b8cc0ee4cab5cb13a889864aa381f1854cef6c88cc1c854144900c02a not found: ID does not exist" Jan 26 17:12:36 crc kubenswrapper[4754]: I0126 17:12:36.242845 4754 scope.go:117] "RemoveContainer" containerID="3f67cb54c8ac7414b3b3fe5d6bc7d975fa28e160e0a72a1773a4721747eb4df3" Jan 26 17:12:36 crc kubenswrapper[4754]: E0126 17:12:36.244813 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3f67cb54c8ac7414b3b3fe5d6bc7d975fa28e160e0a72a1773a4721747eb4df3\": container with ID starting with 3f67cb54c8ac7414b3b3fe5d6bc7d975fa28e160e0a72a1773a4721747eb4df3 not found: ID does not exist" containerID="3f67cb54c8ac7414b3b3fe5d6bc7d975fa28e160e0a72a1773a4721747eb4df3" Jan 26 17:12:36 crc kubenswrapper[4754]: I0126 17:12:36.244847 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3f67cb54c8ac7414b3b3fe5d6bc7d975fa28e160e0a72a1773a4721747eb4df3"} err="failed to get container status \"3f67cb54c8ac7414b3b3fe5d6bc7d975fa28e160e0a72a1773a4721747eb4df3\": rpc error: code = NotFound desc = could not find container \"3f67cb54c8ac7414b3b3fe5d6bc7d975fa28e160e0a72a1773a4721747eb4df3\": container with ID starting with 3f67cb54c8ac7414b3b3fe5d6bc7d975fa28e160e0a72a1773a4721747eb4df3 not found: ID does not exist" Jan 26 17:12:36 crc kubenswrapper[4754]: I0126 17:12:36.250029 4754 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/29be5081-5097-4cd1-b35b-192df0ce8faf-config-data\") on node \"crc\" DevicePath \"\"" Jan 26 17:12:36 crc kubenswrapper[4754]: I0126 17:12:36.250081 4754 reconciler_common.go:293] "Volume detached for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" DevicePath \"\"" Jan 26 17:12:36 crc kubenswrapper[4754]: I0126 17:12:36.250096 4754 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/29be5081-5097-4cd1-b35b-192df0ce8faf-server-conf\") on node \"crc\" DevicePath \"\"" Jan 26 17:12:36 crc kubenswrapper[4754]: I0126 17:12:36.289354 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/29be5081-5097-4cd1-b35b-192df0ce8faf-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "29be5081-5097-4cd1-b35b-192df0ce8faf" (UID: "29be5081-5097-4cd1-b35b-192df0ce8faf"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:12:36 crc kubenswrapper[4754]: I0126 17:12:36.352035 4754 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/29be5081-5097-4cd1-b35b-192df0ce8faf-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Jan 26 17:12:36 crc kubenswrapper[4754]: I0126 17:12:36.513635 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Jan 26 17:12:36 crc kubenswrapper[4754]: I0126 17:12:36.524113 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Jan 26 17:12:36 crc kubenswrapper[4754]: I0126 17:12:36.546116 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Jan 26 17:12:36 crc kubenswrapper[4754]: E0126 17:12:36.546553 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="29be5081-5097-4cd1-b35b-192df0ce8faf" containerName="rabbitmq" Jan 26 17:12:36 crc kubenswrapper[4754]: I0126 17:12:36.546574 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="29be5081-5097-4cd1-b35b-192df0ce8faf" containerName="rabbitmq" Jan 26 17:12:36 crc kubenswrapper[4754]: E0126 17:12:36.546606 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="29be5081-5097-4cd1-b35b-192df0ce8faf" containerName="setup-container" Jan 26 17:12:36 crc kubenswrapper[4754]: I0126 17:12:36.546615 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="29be5081-5097-4cd1-b35b-192df0ce8faf" containerName="setup-container" Jan 26 17:12:36 crc kubenswrapper[4754]: I0126 17:12:36.546915 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="29be5081-5097-4cd1-b35b-192df0ce8faf" containerName="rabbitmq" Jan 26 17:12:36 crc kubenswrapper[4754]: I0126 17:12:36.548034 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Jan 26 17:12:36 crc kubenswrapper[4754]: I0126 17:12:36.550356 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Jan 26 17:12:36 crc kubenswrapper[4754]: I0126 17:12:36.551221 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Jan 26 17:12:36 crc kubenswrapper[4754]: I0126 17:12:36.551393 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Jan 26 17:12:36 crc kubenswrapper[4754]: I0126 17:12:36.551562 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-xwmcb" Jan 26 17:12:36 crc kubenswrapper[4754]: I0126 17:12:36.551762 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Jan 26 17:12:36 crc kubenswrapper[4754]: I0126 17:12:36.551940 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Jan 26 17:12:36 crc kubenswrapper[4754]: I0126 17:12:36.564696 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Jan 26 17:12:36 crc kubenswrapper[4754]: I0126 17:12:36.565577 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Jan 26 17:12:36 crc kubenswrapper[4754]: I0126 17:12:36.668828 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6e2da069-1c54-4801-a91c-241b80e8d17b-config-data\") pod \"rabbitmq-server-0\" (UID: \"6e2da069-1c54-4801-a91c-241b80e8d17b\") " pod="openstack/rabbitmq-server-0" Jan 26 17:12:36 crc kubenswrapper[4754]: I0126 17:12:36.668885 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cwr6n\" (UniqueName: \"kubernetes.io/projected/6e2da069-1c54-4801-a91c-241b80e8d17b-kube-api-access-cwr6n\") pod \"rabbitmq-server-0\" (UID: \"6e2da069-1c54-4801-a91c-241b80e8d17b\") " pod="openstack/rabbitmq-server-0" Jan 26 17:12:36 crc kubenswrapper[4754]: I0126 17:12:36.668927 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/6e2da069-1c54-4801-a91c-241b80e8d17b-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"6e2da069-1c54-4801-a91c-241b80e8d17b\") " pod="openstack/rabbitmq-server-0" Jan 26 17:12:36 crc kubenswrapper[4754]: I0126 17:12:36.668956 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/6e2da069-1c54-4801-a91c-241b80e8d17b-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"6e2da069-1c54-4801-a91c-241b80e8d17b\") " pod="openstack/rabbitmq-server-0" Jan 26 17:12:36 crc kubenswrapper[4754]: I0126 17:12:36.669000 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/6e2da069-1c54-4801-a91c-241b80e8d17b-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"6e2da069-1c54-4801-a91c-241b80e8d17b\") " pod="openstack/rabbitmq-server-0" Jan 26 17:12:36 crc kubenswrapper[4754]: I0126 17:12:36.669061 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/6e2da069-1c54-4801-a91c-241b80e8d17b-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"6e2da069-1c54-4801-a91c-241b80e8d17b\") " pod="openstack/rabbitmq-server-0" Jan 26 17:12:36 crc kubenswrapper[4754]: I0126 17:12:36.669085 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"6e2da069-1c54-4801-a91c-241b80e8d17b\") " pod="openstack/rabbitmq-server-0" Jan 26 17:12:36 crc kubenswrapper[4754]: I0126 17:12:36.669116 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/6e2da069-1c54-4801-a91c-241b80e8d17b-server-conf\") pod \"rabbitmq-server-0\" (UID: \"6e2da069-1c54-4801-a91c-241b80e8d17b\") " pod="openstack/rabbitmq-server-0" Jan 26 17:12:36 crc kubenswrapper[4754]: I0126 17:12:36.669148 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/6e2da069-1c54-4801-a91c-241b80e8d17b-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"6e2da069-1c54-4801-a91c-241b80e8d17b\") " pod="openstack/rabbitmq-server-0" Jan 26 17:12:36 crc kubenswrapper[4754]: I0126 17:12:36.669175 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/6e2da069-1c54-4801-a91c-241b80e8d17b-pod-info\") pod \"rabbitmq-server-0\" (UID: \"6e2da069-1c54-4801-a91c-241b80e8d17b\") " pod="openstack/rabbitmq-server-0" Jan 26 17:12:36 crc kubenswrapper[4754]: I0126 17:12:36.669198 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/6e2da069-1c54-4801-a91c-241b80e8d17b-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"6e2da069-1c54-4801-a91c-241b80e8d17b\") " pod="openstack/rabbitmq-server-0" Jan 26 17:12:36 crc kubenswrapper[4754]: I0126 17:12:36.770377 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6e2da069-1c54-4801-a91c-241b80e8d17b-config-data\") pod \"rabbitmq-server-0\" (UID: \"6e2da069-1c54-4801-a91c-241b80e8d17b\") " pod="openstack/rabbitmq-server-0" Jan 26 17:12:36 crc kubenswrapper[4754]: I0126 17:12:36.770420 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cwr6n\" (UniqueName: \"kubernetes.io/projected/6e2da069-1c54-4801-a91c-241b80e8d17b-kube-api-access-cwr6n\") pod \"rabbitmq-server-0\" (UID: \"6e2da069-1c54-4801-a91c-241b80e8d17b\") " pod="openstack/rabbitmq-server-0" Jan 26 17:12:36 crc kubenswrapper[4754]: I0126 17:12:36.770454 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/6e2da069-1c54-4801-a91c-241b80e8d17b-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"6e2da069-1c54-4801-a91c-241b80e8d17b\") " pod="openstack/rabbitmq-server-0" Jan 26 17:12:36 crc kubenswrapper[4754]: I0126 17:12:36.770480 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/6e2da069-1c54-4801-a91c-241b80e8d17b-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"6e2da069-1c54-4801-a91c-241b80e8d17b\") " pod="openstack/rabbitmq-server-0" Jan 26 17:12:36 crc kubenswrapper[4754]: I0126 17:12:36.770529 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/6e2da069-1c54-4801-a91c-241b80e8d17b-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"6e2da069-1c54-4801-a91c-241b80e8d17b\") " pod="openstack/rabbitmq-server-0" Jan 26 17:12:36 crc kubenswrapper[4754]: I0126 17:12:36.770595 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/6e2da069-1c54-4801-a91c-241b80e8d17b-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"6e2da069-1c54-4801-a91c-241b80e8d17b\") " pod="openstack/rabbitmq-server-0" Jan 26 17:12:36 crc kubenswrapper[4754]: I0126 17:12:36.770617 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"6e2da069-1c54-4801-a91c-241b80e8d17b\") " pod="openstack/rabbitmq-server-0" Jan 26 17:12:36 crc kubenswrapper[4754]: I0126 17:12:36.770649 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/6e2da069-1c54-4801-a91c-241b80e8d17b-server-conf\") pod \"rabbitmq-server-0\" (UID: \"6e2da069-1c54-4801-a91c-241b80e8d17b\") " pod="openstack/rabbitmq-server-0" Jan 26 17:12:36 crc kubenswrapper[4754]: I0126 17:12:36.770713 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/6e2da069-1c54-4801-a91c-241b80e8d17b-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"6e2da069-1c54-4801-a91c-241b80e8d17b\") " pod="openstack/rabbitmq-server-0" Jan 26 17:12:36 crc kubenswrapper[4754]: I0126 17:12:36.770748 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/6e2da069-1c54-4801-a91c-241b80e8d17b-pod-info\") pod \"rabbitmq-server-0\" (UID: \"6e2da069-1c54-4801-a91c-241b80e8d17b\") " pod="openstack/rabbitmq-server-0" Jan 26 17:12:36 crc kubenswrapper[4754]: I0126 17:12:36.770775 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/6e2da069-1c54-4801-a91c-241b80e8d17b-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"6e2da069-1c54-4801-a91c-241b80e8d17b\") " pod="openstack/rabbitmq-server-0" Jan 26 17:12:36 crc kubenswrapper[4754]: I0126 17:12:36.771557 4754 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"6e2da069-1c54-4801-a91c-241b80e8d17b\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/rabbitmq-server-0" Jan 26 17:12:36 crc kubenswrapper[4754]: I0126 17:12:36.771576 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/6e2da069-1c54-4801-a91c-241b80e8d17b-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"6e2da069-1c54-4801-a91c-241b80e8d17b\") " pod="openstack/rabbitmq-server-0" Jan 26 17:12:36 crc kubenswrapper[4754]: I0126 17:12:36.771848 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/6e2da069-1c54-4801-a91c-241b80e8d17b-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"6e2da069-1c54-4801-a91c-241b80e8d17b\") " pod="openstack/rabbitmq-server-0" Jan 26 17:12:36 crc kubenswrapper[4754]: I0126 17:12:36.771966 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/6e2da069-1c54-4801-a91c-241b80e8d17b-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"6e2da069-1c54-4801-a91c-241b80e8d17b\") " pod="openstack/rabbitmq-server-0" Jan 26 17:12:36 crc kubenswrapper[4754]: I0126 17:12:36.772035 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6e2da069-1c54-4801-a91c-241b80e8d17b-config-data\") pod \"rabbitmq-server-0\" (UID: \"6e2da069-1c54-4801-a91c-241b80e8d17b\") " pod="openstack/rabbitmq-server-0" Jan 26 17:12:36 crc kubenswrapper[4754]: I0126 17:12:36.772351 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/6e2da069-1c54-4801-a91c-241b80e8d17b-server-conf\") pod \"rabbitmq-server-0\" (UID: \"6e2da069-1c54-4801-a91c-241b80e8d17b\") " pod="openstack/rabbitmq-server-0" Jan 26 17:12:36 crc kubenswrapper[4754]: I0126 17:12:36.776936 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/6e2da069-1c54-4801-a91c-241b80e8d17b-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"6e2da069-1c54-4801-a91c-241b80e8d17b\") " pod="openstack/rabbitmq-server-0" Jan 26 17:12:36 crc kubenswrapper[4754]: I0126 17:12:36.776936 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/6e2da069-1c54-4801-a91c-241b80e8d17b-pod-info\") pod \"rabbitmq-server-0\" (UID: \"6e2da069-1c54-4801-a91c-241b80e8d17b\") " pod="openstack/rabbitmq-server-0" Jan 26 17:12:36 crc kubenswrapper[4754]: I0126 17:12:36.777021 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/6e2da069-1c54-4801-a91c-241b80e8d17b-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"6e2da069-1c54-4801-a91c-241b80e8d17b\") " pod="openstack/rabbitmq-server-0" Jan 26 17:12:36 crc kubenswrapper[4754]: I0126 17:12:36.777633 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/6e2da069-1c54-4801-a91c-241b80e8d17b-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"6e2da069-1c54-4801-a91c-241b80e8d17b\") " pod="openstack/rabbitmq-server-0" Jan 26 17:12:36 crc kubenswrapper[4754]: I0126 17:12:36.789473 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cwr6n\" (UniqueName: \"kubernetes.io/projected/6e2da069-1c54-4801-a91c-241b80e8d17b-kube-api-access-cwr6n\") pod \"rabbitmq-server-0\" (UID: \"6e2da069-1c54-4801-a91c-241b80e8d17b\") " pod="openstack/rabbitmq-server-0" Jan 26 17:12:36 crc kubenswrapper[4754]: I0126 17:12:36.803889 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"6e2da069-1c54-4801-a91c-241b80e8d17b\") " pod="openstack/rabbitmq-server-0" Jan 26 17:12:36 crc kubenswrapper[4754]: I0126 17:12:36.949213 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Jan 26 17:12:37 crc kubenswrapper[4754]: I0126 17:12:37.129801 4754 patch_prober.go:28] interesting pod/machine-config-daemon-x65wv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 26 17:12:37 crc kubenswrapper[4754]: I0126 17:12:37.130136 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 26 17:12:37 crc kubenswrapper[4754]: I0126 17:12:37.451966 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Jan 26 17:12:37 crc kubenswrapper[4754]: I0126 17:12:37.778230 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="29be5081-5097-4cd1-b35b-192df0ce8faf" path="/var/lib/kubelet/pods/29be5081-5097-4cd1-b35b-192df0ce8faf/volumes" Jan 26 17:12:37 crc kubenswrapper[4754]: I0126 17:12:37.878344 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Jan 26 17:12:37 crc kubenswrapper[4754]: I0126 17:12:37.996112 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p6m5j\" (UniqueName: \"kubernetes.io/projected/48810c5e-3c40-4cdc-8bab-47efa97e76fa-kube-api-access-p6m5j\") pod \"48810c5e-3c40-4cdc-8bab-47efa97e76fa\" (UID: \"48810c5e-3c40-4cdc-8bab-47efa97e76fa\") " Jan 26 17:12:37 crc kubenswrapper[4754]: I0126 17:12:37.996156 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/48810c5e-3c40-4cdc-8bab-47efa97e76fa-server-conf\") pod \"48810c5e-3c40-4cdc-8bab-47efa97e76fa\" (UID: \"48810c5e-3c40-4cdc-8bab-47efa97e76fa\") " Jan 26 17:12:37 crc kubenswrapper[4754]: I0126 17:12:37.996185 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/48810c5e-3c40-4cdc-8bab-47efa97e76fa-rabbitmq-erlang-cookie\") pod \"48810c5e-3c40-4cdc-8bab-47efa97e76fa\" (UID: \"48810c5e-3c40-4cdc-8bab-47efa97e76fa\") " Jan 26 17:12:37 crc kubenswrapper[4754]: I0126 17:12:37.996229 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/48810c5e-3c40-4cdc-8bab-47efa97e76fa-erlang-cookie-secret\") pod \"48810c5e-3c40-4cdc-8bab-47efa97e76fa\" (UID: \"48810c5e-3c40-4cdc-8bab-47efa97e76fa\") " Jan 26 17:12:37 crc kubenswrapper[4754]: I0126 17:12:37.996277 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/48810c5e-3c40-4cdc-8bab-47efa97e76fa-plugins-conf\") pod \"48810c5e-3c40-4cdc-8bab-47efa97e76fa\" (UID: \"48810c5e-3c40-4cdc-8bab-47efa97e76fa\") " Jan 26 17:12:37 crc kubenswrapper[4754]: I0126 17:12:37.996343 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/48810c5e-3c40-4cdc-8bab-47efa97e76fa-pod-info\") pod \"48810c5e-3c40-4cdc-8bab-47efa97e76fa\" (UID: \"48810c5e-3c40-4cdc-8bab-47efa97e76fa\") " Jan 26 17:12:37 crc kubenswrapper[4754]: I0126 17:12:37.996371 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/48810c5e-3c40-4cdc-8bab-47efa97e76fa-config-data\") pod \"48810c5e-3c40-4cdc-8bab-47efa97e76fa\" (UID: \"48810c5e-3c40-4cdc-8bab-47efa97e76fa\") " Jan 26 17:12:37 crc kubenswrapper[4754]: I0126 17:12:37.996436 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/48810c5e-3c40-4cdc-8bab-47efa97e76fa-rabbitmq-plugins\") pod \"48810c5e-3c40-4cdc-8bab-47efa97e76fa\" (UID: \"48810c5e-3c40-4cdc-8bab-47efa97e76fa\") " Jan 26 17:12:37 crc kubenswrapper[4754]: I0126 17:12:37.996464 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"48810c5e-3c40-4cdc-8bab-47efa97e76fa\" (UID: \"48810c5e-3c40-4cdc-8bab-47efa97e76fa\") " Jan 26 17:12:37 crc kubenswrapper[4754]: I0126 17:12:37.996481 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/48810c5e-3c40-4cdc-8bab-47efa97e76fa-rabbitmq-confd\") pod \"48810c5e-3c40-4cdc-8bab-47efa97e76fa\" (UID: \"48810c5e-3c40-4cdc-8bab-47efa97e76fa\") " Jan 26 17:12:37 crc kubenswrapper[4754]: I0126 17:12:37.996506 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/48810c5e-3c40-4cdc-8bab-47efa97e76fa-rabbitmq-tls\") pod \"48810c5e-3c40-4cdc-8bab-47efa97e76fa\" (UID: \"48810c5e-3c40-4cdc-8bab-47efa97e76fa\") " Jan 26 17:12:37 crc kubenswrapper[4754]: I0126 17:12:37.996813 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/48810c5e-3c40-4cdc-8bab-47efa97e76fa-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "48810c5e-3c40-4cdc-8bab-47efa97e76fa" (UID: "48810c5e-3c40-4cdc-8bab-47efa97e76fa"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:12:37 crc kubenswrapper[4754]: I0126 17:12:37.996868 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/48810c5e-3c40-4cdc-8bab-47efa97e76fa-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "48810c5e-3c40-4cdc-8bab-47efa97e76fa" (UID: "48810c5e-3c40-4cdc-8bab-47efa97e76fa"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:12:37 crc kubenswrapper[4754]: I0126 17:12:37.997050 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/48810c5e-3c40-4cdc-8bab-47efa97e76fa-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "48810c5e-3c40-4cdc-8bab-47efa97e76fa" (UID: "48810c5e-3c40-4cdc-8bab-47efa97e76fa"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:12:38 crc kubenswrapper[4754]: I0126 17:12:38.000106 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/48810c5e-3c40-4cdc-8bab-47efa97e76fa-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "48810c5e-3c40-4cdc-8bab-47efa97e76fa" (UID: "48810c5e-3c40-4cdc-8bab-47efa97e76fa"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:12:38 crc kubenswrapper[4754]: I0126 17:12:38.001159 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/48810c5e-3c40-4cdc-8bab-47efa97e76fa-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "48810c5e-3c40-4cdc-8bab-47efa97e76fa" (UID: "48810c5e-3c40-4cdc-8bab-47efa97e76fa"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:12:38 crc kubenswrapper[4754]: I0126 17:12:38.003959 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage07-crc" (OuterVolumeSpecName: "persistence") pod "48810c5e-3c40-4cdc-8bab-47efa97e76fa" (UID: "48810c5e-3c40-4cdc-8bab-47efa97e76fa"). InnerVolumeSpecName "local-storage07-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Jan 26 17:12:38 crc kubenswrapper[4754]: I0126 17:12:38.003959 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/48810c5e-3c40-4cdc-8bab-47efa97e76fa-pod-info" (OuterVolumeSpecName: "pod-info") pod "48810c5e-3c40-4cdc-8bab-47efa97e76fa" (UID: "48810c5e-3c40-4cdc-8bab-47efa97e76fa"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Jan 26 17:12:38 crc kubenswrapper[4754]: I0126 17:12:38.004041 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/48810c5e-3c40-4cdc-8bab-47efa97e76fa-kube-api-access-p6m5j" (OuterVolumeSpecName: "kube-api-access-p6m5j") pod "48810c5e-3c40-4cdc-8bab-47efa97e76fa" (UID: "48810c5e-3c40-4cdc-8bab-47efa97e76fa"). InnerVolumeSpecName "kube-api-access-p6m5j". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:12:38 crc kubenswrapper[4754]: I0126 17:12:38.033171 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/48810c5e-3c40-4cdc-8bab-47efa97e76fa-config-data" (OuterVolumeSpecName: "config-data") pod "48810c5e-3c40-4cdc-8bab-47efa97e76fa" (UID: "48810c5e-3c40-4cdc-8bab-47efa97e76fa"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:12:38 crc kubenswrapper[4754]: I0126 17:12:38.088821 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/48810c5e-3c40-4cdc-8bab-47efa97e76fa-server-conf" (OuterVolumeSpecName: "server-conf") pod "48810c5e-3c40-4cdc-8bab-47efa97e76fa" (UID: "48810c5e-3c40-4cdc-8bab-47efa97e76fa"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:12:38 crc kubenswrapper[4754]: I0126 17:12:38.098391 4754 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/48810c5e-3c40-4cdc-8bab-47efa97e76fa-plugins-conf\") on node \"crc\" DevicePath \"\"" Jan 26 17:12:38 crc kubenswrapper[4754]: I0126 17:12:38.098414 4754 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/48810c5e-3c40-4cdc-8bab-47efa97e76fa-pod-info\") on node \"crc\" DevicePath \"\"" Jan 26 17:12:38 crc kubenswrapper[4754]: I0126 17:12:38.098423 4754 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/48810c5e-3c40-4cdc-8bab-47efa97e76fa-config-data\") on node \"crc\" DevicePath \"\"" Jan 26 17:12:38 crc kubenswrapper[4754]: I0126 17:12:38.098431 4754 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/48810c5e-3c40-4cdc-8bab-47efa97e76fa-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Jan 26 17:12:38 crc kubenswrapper[4754]: I0126 17:12:38.098470 4754 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" " Jan 26 17:12:38 crc kubenswrapper[4754]: I0126 17:12:38.098479 4754 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/48810c5e-3c40-4cdc-8bab-47efa97e76fa-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Jan 26 17:12:38 crc kubenswrapper[4754]: I0126 17:12:38.098501 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p6m5j\" (UniqueName: \"kubernetes.io/projected/48810c5e-3c40-4cdc-8bab-47efa97e76fa-kube-api-access-p6m5j\") on node \"crc\" DevicePath \"\"" Jan 26 17:12:38 crc kubenswrapper[4754]: I0126 17:12:38.098511 4754 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/48810c5e-3c40-4cdc-8bab-47efa97e76fa-server-conf\") on node \"crc\" DevicePath \"\"" Jan 26 17:12:38 crc kubenswrapper[4754]: I0126 17:12:38.098538 4754 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/48810c5e-3c40-4cdc-8bab-47efa97e76fa-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Jan 26 17:12:38 crc kubenswrapper[4754]: I0126 17:12:38.098546 4754 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/48810c5e-3c40-4cdc-8bab-47efa97e76fa-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Jan 26 17:12:38 crc kubenswrapper[4754]: I0126 17:12:38.119753 4754 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage07-crc" (UniqueName: "kubernetes.io/local-volume/local-storage07-crc") on node "crc" Jan 26 17:12:38 crc kubenswrapper[4754]: I0126 17:12:38.201632 4754 reconciler_common.go:293] "Volume detached for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" DevicePath \"\"" Jan 26 17:12:38 crc kubenswrapper[4754]: I0126 17:12:38.201875 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"6e2da069-1c54-4801-a91c-241b80e8d17b","Type":"ContainerStarted","Data":"840f4167b97548a17d3a99d16e3f1ab5201deb1fa1e062afed4242bfb0628cf4"} Jan 26 17:12:38 crc kubenswrapper[4754]: I0126 17:12:38.204958 4754 generic.go:334] "Generic (PLEG): container finished" podID="48810c5e-3c40-4cdc-8bab-47efa97e76fa" containerID="c74f5c36b9d479df65eb065e767730beafd576f0ba47ddcc2bc9a3d7da87b416" exitCode=0 Jan 26 17:12:38 crc kubenswrapper[4754]: I0126 17:12:38.205000 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"48810c5e-3c40-4cdc-8bab-47efa97e76fa","Type":"ContainerDied","Data":"c74f5c36b9d479df65eb065e767730beafd576f0ba47ddcc2bc9a3d7da87b416"} Jan 26 17:12:38 crc kubenswrapper[4754]: I0126 17:12:38.205025 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"48810c5e-3c40-4cdc-8bab-47efa97e76fa","Type":"ContainerDied","Data":"176b2343d92554d1e43a227e02c03aca3ee8b66ec8be27ba113fd85492164c81"} Jan 26 17:12:38 crc kubenswrapper[4754]: I0126 17:12:38.205048 4754 scope.go:117] "RemoveContainer" containerID="c74f5c36b9d479df65eb065e767730beafd576f0ba47ddcc2bc9a3d7da87b416" Jan 26 17:12:38 crc kubenswrapper[4754]: I0126 17:12:38.205382 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Jan 26 17:12:38 crc kubenswrapper[4754]: I0126 17:12:38.223482 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/48810c5e-3c40-4cdc-8bab-47efa97e76fa-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "48810c5e-3c40-4cdc-8bab-47efa97e76fa" (UID: "48810c5e-3c40-4cdc-8bab-47efa97e76fa"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:12:38 crc kubenswrapper[4754]: I0126 17:12:38.232007 4754 scope.go:117] "RemoveContainer" containerID="397f275bbe0a6c7fe3d67ba536948ee64336fe00addd514ba132a01fca3d4cb6" Jan 26 17:12:38 crc kubenswrapper[4754]: I0126 17:12:38.250513 4754 scope.go:117] "RemoveContainer" containerID="c74f5c36b9d479df65eb065e767730beafd576f0ba47ddcc2bc9a3d7da87b416" Jan 26 17:12:38 crc kubenswrapper[4754]: E0126 17:12:38.251067 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c74f5c36b9d479df65eb065e767730beafd576f0ba47ddcc2bc9a3d7da87b416\": container with ID starting with c74f5c36b9d479df65eb065e767730beafd576f0ba47ddcc2bc9a3d7da87b416 not found: ID does not exist" containerID="c74f5c36b9d479df65eb065e767730beafd576f0ba47ddcc2bc9a3d7da87b416" Jan 26 17:12:38 crc kubenswrapper[4754]: I0126 17:12:38.251099 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c74f5c36b9d479df65eb065e767730beafd576f0ba47ddcc2bc9a3d7da87b416"} err="failed to get container status \"c74f5c36b9d479df65eb065e767730beafd576f0ba47ddcc2bc9a3d7da87b416\": rpc error: code = NotFound desc = could not find container \"c74f5c36b9d479df65eb065e767730beafd576f0ba47ddcc2bc9a3d7da87b416\": container with ID starting with c74f5c36b9d479df65eb065e767730beafd576f0ba47ddcc2bc9a3d7da87b416 not found: ID does not exist" Jan 26 17:12:38 crc kubenswrapper[4754]: I0126 17:12:38.251120 4754 scope.go:117] "RemoveContainer" containerID="397f275bbe0a6c7fe3d67ba536948ee64336fe00addd514ba132a01fca3d4cb6" Jan 26 17:12:38 crc kubenswrapper[4754]: E0126 17:12:38.251575 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"397f275bbe0a6c7fe3d67ba536948ee64336fe00addd514ba132a01fca3d4cb6\": container with ID starting with 397f275bbe0a6c7fe3d67ba536948ee64336fe00addd514ba132a01fca3d4cb6 not found: ID does not exist" containerID="397f275bbe0a6c7fe3d67ba536948ee64336fe00addd514ba132a01fca3d4cb6" Jan 26 17:12:38 crc kubenswrapper[4754]: I0126 17:12:38.251598 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"397f275bbe0a6c7fe3d67ba536948ee64336fe00addd514ba132a01fca3d4cb6"} err="failed to get container status \"397f275bbe0a6c7fe3d67ba536948ee64336fe00addd514ba132a01fca3d4cb6\": rpc error: code = NotFound desc = could not find container \"397f275bbe0a6c7fe3d67ba536948ee64336fe00addd514ba132a01fca3d4cb6\": container with ID starting with 397f275bbe0a6c7fe3d67ba536948ee64336fe00addd514ba132a01fca3d4cb6 not found: ID does not exist" Jan 26 17:12:38 crc kubenswrapper[4754]: I0126 17:12:38.302857 4754 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/48810c5e-3c40-4cdc-8bab-47efa97e76fa-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Jan 26 17:12:38 crc kubenswrapper[4754]: I0126 17:12:38.546299 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Jan 26 17:12:38 crc kubenswrapper[4754]: I0126 17:12:38.555639 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Jan 26 17:12:38 crc kubenswrapper[4754]: I0126 17:12:38.575647 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Jan 26 17:12:38 crc kubenswrapper[4754]: E0126 17:12:38.576042 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48810c5e-3c40-4cdc-8bab-47efa97e76fa" containerName="rabbitmq" Jan 26 17:12:38 crc kubenswrapper[4754]: I0126 17:12:38.576058 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="48810c5e-3c40-4cdc-8bab-47efa97e76fa" containerName="rabbitmq" Jan 26 17:12:38 crc kubenswrapper[4754]: E0126 17:12:38.576075 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48810c5e-3c40-4cdc-8bab-47efa97e76fa" containerName="setup-container" Jan 26 17:12:38 crc kubenswrapper[4754]: I0126 17:12:38.576082 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="48810c5e-3c40-4cdc-8bab-47efa97e76fa" containerName="setup-container" Jan 26 17:12:38 crc kubenswrapper[4754]: I0126 17:12:38.576253 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="48810c5e-3c40-4cdc-8bab-47efa97e76fa" containerName="rabbitmq" Jan 26 17:12:38 crc kubenswrapper[4754]: I0126 17:12:38.577269 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Jan 26 17:12:38 crc kubenswrapper[4754]: I0126 17:12:38.579776 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Jan 26 17:12:38 crc kubenswrapper[4754]: I0126 17:12:38.580381 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Jan 26 17:12:38 crc kubenswrapper[4754]: I0126 17:12:38.581037 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Jan 26 17:12:38 crc kubenswrapper[4754]: I0126 17:12:38.581220 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Jan 26 17:12:38 crc kubenswrapper[4754]: I0126 17:12:38.581245 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Jan 26 17:12:38 crc kubenswrapper[4754]: I0126 17:12:38.581257 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-mv2rb" Jan 26 17:12:38 crc kubenswrapper[4754]: I0126 17:12:38.581369 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Jan 26 17:12:38 crc kubenswrapper[4754]: I0126 17:12:38.598685 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Jan 26 17:12:38 crc kubenswrapper[4754]: I0126 17:12:38.709752 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/9acb97f6-82ac-4891-96dc-43a85f9c4e7c-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"9acb97f6-82ac-4891-96dc-43a85f9c4e7c\") " pod="openstack/rabbitmq-cell1-server-0" Jan 26 17:12:38 crc kubenswrapper[4754]: I0126 17:12:38.709982 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9rv69\" (UniqueName: \"kubernetes.io/projected/9acb97f6-82ac-4891-96dc-43a85f9c4e7c-kube-api-access-9rv69\") pod \"rabbitmq-cell1-server-0\" (UID: \"9acb97f6-82ac-4891-96dc-43a85f9c4e7c\") " pod="openstack/rabbitmq-cell1-server-0" Jan 26 17:12:38 crc kubenswrapper[4754]: I0126 17:12:38.710070 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/9acb97f6-82ac-4891-96dc-43a85f9c4e7c-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"9acb97f6-82ac-4891-96dc-43a85f9c4e7c\") " pod="openstack/rabbitmq-cell1-server-0" Jan 26 17:12:38 crc kubenswrapper[4754]: I0126 17:12:38.710153 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/9acb97f6-82ac-4891-96dc-43a85f9c4e7c-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"9acb97f6-82ac-4891-96dc-43a85f9c4e7c\") " pod="openstack/rabbitmq-cell1-server-0" Jan 26 17:12:38 crc kubenswrapper[4754]: I0126 17:12:38.710244 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/9acb97f6-82ac-4891-96dc-43a85f9c4e7c-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"9acb97f6-82ac-4891-96dc-43a85f9c4e7c\") " pod="openstack/rabbitmq-cell1-server-0" Jan 26 17:12:38 crc kubenswrapper[4754]: I0126 17:12:38.710323 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/9acb97f6-82ac-4891-96dc-43a85f9c4e7c-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"9acb97f6-82ac-4891-96dc-43a85f9c4e7c\") " pod="openstack/rabbitmq-cell1-server-0" Jan 26 17:12:38 crc kubenswrapper[4754]: I0126 17:12:38.710446 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/9acb97f6-82ac-4891-96dc-43a85f9c4e7c-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"9acb97f6-82ac-4891-96dc-43a85f9c4e7c\") " pod="openstack/rabbitmq-cell1-server-0" Jan 26 17:12:38 crc kubenswrapper[4754]: I0126 17:12:38.710533 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/9acb97f6-82ac-4891-96dc-43a85f9c4e7c-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"9acb97f6-82ac-4891-96dc-43a85f9c4e7c\") " pod="openstack/rabbitmq-cell1-server-0" Jan 26 17:12:38 crc kubenswrapper[4754]: I0126 17:12:38.710626 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"9acb97f6-82ac-4891-96dc-43a85f9c4e7c\") " pod="openstack/rabbitmq-cell1-server-0" Jan 26 17:12:38 crc kubenswrapper[4754]: I0126 17:12:38.710711 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/9acb97f6-82ac-4891-96dc-43a85f9c4e7c-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"9acb97f6-82ac-4891-96dc-43a85f9c4e7c\") " pod="openstack/rabbitmq-cell1-server-0" Jan 26 17:12:38 crc kubenswrapper[4754]: I0126 17:12:38.710788 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9acb97f6-82ac-4891-96dc-43a85f9c4e7c-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"9acb97f6-82ac-4891-96dc-43a85f9c4e7c\") " pod="openstack/rabbitmq-cell1-server-0" Jan 26 17:12:38 crc kubenswrapper[4754]: I0126 17:12:38.812958 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"9acb97f6-82ac-4891-96dc-43a85f9c4e7c\") " pod="openstack/rabbitmq-cell1-server-0" Jan 26 17:12:38 crc kubenswrapper[4754]: I0126 17:12:38.813301 4754 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"9acb97f6-82ac-4891-96dc-43a85f9c4e7c\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/rabbitmq-cell1-server-0" Jan 26 17:12:38 crc kubenswrapper[4754]: I0126 17:12:38.813327 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/9acb97f6-82ac-4891-96dc-43a85f9c4e7c-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"9acb97f6-82ac-4891-96dc-43a85f9c4e7c\") " pod="openstack/rabbitmq-cell1-server-0" Jan 26 17:12:38 crc kubenswrapper[4754]: I0126 17:12:38.813375 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9acb97f6-82ac-4891-96dc-43a85f9c4e7c-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"9acb97f6-82ac-4891-96dc-43a85f9c4e7c\") " pod="openstack/rabbitmq-cell1-server-0" Jan 26 17:12:38 crc kubenswrapper[4754]: I0126 17:12:38.813404 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/9acb97f6-82ac-4891-96dc-43a85f9c4e7c-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"9acb97f6-82ac-4891-96dc-43a85f9c4e7c\") " pod="openstack/rabbitmq-cell1-server-0" Jan 26 17:12:38 crc kubenswrapper[4754]: I0126 17:12:38.813449 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9rv69\" (UniqueName: \"kubernetes.io/projected/9acb97f6-82ac-4891-96dc-43a85f9c4e7c-kube-api-access-9rv69\") pod \"rabbitmq-cell1-server-0\" (UID: \"9acb97f6-82ac-4891-96dc-43a85f9c4e7c\") " pod="openstack/rabbitmq-cell1-server-0" Jan 26 17:12:38 crc kubenswrapper[4754]: I0126 17:12:38.813488 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/9acb97f6-82ac-4891-96dc-43a85f9c4e7c-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"9acb97f6-82ac-4891-96dc-43a85f9c4e7c\") " pod="openstack/rabbitmq-cell1-server-0" Jan 26 17:12:38 crc kubenswrapper[4754]: I0126 17:12:38.813547 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/9acb97f6-82ac-4891-96dc-43a85f9c4e7c-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"9acb97f6-82ac-4891-96dc-43a85f9c4e7c\") " pod="openstack/rabbitmq-cell1-server-0" Jan 26 17:12:38 crc kubenswrapper[4754]: I0126 17:12:38.813582 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/9acb97f6-82ac-4891-96dc-43a85f9c4e7c-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"9acb97f6-82ac-4891-96dc-43a85f9c4e7c\") " pod="openstack/rabbitmq-cell1-server-0" Jan 26 17:12:38 crc kubenswrapper[4754]: I0126 17:12:38.813627 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/9acb97f6-82ac-4891-96dc-43a85f9c4e7c-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"9acb97f6-82ac-4891-96dc-43a85f9c4e7c\") " pod="openstack/rabbitmq-cell1-server-0" Jan 26 17:12:38 crc kubenswrapper[4754]: I0126 17:12:38.813834 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/9acb97f6-82ac-4891-96dc-43a85f9c4e7c-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"9acb97f6-82ac-4891-96dc-43a85f9c4e7c\") " pod="openstack/rabbitmq-cell1-server-0" Jan 26 17:12:38 crc kubenswrapper[4754]: I0126 17:12:38.813988 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/9acb97f6-82ac-4891-96dc-43a85f9c4e7c-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"9acb97f6-82ac-4891-96dc-43a85f9c4e7c\") " pod="openstack/rabbitmq-cell1-server-0" Jan 26 17:12:38 crc kubenswrapper[4754]: I0126 17:12:38.814460 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9acb97f6-82ac-4891-96dc-43a85f9c4e7c-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"9acb97f6-82ac-4891-96dc-43a85f9c4e7c\") " pod="openstack/rabbitmq-cell1-server-0" Jan 26 17:12:38 crc kubenswrapper[4754]: I0126 17:12:38.814520 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/9acb97f6-82ac-4891-96dc-43a85f9c4e7c-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"9acb97f6-82ac-4891-96dc-43a85f9c4e7c\") " pod="openstack/rabbitmq-cell1-server-0" Jan 26 17:12:38 crc kubenswrapper[4754]: I0126 17:12:38.814608 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/9acb97f6-82ac-4891-96dc-43a85f9c4e7c-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"9acb97f6-82ac-4891-96dc-43a85f9c4e7c\") " pod="openstack/rabbitmq-cell1-server-0" Jan 26 17:12:38 crc kubenswrapper[4754]: I0126 17:12:38.815298 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/9acb97f6-82ac-4891-96dc-43a85f9c4e7c-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"9acb97f6-82ac-4891-96dc-43a85f9c4e7c\") " pod="openstack/rabbitmq-cell1-server-0" Jan 26 17:12:38 crc kubenswrapper[4754]: I0126 17:12:38.815418 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/9acb97f6-82ac-4891-96dc-43a85f9c4e7c-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"9acb97f6-82ac-4891-96dc-43a85f9c4e7c\") " pod="openstack/rabbitmq-cell1-server-0" Jan 26 17:12:38 crc kubenswrapper[4754]: I0126 17:12:38.818545 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/9acb97f6-82ac-4891-96dc-43a85f9c4e7c-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"9acb97f6-82ac-4891-96dc-43a85f9c4e7c\") " pod="openstack/rabbitmq-cell1-server-0" Jan 26 17:12:38 crc kubenswrapper[4754]: I0126 17:12:38.819457 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/9acb97f6-82ac-4891-96dc-43a85f9c4e7c-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"9acb97f6-82ac-4891-96dc-43a85f9c4e7c\") " pod="openstack/rabbitmq-cell1-server-0" Jan 26 17:12:38 crc kubenswrapper[4754]: I0126 17:12:38.819568 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/9acb97f6-82ac-4891-96dc-43a85f9c4e7c-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"9acb97f6-82ac-4891-96dc-43a85f9c4e7c\") " pod="openstack/rabbitmq-cell1-server-0" Jan 26 17:12:38 crc kubenswrapper[4754]: I0126 17:12:38.822524 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/9acb97f6-82ac-4891-96dc-43a85f9c4e7c-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"9acb97f6-82ac-4891-96dc-43a85f9c4e7c\") " pod="openstack/rabbitmq-cell1-server-0" Jan 26 17:12:38 crc kubenswrapper[4754]: I0126 17:12:38.831213 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9rv69\" (UniqueName: \"kubernetes.io/projected/9acb97f6-82ac-4891-96dc-43a85f9c4e7c-kube-api-access-9rv69\") pod \"rabbitmq-cell1-server-0\" (UID: \"9acb97f6-82ac-4891-96dc-43a85f9c4e7c\") " pod="openstack/rabbitmq-cell1-server-0" Jan 26 17:12:38 crc kubenswrapper[4754]: I0126 17:12:38.845053 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"9acb97f6-82ac-4891-96dc-43a85f9c4e7c\") " pod="openstack/rabbitmq-cell1-server-0" Jan 26 17:12:38 crc kubenswrapper[4754]: I0126 17:12:38.894730 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Jan 26 17:12:39 crc kubenswrapper[4754]: I0126 17:12:39.017573 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-79bd4cc8c9-h52kf"] Jan 26 17:12:39 crc kubenswrapper[4754]: I0126 17:12:39.019868 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-79bd4cc8c9-h52kf" Jan 26 17:12:39 crc kubenswrapper[4754]: I0126 17:12:39.022525 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-edpm-ipam" Jan 26 17:12:39 crc kubenswrapper[4754]: I0126 17:12:39.049079 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-79bd4cc8c9-h52kf"] Jan 26 17:12:39 crc kubenswrapper[4754]: I0126 17:12:39.120548 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9cf31d29-b33a-4a3d-ab69-98094ed6364b-config\") pod \"dnsmasq-dns-79bd4cc8c9-h52kf\" (UID: \"9cf31d29-b33a-4a3d-ab69-98094ed6364b\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-h52kf" Jan 26 17:12:39 crc kubenswrapper[4754]: I0126 17:12:39.120608 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9cf31d29-b33a-4a3d-ab69-98094ed6364b-ovsdbserver-nb\") pod \"dnsmasq-dns-79bd4cc8c9-h52kf\" (UID: \"9cf31d29-b33a-4a3d-ab69-98094ed6364b\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-h52kf" Jan 26 17:12:39 crc kubenswrapper[4754]: I0126 17:12:39.120658 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/9cf31d29-b33a-4a3d-ab69-98094ed6364b-openstack-edpm-ipam\") pod \"dnsmasq-dns-79bd4cc8c9-h52kf\" (UID: \"9cf31d29-b33a-4a3d-ab69-98094ed6364b\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-h52kf" Jan 26 17:12:39 crc kubenswrapper[4754]: I0126 17:12:39.120714 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9cf31d29-b33a-4a3d-ab69-98094ed6364b-ovsdbserver-sb\") pod \"dnsmasq-dns-79bd4cc8c9-h52kf\" (UID: \"9cf31d29-b33a-4a3d-ab69-98094ed6364b\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-h52kf" Jan 26 17:12:39 crc kubenswrapper[4754]: I0126 17:12:39.120739 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9cf31d29-b33a-4a3d-ab69-98094ed6364b-dns-svc\") pod \"dnsmasq-dns-79bd4cc8c9-h52kf\" (UID: \"9cf31d29-b33a-4a3d-ab69-98094ed6364b\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-h52kf" Jan 26 17:12:39 crc kubenswrapper[4754]: I0126 17:12:39.120799 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9cf31d29-b33a-4a3d-ab69-98094ed6364b-dns-swift-storage-0\") pod \"dnsmasq-dns-79bd4cc8c9-h52kf\" (UID: \"9cf31d29-b33a-4a3d-ab69-98094ed6364b\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-h52kf" Jan 26 17:12:39 crc kubenswrapper[4754]: I0126 17:12:39.120835 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rqlc7\" (UniqueName: \"kubernetes.io/projected/9cf31d29-b33a-4a3d-ab69-98094ed6364b-kube-api-access-rqlc7\") pod \"dnsmasq-dns-79bd4cc8c9-h52kf\" (UID: \"9cf31d29-b33a-4a3d-ab69-98094ed6364b\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-h52kf" Jan 26 17:12:39 crc kubenswrapper[4754]: I0126 17:12:39.225525 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9cf31d29-b33a-4a3d-ab69-98094ed6364b-dns-swift-storage-0\") pod \"dnsmasq-dns-79bd4cc8c9-h52kf\" (UID: \"9cf31d29-b33a-4a3d-ab69-98094ed6364b\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-h52kf" Jan 26 17:12:39 crc kubenswrapper[4754]: I0126 17:12:39.225590 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rqlc7\" (UniqueName: \"kubernetes.io/projected/9cf31d29-b33a-4a3d-ab69-98094ed6364b-kube-api-access-rqlc7\") pod \"dnsmasq-dns-79bd4cc8c9-h52kf\" (UID: \"9cf31d29-b33a-4a3d-ab69-98094ed6364b\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-h52kf" Jan 26 17:12:39 crc kubenswrapper[4754]: I0126 17:12:39.226399 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9cf31d29-b33a-4a3d-ab69-98094ed6364b-config\") pod \"dnsmasq-dns-79bd4cc8c9-h52kf\" (UID: \"9cf31d29-b33a-4a3d-ab69-98094ed6364b\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-h52kf" Jan 26 17:12:39 crc kubenswrapper[4754]: I0126 17:12:39.226430 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9cf31d29-b33a-4a3d-ab69-98094ed6364b-ovsdbserver-nb\") pod \"dnsmasq-dns-79bd4cc8c9-h52kf\" (UID: \"9cf31d29-b33a-4a3d-ab69-98094ed6364b\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-h52kf" Jan 26 17:12:39 crc kubenswrapper[4754]: I0126 17:12:39.226513 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/9cf31d29-b33a-4a3d-ab69-98094ed6364b-openstack-edpm-ipam\") pod \"dnsmasq-dns-79bd4cc8c9-h52kf\" (UID: \"9cf31d29-b33a-4a3d-ab69-98094ed6364b\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-h52kf" Jan 26 17:12:39 crc kubenswrapper[4754]: I0126 17:12:39.226535 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9cf31d29-b33a-4a3d-ab69-98094ed6364b-ovsdbserver-sb\") pod \"dnsmasq-dns-79bd4cc8c9-h52kf\" (UID: \"9cf31d29-b33a-4a3d-ab69-98094ed6364b\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-h52kf" Jan 26 17:12:39 crc kubenswrapper[4754]: I0126 17:12:39.226587 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9cf31d29-b33a-4a3d-ab69-98094ed6364b-dns-svc\") pod \"dnsmasq-dns-79bd4cc8c9-h52kf\" (UID: \"9cf31d29-b33a-4a3d-ab69-98094ed6364b\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-h52kf" Jan 26 17:12:39 crc kubenswrapper[4754]: I0126 17:12:39.226614 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9cf31d29-b33a-4a3d-ab69-98094ed6364b-dns-swift-storage-0\") pod \"dnsmasq-dns-79bd4cc8c9-h52kf\" (UID: \"9cf31d29-b33a-4a3d-ab69-98094ed6364b\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-h52kf" Jan 26 17:12:39 crc kubenswrapper[4754]: I0126 17:12:39.227155 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9cf31d29-b33a-4a3d-ab69-98094ed6364b-ovsdbserver-nb\") pod \"dnsmasq-dns-79bd4cc8c9-h52kf\" (UID: \"9cf31d29-b33a-4a3d-ab69-98094ed6364b\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-h52kf" Jan 26 17:12:39 crc kubenswrapper[4754]: I0126 17:12:39.227382 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9cf31d29-b33a-4a3d-ab69-98094ed6364b-ovsdbserver-sb\") pod \"dnsmasq-dns-79bd4cc8c9-h52kf\" (UID: \"9cf31d29-b33a-4a3d-ab69-98094ed6364b\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-h52kf" Jan 26 17:12:39 crc kubenswrapper[4754]: I0126 17:12:39.227791 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9cf31d29-b33a-4a3d-ab69-98094ed6364b-dns-svc\") pod \"dnsmasq-dns-79bd4cc8c9-h52kf\" (UID: \"9cf31d29-b33a-4a3d-ab69-98094ed6364b\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-h52kf" Jan 26 17:12:39 crc kubenswrapper[4754]: I0126 17:12:39.229225 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9cf31d29-b33a-4a3d-ab69-98094ed6364b-config\") pod \"dnsmasq-dns-79bd4cc8c9-h52kf\" (UID: \"9cf31d29-b33a-4a3d-ab69-98094ed6364b\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-h52kf" Jan 26 17:12:39 crc kubenswrapper[4754]: I0126 17:12:39.230592 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/9cf31d29-b33a-4a3d-ab69-98094ed6364b-openstack-edpm-ipam\") pod \"dnsmasq-dns-79bd4cc8c9-h52kf\" (UID: \"9cf31d29-b33a-4a3d-ab69-98094ed6364b\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-h52kf" Jan 26 17:12:39 crc kubenswrapper[4754]: I0126 17:12:39.246898 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"6e2da069-1c54-4801-a91c-241b80e8d17b","Type":"ContainerStarted","Data":"667b4c065827e6c2ad1e3c6a95979170a8ced1f3ee0ba5771ec1ce4235e28faa"} Jan 26 17:12:39 crc kubenswrapper[4754]: I0126 17:12:39.267756 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rqlc7\" (UniqueName: \"kubernetes.io/projected/9cf31d29-b33a-4a3d-ab69-98094ed6364b-kube-api-access-rqlc7\") pod \"dnsmasq-dns-79bd4cc8c9-h52kf\" (UID: \"9cf31d29-b33a-4a3d-ab69-98094ed6364b\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-h52kf" Jan 26 17:12:39 crc kubenswrapper[4754]: I0126 17:12:39.323832 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Jan 26 17:12:39 crc kubenswrapper[4754]: W0126 17:12:39.347220 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9acb97f6_82ac_4891_96dc_43a85f9c4e7c.slice/crio-b69eaacd954ae0fc8451b224143d6a870529bfd0010d932e1116f4ada5fa1453 WatchSource:0}: Error finding container b69eaacd954ae0fc8451b224143d6a870529bfd0010d932e1116f4ada5fa1453: Status 404 returned error can't find the container with id b69eaacd954ae0fc8451b224143d6a870529bfd0010d932e1116f4ada5fa1453 Jan 26 17:12:39 crc kubenswrapper[4754]: I0126 17:12:39.351699 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-79bd4cc8c9-h52kf" Jan 26 17:12:39 crc kubenswrapper[4754]: I0126 17:12:39.780322 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="48810c5e-3c40-4cdc-8bab-47efa97e76fa" path="/var/lib/kubelet/pods/48810c5e-3c40-4cdc-8bab-47efa97e76fa/volumes" Jan 26 17:12:39 crc kubenswrapper[4754]: I0126 17:12:39.843710 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-79bd4cc8c9-h52kf"] Jan 26 17:12:39 crc kubenswrapper[4754]: W0126 17:12:39.846832 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9cf31d29_b33a_4a3d_ab69_98094ed6364b.slice/crio-792d3c94e10ebd7364161c71adc541a04bd5e23b0ded389e30409fe05ceaedcd WatchSource:0}: Error finding container 792d3c94e10ebd7364161c71adc541a04bd5e23b0ded389e30409fe05ceaedcd: Status 404 returned error can't find the container with id 792d3c94e10ebd7364161c71adc541a04bd5e23b0ded389e30409fe05ceaedcd Jan 26 17:12:40 crc kubenswrapper[4754]: I0126 17:12:40.316195 4754 generic.go:334] "Generic (PLEG): container finished" podID="9cf31d29-b33a-4a3d-ab69-98094ed6364b" containerID="eecbbaf507684103d27ed11791648478ec4f0b4a963953c60909a7ac24b4e6db" exitCode=0 Jan 26 17:12:40 crc kubenswrapper[4754]: I0126 17:12:40.316592 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79bd4cc8c9-h52kf" event={"ID":"9cf31d29-b33a-4a3d-ab69-98094ed6364b","Type":"ContainerDied","Data":"eecbbaf507684103d27ed11791648478ec4f0b4a963953c60909a7ac24b4e6db"} Jan 26 17:12:40 crc kubenswrapper[4754]: I0126 17:12:40.316623 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79bd4cc8c9-h52kf" event={"ID":"9cf31d29-b33a-4a3d-ab69-98094ed6364b","Type":"ContainerStarted","Data":"792d3c94e10ebd7364161c71adc541a04bd5e23b0ded389e30409fe05ceaedcd"} Jan 26 17:12:40 crc kubenswrapper[4754]: I0126 17:12:40.320310 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"9acb97f6-82ac-4891-96dc-43a85f9c4e7c","Type":"ContainerStarted","Data":"b69eaacd954ae0fc8451b224143d6a870529bfd0010d932e1116f4ada5fa1453"} Jan 26 17:12:41 crc kubenswrapper[4754]: I0126 17:12:41.329712 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79bd4cc8c9-h52kf" event={"ID":"9cf31d29-b33a-4a3d-ab69-98094ed6364b","Type":"ContainerStarted","Data":"867f10be7d6d79bb9f1fbc8fcf04eedd21efb11b71b3248822ebec5b8840613a"} Jan 26 17:12:41 crc kubenswrapper[4754]: I0126 17:12:41.330408 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-79bd4cc8c9-h52kf" Jan 26 17:12:41 crc kubenswrapper[4754]: I0126 17:12:41.331478 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"9acb97f6-82ac-4891-96dc-43a85f9c4e7c","Type":"ContainerStarted","Data":"e3dc5e85d5a3c46330918b93a52510e835514f6976f66406c0fbc1b543517d78"} Jan 26 17:12:41 crc kubenswrapper[4754]: I0126 17:12:41.355376 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-79bd4cc8c9-h52kf" podStartSLOduration=3.355350424 podStartE2EDuration="3.355350424s" podCreationTimestamp="2026-01-26 17:12:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 17:12:41.345174464 +0000 UTC m=+1527.869354898" watchObservedRunningTime="2026-01-26 17:12:41.355350424 +0000 UTC m=+1527.879530868" Jan 26 17:12:49 crc kubenswrapper[4754]: I0126 17:12:49.353644 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-79bd4cc8c9-h52kf" Jan 26 17:12:49 crc kubenswrapper[4754]: I0126 17:12:49.440201 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-89c5cd4d5-8wrps"] Jan 26 17:12:49 crc kubenswrapper[4754]: I0126 17:12:49.440900 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-89c5cd4d5-8wrps" podUID="443a6550-c50d-453e-9f10-3875fcb41c18" containerName="dnsmasq-dns" containerID="cri-o://da940fcb0040028dfd4aebd2f8c5ab637c5ac7300ce1a8c53fa14feb47b58d3b" gracePeriod=10 Jan 26 17:12:49 crc kubenswrapper[4754]: I0126 17:12:49.601329 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-55478c4467-c65ql"] Jan 26 17:12:49 crc kubenswrapper[4754]: I0126 17:12:49.603031 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55478c4467-c65ql" Jan 26 17:12:49 crc kubenswrapper[4754]: I0126 17:12:49.616862 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-55478c4467-c65ql"] Jan 26 17:12:49 crc kubenswrapper[4754]: I0126 17:12:49.732680 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/48e65463-af70-4cca-b90b-a3b4ac9a1619-dns-svc\") pod \"dnsmasq-dns-55478c4467-c65ql\" (UID: \"48e65463-af70-4cca-b90b-a3b4ac9a1619\") " pod="openstack/dnsmasq-dns-55478c4467-c65ql" Jan 26 17:12:49 crc kubenswrapper[4754]: I0126 17:12:49.732741 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/48e65463-af70-4cca-b90b-a3b4ac9a1619-ovsdbserver-nb\") pod \"dnsmasq-dns-55478c4467-c65ql\" (UID: \"48e65463-af70-4cca-b90b-a3b4ac9a1619\") " pod="openstack/dnsmasq-dns-55478c4467-c65ql" Jan 26 17:12:49 crc kubenswrapper[4754]: I0126 17:12:49.732790 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/48e65463-af70-4cca-b90b-a3b4ac9a1619-openstack-edpm-ipam\") pod \"dnsmasq-dns-55478c4467-c65ql\" (UID: \"48e65463-af70-4cca-b90b-a3b4ac9a1619\") " pod="openstack/dnsmasq-dns-55478c4467-c65ql" Jan 26 17:12:49 crc kubenswrapper[4754]: I0126 17:12:49.732813 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/48e65463-af70-4cca-b90b-a3b4ac9a1619-config\") pod \"dnsmasq-dns-55478c4467-c65ql\" (UID: \"48e65463-af70-4cca-b90b-a3b4ac9a1619\") " pod="openstack/dnsmasq-dns-55478c4467-c65ql" Jan 26 17:12:49 crc kubenswrapper[4754]: I0126 17:12:49.732857 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/48e65463-af70-4cca-b90b-a3b4ac9a1619-ovsdbserver-sb\") pod \"dnsmasq-dns-55478c4467-c65ql\" (UID: \"48e65463-af70-4cca-b90b-a3b4ac9a1619\") " pod="openstack/dnsmasq-dns-55478c4467-c65ql" Jan 26 17:12:49 crc kubenswrapper[4754]: I0126 17:12:49.732881 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/48e65463-af70-4cca-b90b-a3b4ac9a1619-dns-swift-storage-0\") pod \"dnsmasq-dns-55478c4467-c65ql\" (UID: \"48e65463-af70-4cca-b90b-a3b4ac9a1619\") " pod="openstack/dnsmasq-dns-55478c4467-c65ql" Jan 26 17:12:49 crc kubenswrapper[4754]: I0126 17:12:49.732910 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rjgl6\" (UniqueName: \"kubernetes.io/projected/48e65463-af70-4cca-b90b-a3b4ac9a1619-kube-api-access-rjgl6\") pod \"dnsmasq-dns-55478c4467-c65ql\" (UID: \"48e65463-af70-4cca-b90b-a3b4ac9a1619\") " pod="openstack/dnsmasq-dns-55478c4467-c65ql" Jan 26 17:12:49 crc kubenswrapper[4754]: I0126 17:12:49.833978 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/48e65463-af70-4cca-b90b-a3b4ac9a1619-openstack-edpm-ipam\") pod \"dnsmasq-dns-55478c4467-c65ql\" (UID: \"48e65463-af70-4cca-b90b-a3b4ac9a1619\") " pod="openstack/dnsmasq-dns-55478c4467-c65ql" Jan 26 17:12:49 crc kubenswrapper[4754]: I0126 17:12:49.834285 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/48e65463-af70-4cca-b90b-a3b4ac9a1619-config\") pod \"dnsmasq-dns-55478c4467-c65ql\" (UID: \"48e65463-af70-4cca-b90b-a3b4ac9a1619\") " pod="openstack/dnsmasq-dns-55478c4467-c65ql" Jan 26 17:12:49 crc kubenswrapper[4754]: I0126 17:12:49.834341 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/48e65463-af70-4cca-b90b-a3b4ac9a1619-ovsdbserver-sb\") pod \"dnsmasq-dns-55478c4467-c65ql\" (UID: \"48e65463-af70-4cca-b90b-a3b4ac9a1619\") " pod="openstack/dnsmasq-dns-55478c4467-c65ql" Jan 26 17:12:49 crc kubenswrapper[4754]: I0126 17:12:49.834364 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/48e65463-af70-4cca-b90b-a3b4ac9a1619-dns-swift-storage-0\") pod \"dnsmasq-dns-55478c4467-c65ql\" (UID: \"48e65463-af70-4cca-b90b-a3b4ac9a1619\") " pod="openstack/dnsmasq-dns-55478c4467-c65ql" Jan 26 17:12:49 crc kubenswrapper[4754]: I0126 17:12:49.834396 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rjgl6\" (UniqueName: \"kubernetes.io/projected/48e65463-af70-4cca-b90b-a3b4ac9a1619-kube-api-access-rjgl6\") pod \"dnsmasq-dns-55478c4467-c65ql\" (UID: \"48e65463-af70-4cca-b90b-a3b4ac9a1619\") " pod="openstack/dnsmasq-dns-55478c4467-c65ql" Jan 26 17:12:49 crc kubenswrapper[4754]: I0126 17:12:49.834472 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/48e65463-af70-4cca-b90b-a3b4ac9a1619-dns-svc\") pod \"dnsmasq-dns-55478c4467-c65ql\" (UID: \"48e65463-af70-4cca-b90b-a3b4ac9a1619\") " pod="openstack/dnsmasq-dns-55478c4467-c65ql" Jan 26 17:12:49 crc kubenswrapper[4754]: I0126 17:12:49.834497 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/48e65463-af70-4cca-b90b-a3b4ac9a1619-ovsdbserver-nb\") pod \"dnsmasq-dns-55478c4467-c65ql\" (UID: \"48e65463-af70-4cca-b90b-a3b4ac9a1619\") " pod="openstack/dnsmasq-dns-55478c4467-c65ql" Jan 26 17:12:49 crc kubenswrapper[4754]: I0126 17:12:49.835429 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/48e65463-af70-4cca-b90b-a3b4ac9a1619-ovsdbserver-nb\") pod \"dnsmasq-dns-55478c4467-c65ql\" (UID: \"48e65463-af70-4cca-b90b-a3b4ac9a1619\") " pod="openstack/dnsmasq-dns-55478c4467-c65ql" Jan 26 17:12:49 crc kubenswrapper[4754]: I0126 17:12:49.839083 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/48e65463-af70-4cca-b90b-a3b4ac9a1619-dns-swift-storage-0\") pod \"dnsmasq-dns-55478c4467-c65ql\" (UID: \"48e65463-af70-4cca-b90b-a3b4ac9a1619\") " pod="openstack/dnsmasq-dns-55478c4467-c65ql" Jan 26 17:12:49 crc kubenswrapper[4754]: I0126 17:12:49.839703 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/48e65463-af70-4cca-b90b-a3b4ac9a1619-openstack-edpm-ipam\") pod \"dnsmasq-dns-55478c4467-c65ql\" (UID: \"48e65463-af70-4cca-b90b-a3b4ac9a1619\") " pod="openstack/dnsmasq-dns-55478c4467-c65ql" Jan 26 17:12:49 crc kubenswrapper[4754]: I0126 17:12:49.839931 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/48e65463-af70-4cca-b90b-a3b4ac9a1619-ovsdbserver-sb\") pod \"dnsmasq-dns-55478c4467-c65ql\" (UID: \"48e65463-af70-4cca-b90b-a3b4ac9a1619\") " pod="openstack/dnsmasq-dns-55478c4467-c65ql" Jan 26 17:12:49 crc kubenswrapper[4754]: I0126 17:12:49.840585 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/48e65463-af70-4cca-b90b-a3b4ac9a1619-dns-svc\") pod \"dnsmasq-dns-55478c4467-c65ql\" (UID: \"48e65463-af70-4cca-b90b-a3b4ac9a1619\") " pod="openstack/dnsmasq-dns-55478c4467-c65ql" Jan 26 17:12:49 crc kubenswrapper[4754]: I0126 17:12:49.841869 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/48e65463-af70-4cca-b90b-a3b4ac9a1619-config\") pod \"dnsmasq-dns-55478c4467-c65ql\" (UID: \"48e65463-af70-4cca-b90b-a3b4ac9a1619\") " pod="openstack/dnsmasq-dns-55478c4467-c65ql" Jan 26 17:12:49 crc kubenswrapper[4754]: I0126 17:12:49.873952 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rjgl6\" (UniqueName: \"kubernetes.io/projected/48e65463-af70-4cca-b90b-a3b4ac9a1619-kube-api-access-rjgl6\") pod \"dnsmasq-dns-55478c4467-c65ql\" (UID: \"48e65463-af70-4cca-b90b-a3b4ac9a1619\") " pod="openstack/dnsmasq-dns-55478c4467-c65ql" Jan 26 17:12:49 crc kubenswrapper[4754]: I0126 17:12:49.960412 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55478c4467-c65ql" Jan 26 17:12:50 crc kubenswrapper[4754]: I0126 17:12:50.074247 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-89c5cd4d5-8wrps" Jan 26 17:12:50 crc kubenswrapper[4754]: I0126 17:12:50.241058 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/443a6550-c50d-453e-9f10-3875fcb41c18-ovsdbserver-nb\") pod \"443a6550-c50d-453e-9f10-3875fcb41c18\" (UID: \"443a6550-c50d-453e-9f10-3875fcb41c18\") " Jan 26 17:12:50 crc kubenswrapper[4754]: I0126 17:12:50.241551 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/443a6550-c50d-453e-9f10-3875fcb41c18-config\") pod \"443a6550-c50d-453e-9f10-3875fcb41c18\" (UID: \"443a6550-c50d-453e-9f10-3875fcb41c18\") " Jan 26 17:12:50 crc kubenswrapper[4754]: I0126 17:12:50.241625 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/443a6550-c50d-453e-9f10-3875fcb41c18-dns-svc\") pod \"443a6550-c50d-453e-9f10-3875fcb41c18\" (UID: \"443a6550-c50d-453e-9f10-3875fcb41c18\") " Jan 26 17:12:50 crc kubenswrapper[4754]: I0126 17:12:50.241644 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/443a6550-c50d-453e-9f10-3875fcb41c18-dns-swift-storage-0\") pod \"443a6550-c50d-453e-9f10-3875fcb41c18\" (UID: \"443a6550-c50d-453e-9f10-3875fcb41c18\") " Jan 26 17:12:50 crc kubenswrapper[4754]: I0126 17:12:50.241750 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/443a6550-c50d-453e-9f10-3875fcb41c18-ovsdbserver-sb\") pod \"443a6550-c50d-453e-9f10-3875fcb41c18\" (UID: \"443a6550-c50d-453e-9f10-3875fcb41c18\") " Jan 26 17:12:50 crc kubenswrapper[4754]: I0126 17:12:50.241833 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bstvk\" (UniqueName: \"kubernetes.io/projected/443a6550-c50d-453e-9f10-3875fcb41c18-kube-api-access-bstvk\") pod \"443a6550-c50d-453e-9f10-3875fcb41c18\" (UID: \"443a6550-c50d-453e-9f10-3875fcb41c18\") " Jan 26 17:12:50 crc kubenswrapper[4754]: I0126 17:12:50.246456 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/443a6550-c50d-453e-9f10-3875fcb41c18-kube-api-access-bstvk" (OuterVolumeSpecName: "kube-api-access-bstvk") pod "443a6550-c50d-453e-9f10-3875fcb41c18" (UID: "443a6550-c50d-453e-9f10-3875fcb41c18"). InnerVolumeSpecName "kube-api-access-bstvk". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:12:50 crc kubenswrapper[4754]: I0126 17:12:50.297501 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/443a6550-c50d-453e-9f10-3875fcb41c18-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "443a6550-c50d-453e-9f10-3875fcb41c18" (UID: "443a6550-c50d-453e-9f10-3875fcb41c18"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:12:50 crc kubenswrapper[4754]: I0126 17:12:50.303365 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/443a6550-c50d-453e-9f10-3875fcb41c18-config" (OuterVolumeSpecName: "config") pod "443a6550-c50d-453e-9f10-3875fcb41c18" (UID: "443a6550-c50d-453e-9f10-3875fcb41c18"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:12:50 crc kubenswrapper[4754]: I0126 17:12:50.304390 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/443a6550-c50d-453e-9f10-3875fcb41c18-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "443a6550-c50d-453e-9f10-3875fcb41c18" (UID: "443a6550-c50d-453e-9f10-3875fcb41c18"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:12:50 crc kubenswrapper[4754]: I0126 17:12:50.317852 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/443a6550-c50d-453e-9f10-3875fcb41c18-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "443a6550-c50d-453e-9f10-3875fcb41c18" (UID: "443a6550-c50d-453e-9f10-3875fcb41c18"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:12:50 crc kubenswrapper[4754]: I0126 17:12:50.318195 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/443a6550-c50d-453e-9f10-3875fcb41c18-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "443a6550-c50d-453e-9f10-3875fcb41c18" (UID: "443a6550-c50d-453e-9f10-3875fcb41c18"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:12:50 crc kubenswrapper[4754]: I0126 17:12:50.353082 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bstvk\" (UniqueName: \"kubernetes.io/projected/443a6550-c50d-453e-9f10-3875fcb41c18-kube-api-access-bstvk\") on node \"crc\" DevicePath \"\"" Jan 26 17:12:50 crc kubenswrapper[4754]: I0126 17:12:50.353207 4754 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/443a6550-c50d-453e-9f10-3875fcb41c18-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Jan 26 17:12:50 crc kubenswrapper[4754]: I0126 17:12:50.353259 4754 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/443a6550-c50d-453e-9f10-3875fcb41c18-config\") on node \"crc\" DevicePath \"\"" Jan 26 17:12:50 crc kubenswrapper[4754]: I0126 17:12:50.353306 4754 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/443a6550-c50d-453e-9f10-3875fcb41c18-dns-svc\") on node \"crc\" DevicePath \"\"" Jan 26 17:12:50 crc kubenswrapper[4754]: I0126 17:12:50.353370 4754 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/443a6550-c50d-453e-9f10-3875fcb41c18-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Jan 26 17:12:50 crc kubenswrapper[4754]: I0126 17:12:50.353431 4754 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/443a6550-c50d-453e-9f10-3875fcb41c18-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Jan 26 17:12:50 crc kubenswrapper[4754]: I0126 17:12:50.418962 4754 generic.go:334] "Generic (PLEG): container finished" podID="443a6550-c50d-453e-9f10-3875fcb41c18" containerID="da940fcb0040028dfd4aebd2f8c5ab637c5ac7300ce1a8c53fa14feb47b58d3b" exitCode=0 Jan 26 17:12:50 crc kubenswrapper[4754]: I0126 17:12:50.419015 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-89c5cd4d5-8wrps" event={"ID":"443a6550-c50d-453e-9f10-3875fcb41c18","Type":"ContainerDied","Data":"da940fcb0040028dfd4aebd2f8c5ab637c5ac7300ce1a8c53fa14feb47b58d3b"} Jan 26 17:12:50 crc kubenswrapper[4754]: I0126 17:12:50.419773 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-89c5cd4d5-8wrps" event={"ID":"443a6550-c50d-453e-9f10-3875fcb41c18","Type":"ContainerDied","Data":"d4197cb2cfe9741f52a6fe12eb036913fc38177359ef56817c5baf702af7553d"} Jan 26 17:12:50 crc kubenswrapper[4754]: I0126 17:12:50.419807 4754 scope.go:117] "RemoveContainer" containerID="da940fcb0040028dfd4aebd2f8c5ab637c5ac7300ce1a8c53fa14feb47b58d3b" Jan 26 17:12:50 crc kubenswrapper[4754]: I0126 17:12:50.419047 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-89c5cd4d5-8wrps" Jan 26 17:12:50 crc kubenswrapper[4754]: I0126 17:12:50.442617 4754 scope.go:117] "RemoveContainer" containerID="121ef5f4c966fb80a1b1b516ebd2e15d37b4b339bcf34b10842645bd2d6f26a4" Jan 26 17:12:50 crc kubenswrapper[4754]: I0126 17:12:50.452885 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-89c5cd4d5-8wrps"] Jan 26 17:12:50 crc kubenswrapper[4754]: I0126 17:12:50.464203 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-89c5cd4d5-8wrps"] Jan 26 17:12:50 crc kubenswrapper[4754]: I0126 17:12:50.480313 4754 scope.go:117] "RemoveContainer" containerID="da940fcb0040028dfd4aebd2f8c5ab637c5ac7300ce1a8c53fa14feb47b58d3b" Jan 26 17:12:50 crc kubenswrapper[4754]: E0126 17:12:50.480797 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"da940fcb0040028dfd4aebd2f8c5ab637c5ac7300ce1a8c53fa14feb47b58d3b\": container with ID starting with da940fcb0040028dfd4aebd2f8c5ab637c5ac7300ce1a8c53fa14feb47b58d3b not found: ID does not exist" containerID="da940fcb0040028dfd4aebd2f8c5ab637c5ac7300ce1a8c53fa14feb47b58d3b" Jan 26 17:12:50 crc kubenswrapper[4754]: I0126 17:12:50.480838 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"da940fcb0040028dfd4aebd2f8c5ab637c5ac7300ce1a8c53fa14feb47b58d3b"} err="failed to get container status \"da940fcb0040028dfd4aebd2f8c5ab637c5ac7300ce1a8c53fa14feb47b58d3b\": rpc error: code = NotFound desc = could not find container \"da940fcb0040028dfd4aebd2f8c5ab637c5ac7300ce1a8c53fa14feb47b58d3b\": container with ID starting with da940fcb0040028dfd4aebd2f8c5ab637c5ac7300ce1a8c53fa14feb47b58d3b not found: ID does not exist" Jan 26 17:12:50 crc kubenswrapper[4754]: I0126 17:12:50.480862 4754 scope.go:117] "RemoveContainer" containerID="121ef5f4c966fb80a1b1b516ebd2e15d37b4b339bcf34b10842645bd2d6f26a4" Jan 26 17:12:50 crc kubenswrapper[4754]: E0126 17:12:50.481185 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"121ef5f4c966fb80a1b1b516ebd2e15d37b4b339bcf34b10842645bd2d6f26a4\": container with ID starting with 121ef5f4c966fb80a1b1b516ebd2e15d37b4b339bcf34b10842645bd2d6f26a4 not found: ID does not exist" containerID="121ef5f4c966fb80a1b1b516ebd2e15d37b4b339bcf34b10842645bd2d6f26a4" Jan 26 17:12:50 crc kubenswrapper[4754]: I0126 17:12:50.481209 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"121ef5f4c966fb80a1b1b516ebd2e15d37b4b339bcf34b10842645bd2d6f26a4"} err="failed to get container status \"121ef5f4c966fb80a1b1b516ebd2e15d37b4b339bcf34b10842645bd2d6f26a4\": rpc error: code = NotFound desc = could not find container \"121ef5f4c966fb80a1b1b516ebd2e15d37b4b339bcf34b10842645bd2d6f26a4\": container with ID starting with 121ef5f4c966fb80a1b1b516ebd2e15d37b4b339bcf34b10842645bd2d6f26a4 not found: ID does not exist" Jan 26 17:12:50 crc kubenswrapper[4754]: I0126 17:12:50.563647 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-55478c4467-c65ql"] Jan 26 17:12:51 crc kubenswrapper[4754]: I0126 17:12:51.435559 4754 generic.go:334] "Generic (PLEG): container finished" podID="48e65463-af70-4cca-b90b-a3b4ac9a1619" containerID="8d37fca8df78e5793839bfbb81a94d4b100bbe35baac53619a52eb008ee2563a" exitCode=0 Jan 26 17:12:51 crc kubenswrapper[4754]: I0126 17:12:51.435618 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55478c4467-c65ql" event={"ID":"48e65463-af70-4cca-b90b-a3b4ac9a1619","Type":"ContainerDied","Data":"8d37fca8df78e5793839bfbb81a94d4b100bbe35baac53619a52eb008ee2563a"} Jan 26 17:12:51 crc kubenswrapper[4754]: I0126 17:12:51.437295 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55478c4467-c65ql" event={"ID":"48e65463-af70-4cca-b90b-a3b4ac9a1619","Type":"ContainerStarted","Data":"78a06bc03b22cf812dc81031e72adc56a73177ef4d2a02186e9c00dd4b10a5db"} Jan 26 17:12:51 crc kubenswrapper[4754]: I0126 17:12:51.779064 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="443a6550-c50d-453e-9f10-3875fcb41c18" path="/var/lib/kubelet/pods/443a6550-c50d-453e-9f10-3875fcb41c18/volumes" Jan 26 17:12:52 crc kubenswrapper[4754]: I0126 17:12:52.450768 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55478c4467-c65ql" event={"ID":"48e65463-af70-4cca-b90b-a3b4ac9a1619","Type":"ContainerStarted","Data":"e5c04e7f18c89e76b216af3bd219e7b88327004379c50141282699ba2da436dd"} Jan 26 17:12:52 crc kubenswrapper[4754]: I0126 17:12:52.451069 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-55478c4467-c65ql" Jan 26 17:12:52 crc kubenswrapper[4754]: I0126 17:12:52.474699 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-55478c4467-c65ql" podStartSLOduration=3.4746619499999998 podStartE2EDuration="3.47466195s" podCreationTimestamp="2026-01-26 17:12:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 17:12:52.468160031 +0000 UTC m=+1538.992340475" watchObservedRunningTime="2026-01-26 17:12:52.47466195 +0000 UTC m=+1538.998842384" Jan 26 17:12:59 crc kubenswrapper[4754]: I0126 17:12:59.962408 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-55478c4467-c65ql" Jan 26 17:13:00 crc kubenswrapper[4754]: I0126 17:13:00.085268 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-79bd4cc8c9-h52kf"] Jan 26 17:13:00 crc kubenswrapper[4754]: I0126 17:13:00.085940 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-79bd4cc8c9-h52kf" podUID="9cf31d29-b33a-4a3d-ab69-98094ed6364b" containerName="dnsmasq-dns" containerID="cri-o://867f10be7d6d79bb9f1fbc8fcf04eedd21efb11b71b3248822ebec5b8840613a" gracePeriod=10 Jan 26 17:13:00 crc kubenswrapper[4754]: E0126 17:13:00.180630 4754 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9cf31d29_b33a_4a3d_ab69_98094ed6364b.slice/crio-conmon-867f10be7d6d79bb9f1fbc8fcf04eedd21efb11b71b3248822ebec5b8840613a.scope\": RecentStats: unable to find data in memory cache]" Jan 26 17:13:00 crc kubenswrapper[4754]: I0126 17:13:00.530197 4754 generic.go:334] "Generic (PLEG): container finished" podID="9cf31d29-b33a-4a3d-ab69-98094ed6364b" containerID="867f10be7d6d79bb9f1fbc8fcf04eedd21efb11b71b3248822ebec5b8840613a" exitCode=0 Jan 26 17:13:00 crc kubenswrapper[4754]: I0126 17:13:00.530260 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79bd4cc8c9-h52kf" event={"ID":"9cf31d29-b33a-4a3d-ab69-98094ed6364b","Type":"ContainerDied","Data":"867f10be7d6d79bb9f1fbc8fcf04eedd21efb11b71b3248822ebec5b8840613a"} Jan 26 17:13:00 crc kubenswrapper[4754]: I0126 17:13:00.530519 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79bd4cc8c9-h52kf" event={"ID":"9cf31d29-b33a-4a3d-ab69-98094ed6364b","Type":"ContainerDied","Data":"792d3c94e10ebd7364161c71adc541a04bd5e23b0ded389e30409fe05ceaedcd"} Jan 26 17:13:00 crc kubenswrapper[4754]: I0126 17:13:00.530536 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="792d3c94e10ebd7364161c71adc541a04bd5e23b0ded389e30409fe05ceaedcd" Jan 26 17:13:00 crc kubenswrapper[4754]: I0126 17:13:00.597274 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-79bd4cc8c9-h52kf" Jan 26 17:13:00 crc kubenswrapper[4754]: I0126 17:13:00.744180 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rqlc7\" (UniqueName: \"kubernetes.io/projected/9cf31d29-b33a-4a3d-ab69-98094ed6364b-kube-api-access-rqlc7\") pod \"9cf31d29-b33a-4a3d-ab69-98094ed6364b\" (UID: \"9cf31d29-b33a-4a3d-ab69-98094ed6364b\") " Jan 26 17:13:00 crc kubenswrapper[4754]: I0126 17:13:00.744283 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/9cf31d29-b33a-4a3d-ab69-98094ed6364b-openstack-edpm-ipam\") pod \"9cf31d29-b33a-4a3d-ab69-98094ed6364b\" (UID: \"9cf31d29-b33a-4a3d-ab69-98094ed6364b\") " Jan 26 17:13:00 crc kubenswrapper[4754]: I0126 17:13:00.744317 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9cf31d29-b33a-4a3d-ab69-98094ed6364b-dns-swift-storage-0\") pod \"9cf31d29-b33a-4a3d-ab69-98094ed6364b\" (UID: \"9cf31d29-b33a-4a3d-ab69-98094ed6364b\") " Jan 26 17:13:00 crc kubenswrapper[4754]: I0126 17:13:00.744375 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9cf31d29-b33a-4a3d-ab69-98094ed6364b-ovsdbserver-nb\") pod \"9cf31d29-b33a-4a3d-ab69-98094ed6364b\" (UID: \"9cf31d29-b33a-4a3d-ab69-98094ed6364b\") " Jan 26 17:13:00 crc kubenswrapper[4754]: I0126 17:13:00.744404 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9cf31d29-b33a-4a3d-ab69-98094ed6364b-ovsdbserver-sb\") pod \"9cf31d29-b33a-4a3d-ab69-98094ed6364b\" (UID: \"9cf31d29-b33a-4a3d-ab69-98094ed6364b\") " Jan 26 17:13:00 crc kubenswrapper[4754]: I0126 17:13:00.744425 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9cf31d29-b33a-4a3d-ab69-98094ed6364b-dns-svc\") pod \"9cf31d29-b33a-4a3d-ab69-98094ed6364b\" (UID: \"9cf31d29-b33a-4a3d-ab69-98094ed6364b\") " Jan 26 17:13:00 crc kubenswrapper[4754]: I0126 17:13:00.744443 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9cf31d29-b33a-4a3d-ab69-98094ed6364b-config\") pod \"9cf31d29-b33a-4a3d-ab69-98094ed6364b\" (UID: \"9cf31d29-b33a-4a3d-ab69-98094ed6364b\") " Jan 26 17:13:00 crc kubenswrapper[4754]: I0126 17:13:00.752900 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9cf31d29-b33a-4a3d-ab69-98094ed6364b-kube-api-access-rqlc7" (OuterVolumeSpecName: "kube-api-access-rqlc7") pod "9cf31d29-b33a-4a3d-ab69-98094ed6364b" (UID: "9cf31d29-b33a-4a3d-ab69-98094ed6364b"). InnerVolumeSpecName "kube-api-access-rqlc7". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:13:00 crc kubenswrapper[4754]: I0126 17:13:00.799923 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9cf31d29-b33a-4a3d-ab69-98094ed6364b-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "9cf31d29-b33a-4a3d-ab69-98094ed6364b" (UID: "9cf31d29-b33a-4a3d-ab69-98094ed6364b"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:13:00 crc kubenswrapper[4754]: I0126 17:13:00.809711 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9cf31d29-b33a-4a3d-ab69-98094ed6364b-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "9cf31d29-b33a-4a3d-ab69-98094ed6364b" (UID: "9cf31d29-b33a-4a3d-ab69-98094ed6364b"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:13:00 crc kubenswrapper[4754]: I0126 17:13:00.812338 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9cf31d29-b33a-4a3d-ab69-98094ed6364b-config" (OuterVolumeSpecName: "config") pod "9cf31d29-b33a-4a3d-ab69-98094ed6364b" (UID: "9cf31d29-b33a-4a3d-ab69-98094ed6364b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:13:00 crc kubenswrapper[4754]: I0126 17:13:00.817042 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9cf31d29-b33a-4a3d-ab69-98094ed6364b-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "9cf31d29-b33a-4a3d-ab69-98094ed6364b" (UID: "9cf31d29-b33a-4a3d-ab69-98094ed6364b"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:13:00 crc kubenswrapper[4754]: I0126 17:13:00.819183 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9cf31d29-b33a-4a3d-ab69-98094ed6364b-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "9cf31d29-b33a-4a3d-ab69-98094ed6364b" (UID: "9cf31d29-b33a-4a3d-ab69-98094ed6364b"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:13:00 crc kubenswrapper[4754]: I0126 17:13:00.828298 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9cf31d29-b33a-4a3d-ab69-98094ed6364b-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "9cf31d29-b33a-4a3d-ab69-98094ed6364b" (UID: "9cf31d29-b33a-4a3d-ab69-98094ed6364b"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:13:00 crc kubenswrapper[4754]: I0126 17:13:00.846463 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rqlc7\" (UniqueName: \"kubernetes.io/projected/9cf31d29-b33a-4a3d-ab69-98094ed6364b-kube-api-access-rqlc7\") on node \"crc\" DevicePath \"\"" Jan 26 17:13:00 crc kubenswrapper[4754]: I0126 17:13:00.846504 4754 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/9cf31d29-b33a-4a3d-ab69-98094ed6364b-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Jan 26 17:13:00 crc kubenswrapper[4754]: I0126 17:13:00.846513 4754 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9cf31d29-b33a-4a3d-ab69-98094ed6364b-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Jan 26 17:13:00 crc kubenswrapper[4754]: I0126 17:13:00.846524 4754 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9cf31d29-b33a-4a3d-ab69-98094ed6364b-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Jan 26 17:13:00 crc kubenswrapper[4754]: I0126 17:13:00.846533 4754 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9cf31d29-b33a-4a3d-ab69-98094ed6364b-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Jan 26 17:13:00 crc kubenswrapper[4754]: I0126 17:13:00.846542 4754 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9cf31d29-b33a-4a3d-ab69-98094ed6364b-dns-svc\") on node \"crc\" DevicePath \"\"" Jan 26 17:13:00 crc kubenswrapper[4754]: I0126 17:13:00.846551 4754 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9cf31d29-b33a-4a3d-ab69-98094ed6364b-config\") on node \"crc\" DevicePath \"\"" Jan 26 17:13:01 crc kubenswrapper[4754]: I0126 17:13:01.539167 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-79bd4cc8c9-h52kf" Jan 26 17:13:01 crc kubenswrapper[4754]: I0126 17:13:01.582228 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-79bd4cc8c9-h52kf"] Jan 26 17:13:01 crc kubenswrapper[4754]: I0126 17:13:01.592383 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-79bd4cc8c9-h52kf"] Jan 26 17:13:01 crc kubenswrapper[4754]: I0126 17:13:01.785565 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9cf31d29-b33a-4a3d-ab69-98094ed6364b" path="/var/lib/kubelet/pods/9cf31d29-b33a-4a3d-ab69-98094ed6364b/volumes" Jan 26 17:13:07 crc kubenswrapper[4754]: I0126 17:13:07.129054 4754 patch_prober.go:28] interesting pod/machine-config-daemon-x65wv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 26 17:13:07 crc kubenswrapper[4754]: I0126 17:13:07.130492 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 26 17:13:07 crc kubenswrapper[4754]: I0126 17:13:07.130536 4754 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" Jan 26 17:13:07 crc kubenswrapper[4754]: I0126 17:13:07.131296 4754 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"19eac0e87afaeef93088a8556aa836389e2ebff93a96b482df571cb7b852f0b7"} pod="openshift-machine-config-operator/machine-config-daemon-x65wv" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 26 17:13:07 crc kubenswrapper[4754]: I0126 17:13:07.131359 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" containerName="machine-config-daemon" containerID="cri-o://19eac0e87afaeef93088a8556aa836389e2ebff93a96b482df571cb7b852f0b7" gracePeriod=600 Jan 26 17:13:07 crc kubenswrapper[4754]: E0126 17:13:07.252558 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x65wv_openshift-machine-config-operator(8c3718a4-f354-4284-92e0-fdfb45a692bd)\"" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" Jan 26 17:13:07 crc kubenswrapper[4754]: I0126 17:13:07.605229 4754 generic.go:334] "Generic (PLEG): container finished" podID="8c3718a4-f354-4284-92e0-fdfb45a692bd" containerID="19eac0e87afaeef93088a8556aa836389e2ebff93a96b482df571cb7b852f0b7" exitCode=0 Jan 26 17:13:07 crc kubenswrapper[4754]: I0126 17:13:07.605279 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" event={"ID":"8c3718a4-f354-4284-92e0-fdfb45a692bd","Type":"ContainerDied","Data":"19eac0e87afaeef93088a8556aa836389e2ebff93a96b482df571cb7b852f0b7"} Jan 26 17:13:07 crc kubenswrapper[4754]: I0126 17:13:07.605314 4754 scope.go:117] "RemoveContainer" containerID="7bdda0a504a055c787cfe163a27b6598986e7e593ba6562f6669a36ea09cda24" Jan 26 17:13:07 crc kubenswrapper[4754]: I0126 17:13:07.606024 4754 scope.go:117] "RemoveContainer" containerID="19eac0e87afaeef93088a8556aa836389e2ebff93a96b482df571cb7b852f0b7" Jan 26 17:13:07 crc kubenswrapper[4754]: E0126 17:13:07.606463 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x65wv_openshift-machine-config-operator(8c3718a4-f354-4284-92e0-fdfb45a692bd)\"" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" Jan 26 17:13:11 crc kubenswrapper[4754]: I0126 17:13:11.646875 4754 generic.go:334] "Generic (PLEG): container finished" podID="6e2da069-1c54-4801-a91c-241b80e8d17b" containerID="667b4c065827e6c2ad1e3c6a95979170a8ced1f3ee0ba5771ec1ce4235e28faa" exitCode=0 Jan 26 17:13:11 crc kubenswrapper[4754]: I0126 17:13:11.646972 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"6e2da069-1c54-4801-a91c-241b80e8d17b","Type":"ContainerDied","Data":"667b4c065827e6c2ad1e3c6a95979170a8ced1f3ee0ba5771ec1ce4235e28faa"} Jan 26 17:13:12 crc kubenswrapper[4754]: I0126 17:13:12.239128 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-gn6jw"] Jan 26 17:13:12 crc kubenswrapper[4754]: E0126 17:13:12.239863 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="443a6550-c50d-453e-9f10-3875fcb41c18" containerName="init" Jan 26 17:13:12 crc kubenswrapper[4754]: I0126 17:13:12.239886 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="443a6550-c50d-453e-9f10-3875fcb41c18" containerName="init" Jan 26 17:13:12 crc kubenswrapper[4754]: E0126 17:13:12.239917 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9cf31d29-b33a-4a3d-ab69-98094ed6364b" containerName="dnsmasq-dns" Jan 26 17:13:12 crc kubenswrapper[4754]: I0126 17:13:12.239925 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="9cf31d29-b33a-4a3d-ab69-98094ed6364b" containerName="dnsmasq-dns" Jan 26 17:13:12 crc kubenswrapper[4754]: E0126 17:13:12.239941 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9cf31d29-b33a-4a3d-ab69-98094ed6364b" containerName="init" Jan 26 17:13:12 crc kubenswrapper[4754]: I0126 17:13:12.239949 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="9cf31d29-b33a-4a3d-ab69-98094ed6364b" containerName="init" Jan 26 17:13:12 crc kubenswrapper[4754]: E0126 17:13:12.239963 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="443a6550-c50d-453e-9f10-3875fcb41c18" containerName="dnsmasq-dns" Jan 26 17:13:12 crc kubenswrapper[4754]: I0126 17:13:12.239969 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="443a6550-c50d-453e-9f10-3875fcb41c18" containerName="dnsmasq-dns" Jan 26 17:13:12 crc kubenswrapper[4754]: I0126 17:13:12.240174 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="443a6550-c50d-453e-9f10-3875fcb41c18" containerName="dnsmasq-dns" Jan 26 17:13:12 crc kubenswrapper[4754]: I0126 17:13:12.240190 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="9cf31d29-b33a-4a3d-ab69-98094ed6364b" containerName="dnsmasq-dns" Jan 26 17:13:12 crc kubenswrapper[4754]: I0126 17:13:12.240885 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-gn6jw" Jan 26 17:13:12 crc kubenswrapper[4754]: I0126 17:13:12.243623 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Jan 26 17:13:12 crc kubenswrapper[4754]: I0126 17:13:12.243922 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Jan 26 17:13:12 crc kubenswrapper[4754]: I0126 17:13:12.244215 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-fz7z8" Jan 26 17:13:12 crc kubenswrapper[4754]: I0126 17:13:12.246132 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Jan 26 17:13:12 crc kubenswrapper[4754]: I0126 17:13:12.257368 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-gn6jw"] Jan 26 17:13:12 crc kubenswrapper[4754]: I0126 17:13:12.367237 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/2d80ff89-c1fc-4331-9683-2740c69d001d-ssh-key-openstack-edpm-ipam\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-gn6jw\" (UID: \"2d80ff89-c1fc-4331-9683-2740c69d001d\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-gn6jw" Jan 26 17:13:12 crc kubenswrapper[4754]: I0126 17:13:12.367296 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d80ff89-c1fc-4331-9683-2740c69d001d-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-gn6jw\" (UID: \"2d80ff89-c1fc-4331-9683-2740c69d001d\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-gn6jw" Jan 26 17:13:12 crc kubenswrapper[4754]: I0126 17:13:12.367319 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2d80ff89-c1fc-4331-9683-2740c69d001d-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-gn6jw\" (UID: \"2d80ff89-c1fc-4331-9683-2740c69d001d\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-gn6jw" Jan 26 17:13:12 crc kubenswrapper[4754]: I0126 17:13:12.367840 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nrp8t\" (UniqueName: \"kubernetes.io/projected/2d80ff89-c1fc-4331-9683-2740c69d001d-kube-api-access-nrp8t\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-gn6jw\" (UID: \"2d80ff89-c1fc-4331-9683-2740c69d001d\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-gn6jw" Jan 26 17:13:12 crc kubenswrapper[4754]: I0126 17:13:12.474727 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nrp8t\" (UniqueName: \"kubernetes.io/projected/2d80ff89-c1fc-4331-9683-2740c69d001d-kube-api-access-nrp8t\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-gn6jw\" (UID: \"2d80ff89-c1fc-4331-9683-2740c69d001d\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-gn6jw" Jan 26 17:13:12 crc kubenswrapper[4754]: I0126 17:13:12.474831 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/2d80ff89-c1fc-4331-9683-2740c69d001d-ssh-key-openstack-edpm-ipam\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-gn6jw\" (UID: \"2d80ff89-c1fc-4331-9683-2740c69d001d\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-gn6jw" Jan 26 17:13:12 crc kubenswrapper[4754]: I0126 17:13:12.474856 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d80ff89-c1fc-4331-9683-2740c69d001d-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-gn6jw\" (UID: \"2d80ff89-c1fc-4331-9683-2740c69d001d\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-gn6jw" Jan 26 17:13:12 crc kubenswrapper[4754]: I0126 17:13:12.474881 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2d80ff89-c1fc-4331-9683-2740c69d001d-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-gn6jw\" (UID: \"2d80ff89-c1fc-4331-9683-2740c69d001d\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-gn6jw" Jan 26 17:13:12 crc kubenswrapper[4754]: I0126 17:13:12.479686 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/2d80ff89-c1fc-4331-9683-2740c69d001d-ssh-key-openstack-edpm-ipam\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-gn6jw\" (UID: \"2d80ff89-c1fc-4331-9683-2740c69d001d\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-gn6jw" Jan 26 17:13:12 crc kubenswrapper[4754]: I0126 17:13:12.480033 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2d80ff89-c1fc-4331-9683-2740c69d001d-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-gn6jw\" (UID: \"2d80ff89-c1fc-4331-9683-2740c69d001d\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-gn6jw" Jan 26 17:13:12 crc kubenswrapper[4754]: I0126 17:13:12.480199 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d80ff89-c1fc-4331-9683-2740c69d001d-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-gn6jw\" (UID: \"2d80ff89-c1fc-4331-9683-2740c69d001d\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-gn6jw" Jan 26 17:13:12 crc kubenswrapper[4754]: I0126 17:13:12.500259 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nrp8t\" (UniqueName: \"kubernetes.io/projected/2d80ff89-c1fc-4331-9683-2740c69d001d-kube-api-access-nrp8t\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-gn6jw\" (UID: \"2d80ff89-c1fc-4331-9683-2740c69d001d\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-gn6jw" Jan 26 17:13:12 crc kubenswrapper[4754]: I0126 17:13:12.559381 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-gn6jw" Jan 26 17:13:12 crc kubenswrapper[4754]: I0126 17:13:12.668649 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"6e2da069-1c54-4801-a91c-241b80e8d17b","Type":"ContainerStarted","Data":"908f049c32633476a654616ab212b6ba869a8b5c8cd214437b8c8b2ca1ebafc8"} Jan 26 17:13:12 crc kubenswrapper[4754]: I0126 17:13:12.670108 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Jan 26 17:13:12 crc kubenswrapper[4754]: I0126 17:13:12.715195 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=36.715173698 podStartE2EDuration="36.715173698s" podCreationTimestamp="2026-01-26 17:12:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 17:13:12.700316479 +0000 UTC m=+1559.224496923" watchObservedRunningTime="2026-01-26 17:13:12.715173698 +0000 UTC m=+1559.239354132" Jan 26 17:13:13 crc kubenswrapper[4754]: I0126 17:13:13.084529 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-gn6jw"] Jan 26 17:13:13 crc kubenswrapper[4754]: I0126 17:13:13.679473 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-gn6jw" event={"ID":"2d80ff89-c1fc-4331-9683-2740c69d001d","Type":"ContainerStarted","Data":"7e6f00ef9cf642bae0ca5c9d187384e5aa2942142ff8ca0a6280ee9a88ebbeb1"} Jan 26 17:13:13 crc kubenswrapper[4754]: I0126 17:13:13.682206 4754 generic.go:334] "Generic (PLEG): container finished" podID="9acb97f6-82ac-4891-96dc-43a85f9c4e7c" containerID="e3dc5e85d5a3c46330918b93a52510e835514f6976f66406c0fbc1b543517d78" exitCode=0 Jan 26 17:13:13 crc kubenswrapper[4754]: I0126 17:13:13.683142 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"9acb97f6-82ac-4891-96dc-43a85f9c4e7c","Type":"ContainerDied","Data":"e3dc5e85d5a3c46330918b93a52510e835514f6976f66406c0fbc1b543517d78"} Jan 26 17:13:14 crc kubenswrapper[4754]: I0126 17:13:14.693721 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"9acb97f6-82ac-4891-96dc-43a85f9c4e7c","Type":"ContainerStarted","Data":"595ab08e25aa679dcd8f1830952958e214655abf97cade8026cc323d7dd752ee"} Jan 26 17:13:14 crc kubenswrapper[4754]: I0126 17:13:14.694402 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Jan 26 17:13:14 crc kubenswrapper[4754]: I0126 17:13:14.726258 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=36.726240561 podStartE2EDuration="36.726240561s" podCreationTimestamp="2026-01-26 17:12:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 17:13:14.714685253 +0000 UTC m=+1561.238865707" watchObservedRunningTime="2026-01-26 17:13:14.726240561 +0000 UTC m=+1561.250420995" Jan 26 17:13:17 crc kubenswrapper[4754]: I0126 17:13:17.190568 4754 scope.go:117] "RemoveContainer" containerID="fe88ff978302d35023844df531d072c3bf22651302bb9dd2dddce4e3e4df7ecc" Jan 26 17:13:21 crc kubenswrapper[4754]: I0126 17:13:21.767704 4754 scope.go:117] "RemoveContainer" containerID="19eac0e87afaeef93088a8556aa836389e2ebff93a96b482df571cb7b852f0b7" Jan 26 17:13:21 crc kubenswrapper[4754]: E0126 17:13:21.768505 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x65wv_openshift-machine-config-operator(8c3718a4-f354-4284-92e0-fdfb45a692bd)\"" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" Jan 26 17:13:22 crc kubenswrapper[4754]: I0126 17:13:22.683342 4754 scope.go:117] "RemoveContainer" containerID="ee7fb2afad886aa0086afd0f8654f9b549226222264cb0f25e231475789ed4b7" Jan 26 17:13:22 crc kubenswrapper[4754]: I0126 17:13:22.729523 4754 scope.go:117] "RemoveContainer" containerID="877b01f8012ec09c2df866b01599ea7ed3a790707891024efe46f3aaca9cc418" Jan 26 17:13:22 crc kubenswrapper[4754]: I0126 17:13:22.776450 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Jan 26 17:13:22 crc kubenswrapper[4754]: I0126 17:13:22.783605 4754 scope.go:117] "RemoveContainer" containerID="420afe55ba4cf912e47db77c3348ccb28bb43686b51ba463c7e761755d35e137" Jan 26 17:13:22 crc kubenswrapper[4754]: I0126 17:13:22.910332 4754 scope.go:117] "RemoveContainer" containerID="b8309f296df385f6aabf8972db308f5cb5f3d17557ebc9629b2a87cdec9cc273" Jan 26 17:13:22 crc kubenswrapper[4754]: I0126 17:13:22.968273 4754 scope.go:117] "RemoveContainer" containerID="18791aca8574779fa910ccd041b92873e7acc45391f6423c1c9e799ef87424a0" Jan 26 17:13:22 crc kubenswrapper[4754]: I0126 17:13:22.989234 4754 scope.go:117] "RemoveContainer" containerID="316130fcccc0bf944a3b73afaa9f2d0f8d0d8fa2c0ccc61283a5ae24c9d6366c" Jan 26 17:13:23 crc kubenswrapper[4754]: I0126 17:13:23.016573 4754 scope.go:117] "RemoveContainer" containerID="33a0ab55bbd24b76d21481a4d6e3bb9482f39506cf37531f314c5be370502159" Jan 26 17:13:23 crc kubenswrapper[4754]: I0126 17:13:23.043659 4754 scope.go:117] "RemoveContainer" containerID="890780a4f6df86cd0474c7787d35ee7d2ca45b9ed51a3fabc57b69c719947625" Jan 26 17:13:23 crc kubenswrapper[4754]: I0126 17:13:23.817784 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-gn6jw" event={"ID":"2d80ff89-c1fc-4331-9683-2740c69d001d","Type":"ContainerStarted","Data":"d367716d71a8c14ceb5ebd4f719871d831b7150c060ead66829167212b860c22"} Jan 26 17:13:23 crc kubenswrapper[4754]: I0126 17:13:23.836655 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-gn6jw" podStartSLOduration=2.168077609 podStartE2EDuration="11.836638713s" podCreationTimestamp="2026-01-26 17:13:12 +0000 UTC" firstStartedPulling="2026-01-26 17:13:13.103772088 +0000 UTC m=+1559.627952522" lastFinishedPulling="2026-01-26 17:13:22.772333192 +0000 UTC m=+1569.296513626" observedRunningTime="2026-01-26 17:13:23.835753529 +0000 UTC m=+1570.359933973" watchObservedRunningTime="2026-01-26 17:13:23.836638713 +0000 UTC m=+1570.360819147" Jan 26 17:13:26 crc kubenswrapper[4754]: I0126 17:13:26.957871 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Jan 26 17:13:28 crc kubenswrapper[4754]: I0126 17:13:28.898447 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Jan 26 17:13:31 crc kubenswrapper[4754]: I0126 17:13:31.868552 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-psxqw"] Jan 26 17:13:31 crc kubenswrapper[4754]: I0126 17:13:31.870805 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-psxqw" Jan 26 17:13:31 crc kubenswrapper[4754]: I0126 17:13:31.898960 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-psxqw"] Jan 26 17:13:31 crc kubenswrapper[4754]: I0126 17:13:31.988845 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/41e45928-06b9-4854-9061-16e4053b0fb1-utilities\") pod \"redhat-marketplace-psxqw\" (UID: \"41e45928-06b9-4854-9061-16e4053b0fb1\") " pod="openshift-marketplace/redhat-marketplace-psxqw" Jan 26 17:13:31 crc kubenswrapper[4754]: I0126 17:13:31.988961 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4pgfb\" (UniqueName: \"kubernetes.io/projected/41e45928-06b9-4854-9061-16e4053b0fb1-kube-api-access-4pgfb\") pod \"redhat-marketplace-psxqw\" (UID: \"41e45928-06b9-4854-9061-16e4053b0fb1\") " pod="openshift-marketplace/redhat-marketplace-psxqw" Jan 26 17:13:31 crc kubenswrapper[4754]: I0126 17:13:31.989045 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/41e45928-06b9-4854-9061-16e4053b0fb1-catalog-content\") pod \"redhat-marketplace-psxqw\" (UID: \"41e45928-06b9-4854-9061-16e4053b0fb1\") " pod="openshift-marketplace/redhat-marketplace-psxqw" Jan 26 17:13:32 crc kubenswrapper[4754]: I0126 17:13:32.091051 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4pgfb\" (UniqueName: \"kubernetes.io/projected/41e45928-06b9-4854-9061-16e4053b0fb1-kube-api-access-4pgfb\") pod \"redhat-marketplace-psxqw\" (UID: \"41e45928-06b9-4854-9061-16e4053b0fb1\") " pod="openshift-marketplace/redhat-marketplace-psxqw" Jan 26 17:13:32 crc kubenswrapper[4754]: I0126 17:13:32.091158 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/41e45928-06b9-4854-9061-16e4053b0fb1-catalog-content\") pod \"redhat-marketplace-psxqw\" (UID: \"41e45928-06b9-4854-9061-16e4053b0fb1\") " pod="openshift-marketplace/redhat-marketplace-psxqw" Jan 26 17:13:32 crc kubenswrapper[4754]: I0126 17:13:32.091272 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/41e45928-06b9-4854-9061-16e4053b0fb1-utilities\") pod \"redhat-marketplace-psxqw\" (UID: \"41e45928-06b9-4854-9061-16e4053b0fb1\") " pod="openshift-marketplace/redhat-marketplace-psxqw" Jan 26 17:13:32 crc kubenswrapper[4754]: I0126 17:13:32.091628 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/41e45928-06b9-4854-9061-16e4053b0fb1-catalog-content\") pod \"redhat-marketplace-psxqw\" (UID: \"41e45928-06b9-4854-9061-16e4053b0fb1\") " pod="openshift-marketplace/redhat-marketplace-psxqw" Jan 26 17:13:32 crc kubenswrapper[4754]: I0126 17:13:32.091684 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/41e45928-06b9-4854-9061-16e4053b0fb1-utilities\") pod \"redhat-marketplace-psxqw\" (UID: \"41e45928-06b9-4854-9061-16e4053b0fb1\") " pod="openshift-marketplace/redhat-marketplace-psxqw" Jan 26 17:13:32 crc kubenswrapper[4754]: I0126 17:13:32.120692 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4pgfb\" (UniqueName: \"kubernetes.io/projected/41e45928-06b9-4854-9061-16e4053b0fb1-kube-api-access-4pgfb\") pod \"redhat-marketplace-psxqw\" (UID: \"41e45928-06b9-4854-9061-16e4053b0fb1\") " pod="openshift-marketplace/redhat-marketplace-psxqw" Jan 26 17:13:32 crc kubenswrapper[4754]: I0126 17:13:32.211339 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-psxqw" Jan 26 17:13:32 crc kubenswrapper[4754]: I0126 17:13:32.684410 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-psxqw"] Jan 26 17:13:32 crc kubenswrapper[4754]: W0126 17:13:32.687844 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod41e45928_06b9_4854_9061_16e4053b0fb1.slice/crio-07502917c23bee9260c0a4e664053315ac37248137a54c44c4c683a0f3a2fb77 WatchSource:0}: Error finding container 07502917c23bee9260c0a4e664053315ac37248137a54c44c4c683a0f3a2fb77: Status 404 returned error can't find the container with id 07502917c23bee9260c0a4e664053315ac37248137a54c44c4c683a0f3a2fb77 Jan 26 17:13:32 crc kubenswrapper[4754]: I0126 17:13:32.900029 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-psxqw" event={"ID":"41e45928-06b9-4854-9061-16e4053b0fb1","Type":"ContainerStarted","Data":"07502917c23bee9260c0a4e664053315ac37248137a54c44c4c683a0f3a2fb77"} Jan 26 17:13:33 crc kubenswrapper[4754]: I0126 17:13:33.909312 4754 generic.go:334] "Generic (PLEG): container finished" podID="41e45928-06b9-4854-9061-16e4053b0fb1" containerID="7c0abec0fb9f216ff3607b408cf83f598056e0e1f76ed6150476358682803cbb" exitCode=0 Jan 26 17:13:33 crc kubenswrapper[4754]: I0126 17:13:33.909372 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-psxqw" event={"ID":"41e45928-06b9-4854-9061-16e4053b0fb1","Type":"ContainerDied","Data":"7c0abec0fb9f216ff3607b408cf83f598056e0e1f76ed6150476358682803cbb"} Jan 26 17:13:35 crc kubenswrapper[4754]: I0126 17:13:35.768148 4754 scope.go:117] "RemoveContainer" containerID="19eac0e87afaeef93088a8556aa836389e2ebff93a96b482df571cb7b852f0b7" Jan 26 17:13:35 crc kubenswrapper[4754]: E0126 17:13:35.768568 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x65wv_openshift-machine-config-operator(8c3718a4-f354-4284-92e0-fdfb45a692bd)\"" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" Jan 26 17:13:37 crc kubenswrapper[4754]: I0126 17:13:37.948115 4754 generic.go:334] "Generic (PLEG): container finished" podID="2d80ff89-c1fc-4331-9683-2740c69d001d" containerID="d367716d71a8c14ceb5ebd4f719871d831b7150c060ead66829167212b860c22" exitCode=0 Jan 26 17:13:37 crc kubenswrapper[4754]: I0126 17:13:37.948231 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-gn6jw" event={"ID":"2d80ff89-c1fc-4331-9683-2740c69d001d","Type":"ContainerDied","Data":"d367716d71a8c14ceb5ebd4f719871d831b7150c060ead66829167212b860c22"} Jan 26 17:13:38 crc kubenswrapper[4754]: I0126 17:13:38.958792 4754 generic.go:334] "Generic (PLEG): container finished" podID="41e45928-06b9-4854-9061-16e4053b0fb1" containerID="2dfa1158b61bfd4ef60a6d8bfe54dc3f9125b8f860498c7fd2a186296526ccef" exitCode=0 Jan 26 17:13:38 crc kubenswrapper[4754]: I0126 17:13:38.958854 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-psxqw" event={"ID":"41e45928-06b9-4854-9061-16e4053b0fb1","Type":"ContainerDied","Data":"2dfa1158b61bfd4ef60a6d8bfe54dc3f9125b8f860498c7fd2a186296526ccef"} Jan 26 17:13:39 crc kubenswrapper[4754]: I0126 17:13:39.387080 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-gn6jw" Jan 26 17:13:39 crc kubenswrapper[4754]: I0126 17:13:39.523380 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d80ff89-c1fc-4331-9683-2740c69d001d-repo-setup-combined-ca-bundle\") pod \"2d80ff89-c1fc-4331-9683-2740c69d001d\" (UID: \"2d80ff89-c1fc-4331-9683-2740c69d001d\") " Jan 26 17:13:39 crc kubenswrapper[4754]: I0126 17:13:39.523800 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nrp8t\" (UniqueName: \"kubernetes.io/projected/2d80ff89-c1fc-4331-9683-2740c69d001d-kube-api-access-nrp8t\") pod \"2d80ff89-c1fc-4331-9683-2740c69d001d\" (UID: \"2d80ff89-c1fc-4331-9683-2740c69d001d\") " Jan 26 17:13:39 crc kubenswrapper[4754]: I0126 17:13:39.523927 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/2d80ff89-c1fc-4331-9683-2740c69d001d-ssh-key-openstack-edpm-ipam\") pod \"2d80ff89-c1fc-4331-9683-2740c69d001d\" (UID: \"2d80ff89-c1fc-4331-9683-2740c69d001d\") " Jan 26 17:13:39 crc kubenswrapper[4754]: I0126 17:13:39.523971 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2d80ff89-c1fc-4331-9683-2740c69d001d-inventory\") pod \"2d80ff89-c1fc-4331-9683-2740c69d001d\" (UID: \"2d80ff89-c1fc-4331-9683-2740c69d001d\") " Jan 26 17:13:39 crc kubenswrapper[4754]: I0126 17:13:39.529492 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2d80ff89-c1fc-4331-9683-2740c69d001d-kube-api-access-nrp8t" (OuterVolumeSpecName: "kube-api-access-nrp8t") pod "2d80ff89-c1fc-4331-9683-2740c69d001d" (UID: "2d80ff89-c1fc-4331-9683-2740c69d001d"). InnerVolumeSpecName "kube-api-access-nrp8t". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:13:39 crc kubenswrapper[4754]: I0126 17:13:39.530086 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2d80ff89-c1fc-4331-9683-2740c69d001d-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "2d80ff89-c1fc-4331-9683-2740c69d001d" (UID: "2d80ff89-c1fc-4331-9683-2740c69d001d"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:13:39 crc kubenswrapper[4754]: I0126 17:13:39.556356 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2d80ff89-c1fc-4331-9683-2740c69d001d-inventory" (OuterVolumeSpecName: "inventory") pod "2d80ff89-c1fc-4331-9683-2740c69d001d" (UID: "2d80ff89-c1fc-4331-9683-2740c69d001d"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:13:39 crc kubenswrapper[4754]: I0126 17:13:39.563173 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2d80ff89-c1fc-4331-9683-2740c69d001d-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "2d80ff89-c1fc-4331-9683-2740c69d001d" (UID: "2d80ff89-c1fc-4331-9683-2740c69d001d"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:13:39 crc kubenswrapper[4754]: I0126 17:13:39.626357 4754 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/2d80ff89-c1fc-4331-9683-2740c69d001d-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Jan 26 17:13:39 crc kubenswrapper[4754]: I0126 17:13:39.626405 4754 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2d80ff89-c1fc-4331-9683-2740c69d001d-inventory\") on node \"crc\" DevicePath \"\"" Jan 26 17:13:39 crc kubenswrapper[4754]: I0126 17:13:39.626418 4754 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d80ff89-c1fc-4331-9683-2740c69d001d-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 26 17:13:39 crc kubenswrapper[4754]: I0126 17:13:39.626430 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nrp8t\" (UniqueName: \"kubernetes.io/projected/2d80ff89-c1fc-4331-9683-2740c69d001d-kube-api-access-nrp8t\") on node \"crc\" DevicePath \"\"" Jan 26 17:13:39 crc kubenswrapper[4754]: I0126 17:13:39.969801 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-gn6jw" event={"ID":"2d80ff89-c1fc-4331-9683-2740c69d001d","Type":"ContainerDied","Data":"7e6f00ef9cf642bae0ca5c9d187384e5aa2942142ff8ca0a6280ee9a88ebbeb1"} Jan 26 17:13:39 crc kubenswrapper[4754]: I0126 17:13:39.969856 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7e6f00ef9cf642bae0ca5c9d187384e5aa2942142ff8ca0a6280ee9a88ebbeb1" Jan 26 17:13:39 crc kubenswrapper[4754]: I0126 17:13:39.969853 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-gn6jw" Jan 26 17:13:39 crc kubenswrapper[4754]: I0126 17:13:39.974067 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-psxqw" event={"ID":"41e45928-06b9-4854-9061-16e4053b0fb1","Type":"ContainerStarted","Data":"1aaf8b0b7184d28bc4bf839bf6445ddb23ff15a4eb3356cbb0a6745737762525"} Jan 26 17:13:40 crc kubenswrapper[4754]: I0126 17:13:40.009267 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-psxqw" podStartSLOduration=3.535327947 podStartE2EDuration="9.009248562s" podCreationTimestamp="2026-01-26 17:13:31 +0000 UTC" firstStartedPulling="2026-01-26 17:13:33.912263986 +0000 UTC m=+1580.436444430" lastFinishedPulling="2026-01-26 17:13:39.386184611 +0000 UTC m=+1585.910365045" observedRunningTime="2026-01-26 17:13:39.994242629 +0000 UTC m=+1586.518423063" watchObservedRunningTime="2026-01-26 17:13:40.009248562 +0000 UTC m=+1586.533428996" Jan 26 17:13:40 crc kubenswrapper[4754]: I0126 17:13:40.053541 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-6w45w"] Jan 26 17:13:40 crc kubenswrapper[4754]: E0126 17:13:40.053934 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d80ff89-c1fc-4331-9683-2740c69d001d" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Jan 26 17:13:40 crc kubenswrapper[4754]: I0126 17:13:40.053956 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d80ff89-c1fc-4331-9683-2740c69d001d" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Jan 26 17:13:40 crc kubenswrapper[4754]: I0126 17:13:40.054131 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="2d80ff89-c1fc-4331-9683-2740c69d001d" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Jan 26 17:13:40 crc kubenswrapper[4754]: I0126 17:13:40.054694 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-6w45w" Jan 26 17:13:40 crc kubenswrapper[4754]: I0126 17:13:40.056720 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-fz7z8" Jan 26 17:13:40 crc kubenswrapper[4754]: I0126 17:13:40.056746 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Jan 26 17:13:40 crc kubenswrapper[4754]: I0126 17:13:40.057049 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Jan 26 17:13:40 crc kubenswrapper[4754]: I0126 17:13:40.057490 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Jan 26 17:13:40 crc kubenswrapper[4754]: I0126 17:13:40.067528 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-6w45w"] Jan 26 17:13:40 crc kubenswrapper[4754]: I0126 17:13:40.237737 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/09f02458-d4e3-49d4-8735-467141a57b6c-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-6w45w\" (UID: \"09f02458-d4e3-49d4-8735-467141a57b6c\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-6w45w" Jan 26 17:13:40 crc kubenswrapper[4754]: I0126 17:13:40.237835 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/09f02458-d4e3-49d4-8735-467141a57b6c-ssh-key-openstack-edpm-ipam\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-6w45w\" (UID: \"09f02458-d4e3-49d4-8735-467141a57b6c\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-6w45w" Jan 26 17:13:40 crc kubenswrapper[4754]: I0126 17:13:40.237897 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vwhbp\" (UniqueName: \"kubernetes.io/projected/09f02458-d4e3-49d4-8735-467141a57b6c-kube-api-access-vwhbp\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-6w45w\" (UID: \"09f02458-d4e3-49d4-8735-467141a57b6c\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-6w45w" Jan 26 17:13:40 crc kubenswrapper[4754]: I0126 17:13:40.339215 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vwhbp\" (UniqueName: \"kubernetes.io/projected/09f02458-d4e3-49d4-8735-467141a57b6c-kube-api-access-vwhbp\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-6w45w\" (UID: \"09f02458-d4e3-49d4-8735-467141a57b6c\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-6w45w" Jan 26 17:13:40 crc kubenswrapper[4754]: I0126 17:13:40.339348 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/09f02458-d4e3-49d4-8735-467141a57b6c-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-6w45w\" (UID: \"09f02458-d4e3-49d4-8735-467141a57b6c\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-6w45w" Jan 26 17:13:40 crc kubenswrapper[4754]: I0126 17:13:40.339406 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/09f02458-d4e3-49d4-8735-467141a57b6c-ssh-key-openstack-edpm-ipam\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-6w45w\" (UID: \"09f02458-d4e3-49d4-8735-467141a57b6c\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-6w45w" Jan 26 17:13:40 crc kubenswrapper[4754]: I0126 17:13:40.344906 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/09f02458-d4e3-49d4-8735-467141a57b6c-ssh-key-openstack-edpm-ipam\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-6w45w\" (UID: \"09f02458-d4e3-49d4-8735-467141a57b6c\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-6w45w" Jan 26 17:13:40 crc kubenswrapper[4754]: I0126 17:13:40.345187 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/09f02458-d4e3-49d4-8735-467141a57b6c-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-6w45w\" (UID: \"09f02458-d4e3-49d4-8735-467141a57b6c\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-6w45w" Jan 26 17:13:40 crc kubenswrapper[4754]: I0126 17:13:40.357803 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vwhbp\" (UniqueName: \"kubernetes.io/projected/09f02458-d4e3-49d4-8735-467141a57b6c-kube-api-access-vwhbp\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-6w45w\" (UID: \"09f02458-d4e3-49d4-8735-467141a57b6c\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-6w45w" Jan 26 17:13:40 crc kubenswrapper[4754]: I0126 17:13:40.370890 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-6w45w" Jan 26 17:13:40 crc kubenswrapper[4754]: W0126 17:13:40.866156 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod09f02458_d4e3_49d4_8735_467141a57b6c.slice/crio-207a2252abc5ec3807b9076db33adaf14ebce75d0ca945c6fd63fbf920d583c8 WatchSource:0}: Error finding container 207a2252abc5ec3807b9076db33adaf14ebce75d0ca945c6fd63fbf920d583c8: Status 404 returned error can't find the container with id 207a2252abc5ec3807b9076db33adaf14ebce75d0ca945c6fd63fbf920d583c8 Jan 26 17:13:40 crc kubenswrapper[4754]: I0126 17:13:40.867225 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-6w45w"] Jan 26 17:13:40 crc kubenswrapper[4754]: I0126 17:13:40.982319 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-6w45w" event={"ID":"09f02458-d4e3-49d4-8735-467141a57b6c","Type":"ContainerStarted","Data":"207a2252abc5ec3807b9076db33adaf14ebce75d0ca945c6fd63fbf920d583c8"} Jan 26 17:13:41 crc kubenswrapper[4754]: I0126 17:13:41.992330 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-6w45w" event={"ID":"09f02458-d4e3-49d4-8735-467141a57b6c","Type":"ContainerStarted","Data":"825815769fbdbb9171aa38fbed2026fdcd77127d5bf52fd0dad4cdc6dad3bad9"} Jan 26 17:13:42 crc kubenswrapper[4754]: I0126 17:13:42.015827 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-6w45w" podStartSLOduration=1.57600005 podStartE2EDuration="2.015807041s" podCreationTimestamp="2026-01-26 17:13:40 +0000 UTC" firstStartedPulling="2026-01-26 17:13:40.868175813 +0000 UTC m=+1587.392356247" lastFinishedPulling="2026-01-26 17:13:41.307982794 +0000 UTC m=+1587.832163238" observedRunningTime="2026-01-26 17:13:42.010101614 +0000 UTC m=+1588.534282058" watchObservedRunningTime="2026-01-26 17:13:42.015807041 +0000 UTC m=+1588.539987495" Jan 26 17:13:42 crc kubenswrapper[4754]: I0126 17:13:42.212170 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-psxqw" Jan 26 17:13:42 crc kubenswrapper[4754]: I0126 17:13:42.212551 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-psxqw" Jan 26 17:13:42 crc kubenswrapper[4754]: I0126 17:13:42.282937 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-psxqw" Jan 26 17:13:45 crc kubenswrapper[4754]: I0126 17:13:45.015974 4754 generic.go:334] "Generic (PLEG): container finished" podID="09f02458-d4e3-49d4-8735-467141a57b6c" containerID="825815769fbdbb9171aa38fbed2026fdcd77127d5bf52fd0dad4cdc6dad3bad9" exitCode=0 Jan 26 17:13:45 crc kubenswrapper[4754]: I0126 17:13:45.016062 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-6w45w" event={"ID":"09f02458-d4e3-49d4-8735-467141a57b6c","Type":"ContainerDied","Data":"825815769fbdbb9171aa38fbed2026fdcd77127d5bf52fd0dad4cdc6dad3bad9"} Jan 26 17:13:46 crc kubenswrapper[4754]: I0126 17:13:46.465214 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-6w45w" Jan 26 17:13:46 crc kubenswrapper[4754]: I0126 17:13:46.656160 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vwhbp\" (UniqueName: \"kubernetes.io/projected/09f02458-d4e3-49d4-8735-467141a57b6c-kube-api-access-vwhbp\") pod \"09f02458-d4e3-49d4-8735-467141a57b6c\" (UID: \"09f02458-d4e3-49d4-8735-467141a57b6c\") " Jan 26 17:13:46 crc kubenswrapper[4754]: I0126 17:13:46.656548 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/09f02458-d4e3-49d4-8735-467141a57b6c-ssh-key-openstack-edpm-ipam\") pod \"09f02458-d4e3-49d4-8735-467141a57b6c\" (UID: \"09f02458-d4e3-49d4-8735-467141a57b6c\") " Jan 26 17:13:46 crc kubenswrapper[4754]: I0126 17:13:46.656765 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/09f02458-d4e3-49d4-8735-467141a57b6c-inventory\") pod \"09f02458-d4e3-49d4-8735-467141a57b6c\" (UID: \"09f02458-d4e3-49d4-8735-467141a57b6c\") " Jan 26 17:13:46 crc kubenswrapper[4754]: I0126 17:13:46.661736 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09f02458-d4e3-49d4-8735-467141a57b6c-kube-api-access-vwhbp" (OuterVolumeSpecName: "kube-api-access-vwhbp") pod "09f02458-d4e3-49d4-8735-467141a57b6c" (UID: "09f02458-d4e3-49d4-8735-467141a57b6c"). InnerVolumeSpecName "kube-api-access-vwhbp". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:13:46 crc kubenswrapper[4754]: I0126 17:13:46.682900 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09f02458-d4e3-49d4-8735-467141a57b6c-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "09f02458-d4e3-49d4-8735-467141a57b6c" (UID: "09f02458-d4e3-49d4-8735-467141a57b6c"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:13:46 crc kubenswrapper[4754]: I0126 17:13:46.684398 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09f02458-d4e3-49d4-8735-467141a57b6c-inventory" (OuterVolumeSpecName: "inventory") pod "09f02458-d4e3-49d4-8735-467141a57b6c" (UID: "09f02458-d4e3-49d4-8735-467141a57b6c"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:13:46 crc kubenswrapper[4754]: I0126 17:13:46.758728 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vwhbp\" (UniqueName: \"kubernetes.io/projected/09f02458-d4e3-49d4-8735-467141a57b6c-kube-api-access-vwhbp\") on node \"crc\" DevicePath \"\"" Jan 26 17:13:46 crc kubenswrapper[4754]: I0126 17:13:46.758757 4754 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/09f02458-d4e3-49d4-8735-467141a57b6c-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Jan 26 17:13:46 crc kubenswrapper[4754]: I0126 17:13:46.758769 4754 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/09f02458-d4e3-49d4-8735-467141a57b6c-inventory\") on node \"crc\" DevicePath \"\"" Jan 26 17:13:47 crc kubenswrapper[4754]: I0126 17:13:47.039290 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-6w45w" event={"ID":"09f02458-d4e3-49d4-8735-467141a57b6c","Type":"ContainerDied","Data":"207a2252abc5ec3807b9076db33adaf14ebce75d0ca945c6fd63fbf920d583c8"} Jan 26 17:13:47 crc kubenswrapper[4754]: I0126 17:13:47.039337 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="207a2252abc5ec3807b9076db33adaf14ebce75d0ca945c6fd63fbf920d583c8" Jan 26 17:13:47 crc kubenswrapper[4754]: I0126 17:13:47.039310 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-6w45w" Jan 26 17:13:47 crc kubenswrapper[4754]: I0126 17:13:47.097927 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-nng92"] Jan 26 17:13:47 crc kubenswrapper[4754]: E0126 17:13:47.098428 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09f02458-d4e3-49d4-8735-467141a57b6c" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Jan 26 17:13:47 crc kubenswrapper[4754]: I0126 17:13:47.098455 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="09f02458-d4e3-49d4-8735-467141a57b6c" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Jan 26 17:13:47 crc kubenswrapper[4754]: I0126 17:13:47.098755 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="09f02458-d4e3-49d4-8735-467141a57b6c" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Jan 26 17:13:47 crc kubenswrapper[4754]: I0126 17:13:47.099570 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-nng92" Jan 26 17:13:47 crc kubenswrapper[4754]: I0126 17:13:47.101354 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Jan 26 17:13:47 crc kubenswrapper[4754]: I0126 17:13:47.101579 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Jan 26 17:13:47 crc kubenswrapper[4754]: I0126 17:13:47.101685 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-fz7z8" Jan 26 17:13:47 crc kubenswrapper[4754]: I0126 17:13:47.101746 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Jan 26 17:13:47 crc kubenswrapper[4754]: I0126 17:13:47.107032 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-nng92"] Jan 26 17:13:47 crc kubenswrapper[4754]: I0126 17:13:47.270533 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c5j77\" (UniqueName: \"kubernetes.io/projected/fc4da904-699d-44a4-995d-d5ac9b05695b-kube-api-access-c5j77\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-nng92\" (UID: \"fc4da904-699d-44a4-995d-d5ac9b05695b\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-nng92" Jan 26 17:13:47 crc kubenswrapper[4754]: I0126 17:13:47.270723 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fc4da904-699d-44a4-995d-d5ac9b05695b-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-nng92\" (UID: \"fc4da904-699d-44a4-995d-d5ac9b05695b\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-nng92" Jan 26 17:13:47 crc kubenswrapper[4754]: I0126 17:13:47.270820 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc4da904-699d-44a4-995d-d5ac9b05695b-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-nng92\" (UID: \"fc4da904-699d-44a4-995d-d5ac9b05695b\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-nng92" Jan 26 17:13:47 crc kubenswrapper[4754]: I0126 17:13:47.270875 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/fc4da904-699d-44a4-995d-d5ac9b05695b-ssh-key-openstack-edpm-ipam\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-nng92\" (UID: \"fc4da904-699d-44a4-995d-d5ac9b05695b\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-nng92" Jan 26 17:13:47 crc kubenswrapper[4754]: I0126 17:13:47.372329 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc4da904-699d-44a4-995d-d5ac9b05695b-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-nng92\" (UID: \"fc4da904-699d-44a4-995d-d5ac9b05695b\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-nng92" Jan 26 17:13:47 crc kubenswrapper[4754]: I0126 17:13:47.372407 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/fc4da904-699d-44a4-995d-d5ac9b05695b-ssh-key-openstack-edpm-ipam\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-nng92\" (UID: \"fc4da904-699d-44a4-995d-d5ac9b05695b\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-nng92" Jan 26 17:13:47 crc kubenswrapper[4754]: I0126 17:13:47.372484 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c5j77\" (UniqueName: \"kubernetes.io/projected/fc4da904-699d-44a4-995d-d5ac9b05695b-kube-api-access-c5j77\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-nng92\" (UID: \"fc4da904-699d-44a4-995d-d5ac9b05695b\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-nng92" Jan 26 17:13:47 crc kubenswrapper[4754]: I0126 17:13:47.372562 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fc4da904-699d-44a4-995d-d5ac9b05695b-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-nng92\" (UID: \"fc4da904-699d-44a4-995d-d5ac9b05695b\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-nng92" Jan 26 17:13:47 crc kubenswrapper[4754]: I0126 17:13:47.378115 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fc4da904-699d-44a4-995d-d5ac9b05695b-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-nng92\" (UID: \"fc4da904-699d-44a4-995d-d5ac9b05695b\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-nng92" Jan 26 17:13:47 crc kubenswrapper[4754]: I0126 17:13:47.378115 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/fc4da904-699d-44a4-995d-d5ac9b05695b-ssh-key-openstack-edpm-ipam\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-nng92\" (UID: \"fc4da904-699d-44a4-995d-d5ac9b05695b\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-nng92" Jan 26 17:13:47 crc kubenswrapper[4754]: I0126 17:13:47.387224 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc4da904-699d-44a4-995d-d5ac9b05695b-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-nng92\" (UID: \"fc4da904-699d-44a4-995d-d5ac9b05695b\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-nng92" Jan 26 17:13:47 crc kubenswrapper[4754]: I0126 17:13:47.390281 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c5j77\" (UniqueName: \"kubernetes.io/projected/fc4da904-699d-44a4-995d-d5ac9b05695b-kube-api-access-c5j77\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-nng92\" (UID: \"fc4da904-699d-44a4-995d-d5ac9b05695b\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-nng92" Jan 26 17:13:47 crc kubenswrapper[4754]: I0126 17:13:47.419161 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-nng92" Jan 26 17:13:47 crc kubenswrapper[4754]: I0126 17:13:47.767413 4754 scope.go:117] "RemoveContainer" containerID="19eac0e87afaeef93088a8556aa836389e2ebff93a96b482df571cb7b852f0b7" Jan 26 17:13:47 crc kubenswrapper[4754]: E0126 17:13:47.768157 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x65wv_openshift-machine-config-operator(8c3718a4-f354-4284-92e0-fdfb45a692bd)\"" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" Jan 26 17:13:47 crc kubenswrapper[4754]: I0126 17:13:47.928457 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-nng92"] Jan 26 17:13:48 crc kubenswrapper[4754]: I0126 17:13:48.049885 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-nng92" event={"ID":"fc4da904-699d-44a4-995d-d5ac9b05695b","Type":"ContainerStarted","Data":"66160c7aaedd8a89bf0803297b66b87e65360a06988bcea8ba8c683a47b1700a"} Jan 26 17:13:49 crc kubenswrapper[4754]: I0126 17:13:49.060592 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-nng92" event={"ID":"fc4da904-699d-44a4-995d-d5ac9b05695b","Type":"ContainerStarted","Data":"2d39c7847d06705a6d00d870977faf3d2436a164649c4f38d2d888a0460f6cb1"} Jan 26 17:13:49 crc kubenswrapper[4754]: I0126 17:13:49.090572 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-nng92" podStartSLOduration=1.6890021179999999 podStartE2EDuration="2.090547573s" podCreationTimestamp="2026-01-26 17:13:47 +0000 UTC" firstStartedPulling="2026-01-26 17:13:47.93651954 +0000 UTC m=+1594.460699974" lastFinishedPulling="2026-01-26 17:13:48.338064995 +0000 UTC m=+1594.862245429" observedRunningTime="2026-01-26 17:13:49.076946028 +0000 UTC m=+1595.601126492" watchObservedRunningTime="2026-01-26 17:13:49.090547573 +0000 UTC m=+1595.614728007" Jan 26 17:13:52 crc kubenswrapper[4754]: I0126 17:13:52.255213 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-psxqw" Jan 26 17:13:52 crc kubenswrapper[4754]: I0126 17:13:52.332609 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-psxqw"] Jan 26 17:13:52 crc kubenswrapper[4754]: I0126 17:13:52.392458 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-xc89f"] Jan 26 17:13:52 crc kubenswrapper[4754]: I0126 17:13:52.393059 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-xc89f" podUID="231baab8-113c-494b-b5d3-b169370901bf" containerName="registry-server" containerID="cri-o://4cb9df5ec9b0b2595514b3cc038c99ed9bf3e913fa175e54d7db94646d7f9fa8" gracePeriod=2 Jan 26 17:13:53 crc kubenswrapper[4754]: I0126 17:13:53.100047 4754 generic.go:334] "Generic (PLEG): container finished" podID="231baab8-113c-494b-b5d3-b169370901bf" containerID="4cb9df5ec9b0b2595514b3cc038c99ed9bf3e913fa175e54d7db94646d7f9fa8" exitCode=0 Jan 26 17:13:53 crc kubenswrapper[4754]: I0126 17:13:53.100115 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xc89f" event={"ID":"231baab8-113c-494b-b5d3-b169370901bf","Type":"ContainerDied","Data":"4cb9df5ec9b0b2595514b3cc038c99ed9bf3e913fa175e54d7db94646d7f9fa8"} Jan 26 17:13:55 crc kubenswrapper[4754]: I0126 17:13:55.099090 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xc89f" Jan 26 17:13:55 crc kubenswrapper[4754]: I0126 17:13:55.130525 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xc89f" event={"ID":"231baab8-113c-494b-b5d3-b169370901bf","Type":"ContainerDied","Data":"3ebfa83a731a856ba08337b6e788ae8844d894f48770a85b5a57ec9111aa69d4"} Jan 26 17:13:55 crc kubenswrapper[4754]: I0126 17:13:55.130580 4754 scope.go:117] "RemoveContainer" containerID="4cb9df5ec9b0b2595514b3cc038c99ed9bf3e913fa175e54d7db94646d7f9fa8" Jan 26 17:13:55 crc kubenswrapper[4754]: I0126 17:13:55.130589 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xc89f" Jan 26 17:13:55 crc kubenswrapper[4754]: I0126 17:13:55.155149 4754 scope.go:117] "RemoveContainer" containerID="173fad659b5a84e1d302d02094b68c7801eaf468a6f6dbab00f625bca08b57db" Jan 26 17:13:55 crc kubenswrapper[4754]: I0126 17:13:55.176964 4754 scope.go:117] "RemoveContainer" containerID="6ea4becaf6122b4a10f899d30011e0eac95621997405e503d8feb75922612aa8" Jan 26 17:13:55 crc kubenswrapper[4754]: I0126 17:13:55.221706 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6qskt\" (UniqueName: \"kubernetes.io/projected/231baab8-113c-494b-b5d3-b169370901bf-kube-api-access-6qskt\") pod \"231baab8-113c-494b-b5d3-b169370901bf\" (UID: \"231baab8-113c-494b-b5d3-b169370901bf\") " Jan 26 17:13:55 crc kubenswrapper[4754]: I0126 17:13:55.221844 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/231baab8-113c-494b-b5d3-b169370901bf-utilities\") pod \"231baab8-113c-494b-b5d3-b169370901bf\" (UID: \"231baab8-113c-494b-b5d3-b169370901bf\") " Jan 26 17:13:55 crc kubenswrapper[4754]: I0126 17:13:55.221937 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/231baab8-113c-494b-b5d3-b169370901bf-catalog-content\") pod \"231baab8-113c-494b-b5d3-b169370901bf\" (UID: \"231baab8-113c-494b-b5d3-b169370901bf\") " Jan 26 17:13:55 crc kubenswrapper[4754]: I0126 17:13:55.222565 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/231baab8-113c-494b-b5d3-b169370901bf-utilities" (OuterVolumeSpecName: "utilities") pod "231baab8-113c-494b-b5d3-b169370901bf" (UID: "231baab8-113c-494b-b5d3-b169370901bf"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:13:55 crc kubenswrapper[4754]: I0126 17:13:55.227481 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/231baab8-113c-494b-b5d3-b169370901bf-kube-api-access-6qskt" (OuterVolumeSpecName: "kube-api-access-6qskt") pod "231baab8-113c-494b-b5d3-b169370901bf" (UID: "231baab8-113c-494b-b5d3-b169370901bf"). InnerVolumeSpecName "kube-api-access-6qskt". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:13:55 crc kubenswrapper[4754]: I0126 17:13:55.241240 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/231baab8-113c-494b-b5d3-b169370901bf-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "231baab8-113c-494b-b5d3-b169370901bf" (UID: "231baab8-113c-494b-b5d3-b169370901bf"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:13:55 crc kubenswrapper[4754]: I0126 17:13:55.324182 4754 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/231baab8-113c-494b-b5d3-b169370901bf-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 26 17:13:55 crc kubenswrapper[4754]: I0126 17:13:55.324231 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6qskt\" (UniqueName: \"kubernetes.io/projected/231baab8-113c-494b-b5d3-b169370901bf-kube-api-access-6qskt\") on node \"crc\" DevicePath \"\"" Jan 26 17:13:55 crc kubenswrapper[4754]: I0126 17:13:55.324248 4754 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/231baab8-113c-494b-b5d3-b169370901bf-utilities\") on node \"crc\" DevicePath \"\"" Jan 26 17:13:55 crc kubenswrapper[4754]: I0126 17:13:55.472693 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-xc89f"] Jan 26 17:13:55 crc kubenswrapper[4754]: I0126 17:13:55.480455 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-xc89f"] Jan 26 17:13:55 crc kubenswrapper[4754]: I0126 17:13:55.779143 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="231baab8-113c-494b-b5d3-b169370901bf" path="/var/lib/kubelet/pods/231baab8-113c-494b-b5d3-b169370901bf/volumes" Jan 26 17:14:02 crc kubenswrapper[4754]: I0126 17:14:02.767843 4754 scope.go:117] "RemoveContainer" containerID="19eac0e87afaeef93088a8556aa836389e2ebff93a96b482df571cb7b852f0b7" Jan 26 17:14:02 crc kubenswrapper[4754]: E0126 17:14:02.768717 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x65wv_openshift-machine-config-operator(8c3718a4-f354-4284-92e0-fdfb45a692bd)\"" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" Jan 26 17:14:14 crc kubenswrapper[4754]: I0126 17:14:14.767801 4754 scope.go:117] "RemoveContainer" containerID="19eac0e87afaeef93088a8556aa836389e2ebff93a96b482df571cb7b852f0b7" Jan 26 17:14:14 crc kubenswrapper[4754]: E0126 17:14:14.768700 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x65wv_openshift-machine-config-operator(8c3718a4-f354-4284-92e0-fdfb45a692bd)\"" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" Jan 26 17:14:23 crc kubenswrapper[4754]: I0126 17:14:23.253902 4754 scope.go:117] "RemoveContainer" containerID="e14f46f7caf5b8f1f4c6a0e93b774795116f970e960a9fda41198bd120576a1b" Jan 26 17:14:23 crc kubenswrapper[4754]: I0126 17:14:23.522330 4754 scope.go:117] "RemoveContainer" containerID="a63002959b317ff98d2e99a38ab4b17c6e3ea0a8c5ee9526988780fb2de6d4a2" Jan 26 17:14:23 crc kubenswrapper[4754]: I0126 17:14:23.697242 4754 scope.go:117] "RemoveContainer" containerID="e920e8dd93fe23ed63bbaaac5aa05943a59d0ede5d920fc5ddf3c89066cffba4" Jan 26 17:14:25 crc kubenswrapper[4754]: I0126 17:14:25.768861 4754 scope.go:117] "RemoveContainer" containerID="19eac0e87afaeef93088a8556aa836389e2ebff93a96b482df571cb7b852f0b7" Jan 26 17:14:25 crc kubenswrapper[4754]: E0126 17:14:25.769511 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x65wv_openshift-machine-config-operator(8c3718a4-f354-4284-92e0-fdfb45a692bd)\"" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" Jan 26 17:14:39 crc kubenswrapper[4754]: I0126 17:14:39.767919 4754 scope.go:117] "RemoveContainer" containerID="19eac0e87afaeef93088a8556aa836389e2ebff93a96b482df571cb7b852f0b7" Jan 26 17:14:39 crc kubenswrapper[4754]: E0126 17:14:39.768872 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x65wv_openshift-machine-config-operator(8c3718a4-f354-4284-92e0-fdfb45a692bd)\"" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" Jan 26 17:14:53 crc kubenswrapper[4754]: I0126 17:14:53.774303 4754 scope.go:117] "RemoveContainer" containerID="19eac0e87afaeef93088a8556aa836389e2ebff93a96b482df571cb7b852f0b7" Jan 26 17:14:53 crc kubenswrapper[4754]: E0126 17:14:53.775139 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x65wv_openshift-machine-config-operator(8c3718a4-f354-4284-92e0-fdfb45a692bd)\"" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" Jan 26 17:15:00 crc kubenswrapper[4754]: I0126 17:15:00.148343 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29490795-zg9wx"] Jan 26 17:15:00 crc kubenswrapper[4754]: E0126 17:15:00.149353 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="231baab8-113c-494b-b5d3-b169370901bf" containerName="registry-server" Jan 26 17:15:00 crc kubenswrapper[4754]: I0126 17:15:00.149371 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="231baab8-113c-494b-b5d3-b169370901bf" containerName="registry-server" Jan 26 17:15:00 crc kubenswrapper[4754]: E0126 17:15:00.149398 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="231baab8-113c-494b-b5d3-b169370901bf" containerName="extract-utilities" Jan 26 17:15:00 crc kubenswrapper[4754]: I0126 17:15:00.149407 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="231baab8-113c-494b-b5d3-b169370901bf" containerName="extract-utilities" Jan 26 17:15:00 crc kubenswrapper[4754]: E0126 17:15:00.149419 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="231baab8-113c-494b-b5d3-b169370901bf" containerName="extract-content" Jan 26 17:15:00 crc kubenswrapper[4754]: I0126 17:15:00.149429 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="231baab8-113c-494b-b5d3-b169370901bf" containerName="extract-content" Jan 26 17:15:00 crc kubenswrapper[4754]: I0126 17:15:00.149680 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="231baab8-113c-494b-b5d3-b169370901bf" containerName="registry-server" Jan 26 17:15:00 crc kubenswrapper[4754]: I0126 17:15:00.150480 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29490795-zg9wx" Jan 26 17:15:00 crc kubenswrapper[4754]: I0126 17:15:00.154849 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Jan 26 17:15:00 crc kubenswrapper[4754]: I0126 17:15:00.155480 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Jan 26 17:15:00 crc kubenswrapper[4754]: I0126 17:15:00.159278 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29490795-zg9wx"] Jan 26 17:15:00 crc kubenswrapper[4754]: I0126 17:15:00.273619 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5x7gq\" (UniqueName: \"kubernetes.io/projected/abd6ded6-5cbf-43d4-99d1-5e03586f1ee6-kube-api-access-5x7gq\") pod \"collect-profiles-29490795-zg9wx\" (UID: \"abd6ded6-5cbf-43d4-99d1-5e03586f1ee6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29490795-zg9wx" Jan 26 17:15:00 crc kubenswrapper[4754]: I0126 17:15:00.273696 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/abd6ded6-5cbf-43d4-99d1-5e03586f1ee6-secret-volume\") pod \"collect-profiles-29490795-zg9wx\" (UID: \"abd6ded6-5cbf-43d4-99d1-5e03586f1ee6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29490795-zg9wx" Jan 26 17:15:00 crc kubenswrapper[4754]: I0126 17:15:00.273731 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/abd6ded6-5cbf-43d4-99d1-5e03586f1ee6-config-volume\") pod \"collect-profiles-29490795-zg9wx\" (UID: \"abd6ded6-5cbf-43d4-99d1-5e03586f1ee6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29490795-zg9wx" Jan 26 17:15:00 crc kubenswrapper[4754]: I0126 17:15:00.375245 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5x7gq\" (UniqueName: \"kubernetes.io/projected/abd6ded6-5cbf-43d4-99d1-5e03586f1ee6-kube-api-access-5x7gq\") pod \"collect-profiles-29490795-zg9wx\" (UID: \"abd6ded6-5cbf-43d4-99d1-5e03586f1ee6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29490795-zg9wx" Jan 26 17:15:00 crc kubenswrapper[4754]: I0126 17:15:00.375307 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/abd6ded6-5cbf-43d4-99d1-5e03586f1ee6-secret-volume\") pod \"collect-profiles-29490795-zg9wx\" (UID: \"abd6ded6-5cbf-43d4-99d1-5e03586f1ee6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29490795-zg9wx" Jan 26 17:15:00 crc kubenswrapper[4754]: I0126 17:15:00.375344 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/abd6ded6-5cbf-43d4-99d1-5e03586f1ee6-config-volume\") pod \"collect-profiles-29490795-zg9wx\" (UID: \"abd6ded6-5cbf-43d4-99d1-5e03586f1ee6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29490795-zg9wx" Jan 26 17:15:00 crc kubenswrapper[4754]: I0126 17:15:00.376241 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/abd6ded6-5cbf-43d4-99d1-5e03586f1ee6-config-volume\") pod \"collect-profiles-29490795-zg9wx\" (UID: \"abd6ded6-5cbf-43d4-99d1-5e03586f1ee6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29490795-zg9wx" Jan 26 17:15:00 crc kubenswrapper[4754]: I0126 17:15:00.389524 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/abd6ded6-5cbf-43d4-99d1-5e03586f1ee6-secret-volume\") pod \"collect-profiles-29490795-zg9wx\" (UID: \"abd6ded6-5cbf-43d4-99d1-5e03586f1ee6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29490795-zg9wx" Jan 26 17:15:00 crc kubenswrapper[4754]: I0126 17:15:00.393083 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5x7gq\" (UniqueName: \"kubernetes.io/projected/abd6ded6-5cbf-43d4-99d1-5e03586f1ee6-kube-api-access-5x7gq\") pod \"collect-profiles-29490795-zg9wx\" (UID: \"abd6ded6-5cbf-43d4-99d1-5e03586f1ee6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29490795-zg9wx" Jan 26 17:15:00 crc kubenswrapper[4754]: I0126 17:15:00.496681 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29490795-zg9wx" Jan 26 17:15:00 crc kubenswrapper[4754]: I0126 17:15:00.914166 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29490795-zg9wx"] Jan 26 17:15:00 crc kubenswrapper[4754]: W0126 17:15:00.926738 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podabd6ded6_5cbf_43d4_99d1_5e03586f1ee6.slice/crio-405ad1a7237fe83f6044a7d9c5984980b6895dc247d418b796adb08181113791 WatchSource:0}: Error finding container 405ad1a7237fe83f6044a7d9c5984980b6895dc247d418b796adb08181113791: Status 404 returned error can't find the container with id 405ad1a7237fe83f6044a7d9c5984980b6895dc247d418b796adb08181113791 Jan 26 17:15:00 crc kubenswrapper[4754]: I0126 17:15:00.951682 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29490795-zg9wx" event={"ID":"abd6ded6-5cbf-43d4-99d1-5e03586f1ee6","Type":"ContainerStarted","Data":"405ad1a7237fe83f6044a7d9c5984980b6895dc247d418b796adb08181113791"} Jan 26 17:15:01 crc kubenswrapper[4754]: I0126 17:15:01.960121 4754 generic.go:334] "Generic (PLEG): container finished" podID="abd6ded6-5cbf-43d4-99d1-5e03586f1ee6" containerID="ab2be4d7b3c6674c982dfb9a730a8d5f3b38c70d808dcf1f5a3a8ebaf1dd7013" exitCode=0 Jan 26 17:15:01 crc kubenswrapper[4754]: I0126 17:15:01.960211 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29490795-zg9wx" event={"ID":"abd6ded6-5cbf-43d4-99d1-5e03586f1ee6","Type":"ContainerDied","Data":"ab2be4d7b3c6674c982dfb9a730a8d5f3b38c70d808dcf1f5a3a8ebaf1dd7013"} Jan 26 17:15:03 crc kubenswrapper[4754]: I0126 17:15:03.287451 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29490795-zg9wx" Jan 26 17:15:03 crc kubenswrapper[4754]: I0126 17:15:03.428220 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/abd6ded6-5cbf-43d4-99d1-5e03586f1ee6-config-volume\") pod \"abd6ded6-5cbf-43d4-99d1-5e03586f1ee6\" (UID: \"abd6ded6-5cbf-43d4-99d1-5e03586f1ee6\") " Jan 26 17:15:03 crc kubenswrapper[4754]: I0126 17:15:03.428381 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5x7gq\" (UniqueName: \"kubernetes.io/projected/abd6ded6-5cbf-43d4-99d1-5e03586f1ee6-kube-api-access-5x7gq\") pod \"abd6ded6-5cbf-43d4-99d1-5e03586f1ee6\" (UID: \"abd6ded6-5cbf-43d4-99d1-5e03586f1ee6\") " Jan 26 17:15:03 crc kubenswrapper[4754]: I0126 17:15:03.428449 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/abd6ded6-5cbf-43d4-99d1-5e03586f1ee6-secret-volume\") pod \"abd6ded6-5cbf-43d4-99d1-5e03586f1ee6\" (UID: \"abd6ded6-5cbf-43d4-99d1-5e03586f1ee6\") " Jan 26 17:15:03 crc kubenswrapper[4754]: I0126 17:15:03.429192 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/abd6ded6-5cbf-43d4-99d1-5e03586f1ee6-config-volume" (OuterVolumeSpecName: "config-volume") pod "abd6ded6-5cbf-43d4-99d1-5e03586f1ee6" (UID: "abd6ded6-5cbf-43d4-99d1-5e03586f1ee6"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:15:03 crc kubenswrapper[4754]: I0126 17:15:03.429418 4754 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/abd6ded6-5cbf-43d4-99d1-5e03586f1ee6-config-volume\") on node \"crc\" DevicePath \"\"" Jan 26 17:15:03 crc kubenswrapper[4754]: I0126 17:15:03.435454 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/abd6ded6-5cbf-43d4-99d1-5e03586f1ee6-kube-api-access-5x7gq" (OuterVolumeSpecName: "kube-api-access-5x7gq") pod "abd6ded6-5cbf-43d4-99d1-5e03586f1ee6" (UID: "abd6ded6-5cbf-43d4-99d1-5e03586f1ee6"). InnerVolumeSpecName "kube-api-access-5x7gq". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:15:03 crc kubenswrapper[4754]: I0126 17:15:03.435483 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/abd6ded6-5cbf-43d4-99d1-5e03586f1ee6-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "abd6ded6-5cbf-43d4-99d1-5e03586f1ee6" (UID: "abd6ded6-5cbf-43d4-99d1-5e03586f1ee6"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:15:03 crc kubenswrapper[4754]: I0126 17:15:03.531747 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5x7gq\" (UniqueName: \"kubernetes.io/projected/abd6ded6-5cbf-43d4-99d1-5e03586f1ee6-kube-api-access-5x7gq\") on node \"crc\" DevicePath \"\"" Jan 26 17:15:03 crc kubenswrapper[4754]: I0126 17:15:03.531796 4754 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/abd6ded6-5cbf-43d4-99d1-5e03586f1ee6-secret-volume\") on node \"crc\" DevicePath \"\"" Jan 26 17:15:03 crc kubenswrapper[4754]: I0126 17:15:03.992301 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29490795-zg9wx" event={"ID":"abd6ded6-5cbf-43d4-99d1-5e03586f1ee6","Type":"ContainerDied","Data":"405ad1a7237fe83f6044a7d9c5984980b6895dc247d418b796adb08181113791"} Jan 26 17:15:03 crc kubenswrapper[4754]: I0126 17:15:03.992585 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="405ad1a7237fe83f6044a7d9c5984980b6895dc247d418b796adb08181113791" Jan 26 17:15:03 crc kubenswrapper[4754]: I0126 17:15:03.992404 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29490795-zg9wx" Jan 26 17:15:06 crc kubenswrapper[4754]: I0126 17:15:06.767928 4754 scope.go:117] "RemoveContainer" containerID="19eac0e87afaeef93088a8556aa836389e2ebff93a96b482df571cb7b852f0b7" Jan 26 17:15:06 crc kubenswrapper[4754]: E0126 17:15:06.768425 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x65wv_openshift-machine-config-operator(8c3718a4-f354-4284-92e0-fdfb45a692bd)\"" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" Jan 26 17:15:19 crc kubenswrapper[4754]: I0126 17:15:19.767976 4754 scope.go:117] "RemoveContainer" containerID="19eac0e87afaeef93088a8556aa836389e2ebff93a96b482df571cb7b852f0b7" Jan 26 17:15:19 crc kubenswrapper[4754]: E0126 17:15:19.768542 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x65wv_openshift-machine-config-operator(8c3718a4-f354-4284-92e0-fdfb45a692bd)\"" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" Jan 26 17:15:30 crc kubenswrapper[4754]: I0126 17:15:30.767227 4754 scope.go:117] "RemoveContainer" containerID="19eac0e87afaeef93088a8556aa836389e2ebff93a96b482df571cb7b852f0b7" Jan 26 17:15:30 crc kubenswrapper[4754]: E0126 17:15:30.768066 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x65wv_openshift-machine-config-operator(8c3718a4-f354-4284-92e0-fdfb45a692bd)\"" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" Jan 26 17:15:43 crc kubenswrapper[4754]: I0126 17:15:43.774913 4754 scope.go:117] "RemoveContainer" containerID="19eac0e87afaeef93088a8556aa836389e2ebff93a96b482df571cb7b852f0b7" Jan 26 17:15:43 crc kubenswrapper[4754]: E0126 17:15:43.776442 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x65wv_openshift-machine-config-operator(8c3718a4-f354-4284-92e0-fdfb45a692bd)\"" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" Jan 26 17:15:54 crc kubenswrapper[4754]: I0126 17:15:54.767563 4754 scope.go:117] "RemoveContainer" containerID="19eac0e87afaeef93088a8556aa836389e2ebff93a96b482df571cb7b852f0b7" Jan 26 17:15:54 crc kubenswrapper[4754]: E0126 17:15:54.768280 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x65wv_openshift-machine-config-operator(8c3718a4-f354-4284-92e0-fdfb45a692bd)\"" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" Jan 26 17:16:07 crc kubenswrapper[4754]: I0126 17:16:07.768320 4754 scope.go:117] "RemoveContainer" containerID="19eac0e87afaeef93088a8556aa836389e2ebff93a96b482df571cb7b852f0b7" Jan 26 17:16:07 crc kubenswrapper[4754]: E0126 17:16:07.769583 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x65wv_openshift-machine-config-operator(8c3718a4-f354-4284-92e0-fdfb45a692bd)\"" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" Jan 26 17:16:19 crc kubenswrapper[4754]: I0126 17:16:19.768327 4754 scope.go:117] "RemoveContainer" containerID="19eac0e87afaeef93088a8556aa836389e2ebff93a96b482df571cb7b852f0b7" Jan 26 17:16:19 crc kubenswrapper[4754]: E0126 17:16:19.769189 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x65wv_openshift-machine-config-operator(8c3718a4-f354-4284-92e0-fdfb45a692bd)\"" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" Jan 26 17:16:26 crc kubenswrapper[4754]: I0126 17:16:26.049140 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-f12b-account-create-update-scv4g"] Jan 26 17:16:26 crc kubenswrapper[4754]: I0126 17:16:26.062586 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-c7cxf"] Jan 26 17:16:26 crc kubenswrapper[4754]: I0126 17:16:26.071104 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-4c00-account-create-update-h9g8d"] Jan 26 17:16:26 crc kubenswrapper[4754]: I0126 17:16:26.078530 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-8dnlq"] Jan 26 17:16:26 crc kubenswrapper[4754]: I0126 17:16:26.087489 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-4c00-account-create-update-h9g8d"] Jan 26 17:16:26 crc kubenswrapper[4754]: I0126 17:16:26.097105 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-c7cxf"] Jan 26 17:16:26 crc kubenswrapper[4754]: I0126 17:16:26.105292 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-f12b-account-create-update-scv4g"] Jan 26 17:16:26 crc kubenswrapper[4754]: I0126 17:16:26.115003 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-8dnlq"] Jan 26 17:16:27 crc kubenswrapper[4754]: I0126 17:16:27.780325 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9295f950-6f88-42d3-a82e-688cda9cce76" path="/var/lib/kubelet/pods/9295f950-6f88-42d3-a82e-688cda9cce76/volumes" Jan 26 17:16:27 crc kubenswrapper[4754]: I0126 17:16:27.781595 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d1acc0d8-ac6a-4b56-bc9b-d926176314c3" path="/var/lib/kubelet/pods/d1acc0d8-ac6a-4b56-bc9b-d926176314c3/volumes" Jan 26 17:16:27 crc kubenswrapper[4754]: I0126 17:16:27.782337 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dce62524-1013-4b38-b563-d9f6d7701523" path="/var/lib/kubelet/pods/dce62524-1013-4b38-b563-d9f6d7701523/volumes" Jan 26 17:16:27 crc kubenswrapper[4754]: I0126 17:16:27.783065 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fe272255-848d-4761-a7f9-347103832e7d" path="/var/lib/kubelet/pods/fe272255-848d-4761-a7f9-347103832e7d/volumes" Jan 26 17:16:30 crc kubenswrapper[4754]: I0126 17:16:30.056346 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-fc80-account-create-update-jh5xv"] Jan 26 17:16:30 crc kubenswrapper[4754]: I0126 17:16:30.075269 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-plgbz"] Jan 26 17:16:30 crc kubenswrapper[4754]: I0126 17:16:30.095985 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-plgbz"] Jan 26 17:16:30 crc kubenswrapper[4754]: I0126 17:16:30.114257 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-fc80-account-create-update-jh5xv"] Jan 26 17:16:31 crc kubenswrapper[4754]: I0126 17:16:31.778631 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="07ac2529-0a5d-4a79-95cf-2b5040e67f46" path="/var/lib/kubelet/pods/07ac2529-0a5d-4a79-95cf-2b5040e67f46/volumes" Jan 26 17:16:31 crc kubenswrapper[4754]: I0126 17:16:31.779802 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e2c6020a-c674-43b0-9698-b6b31e0ab5b8" path="/var/lib/kubelet/pods/e2c6020a-c674-43b0-9698-b6b31e0ab5b8/volumes" Jan 26 17:16:33 crc kubenswrapper[4754]: I0126 17:16:33.777768 4754 scope.go:117] "RemoveContainer" containerID="19eac0e87afaeef93088a8556aa836389e2ebff93a96b482df571cb7b852f0b7" Jan 26 17:16:33 crc kubenswrapper[4754]: E0126 17:16:33.779480 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x65wv_openshift-machine-config-operator(8c3718a4-f354-4284-92e0-fdfb45a692bd)\"" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" Jan 26 17:16:44 crc kubenswrapper[4754]: I0126 17:16:44.767789 4754 scope.go:117] "RemoveContainer" containerID="19eac0e87afaeef93088a8556aa836389e2ebff93a96b482df571cb7b852f0b7" Jan 26 17:16:44 crc kubenswrapper[4754]: E0126 17:16:44.768599 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x65wv_openshift-machine-config-operator(8c3718a4-f354-4284-92e0-fdfb45a692bd)\"" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" Jan 26 17:16:56 crc kubenswrapper[4754]: I0126 17:16:56.768839 4754 scope.go:117] "RemoveContainer" containerID="19eac0e87afaeef93088a8556aa836389e2ebff93a96b482df571cb7b852f0b7" Jan 26 17:16:56 crc kubenswrapper[4754]: E0126 17:16:56.769776 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x65wv_openshift-machine-config-operator(8c3718a4-f354-4284-92e0-fdfb45a692bd)\"" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" Jan 26 17:17:09 crc kubenswrapper[4754]: I0126 17:17:09.767523 4754 scope.go:117] "RemoveContainer" containerID="19eac0e87afaeef93088a8556aa836389e2ebff93a96b482df571cb7b852f0b7" Jan 26 17:17:09 crc kubenswrapper[4754]: E0126 17:17:09.768374 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x65wv_openshift-machine-config-operator(8c3718a4-f354-4284-92e0-fdfb45a692bd)\"" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" Jan 26 17:17:14 crc kubenswrapper[4754]: I0126 17:17:14.048207 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-899gb"] Jan 26 17:17:14 crc kubenswrapper[4754]: I0126 17:17:14.067771 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-899gb"] Jan 26 17:17:15 crc kubenswrapper[4754]: I0126 17:17:15.043042 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-92jjf"] Jan 26 17:17:15 crc kubenswrapper[4754]: I0126 17:17:15.058194 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-b037-account-create-update-sb2vx"] Jan 26 17:17:15 crc kubenswrapper[4754]: I0126 17:17:15.068496 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-dcjt9"] Jan 26 17:17:15 crc kubenswrapper[4754]: I0126 17:17:15.077311 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-92jjf"] Jan 26 17:17:15 crc kubenswrapper[4754]: I0126 17:17:15.084771 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-b037-account-create-update-sb2vx"] Jan 26 17:17:15 crc kubenswrapper[4754]: I0126 17:17:15.093637 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-dcjt9"] Jan 26 17:17:15 crc kubenswrapper[4754]: I0126 17:17:15.778526 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0bc80e48-26f1-47c6-9fc2-95057bfd06b9" path="/var/lib/kubelet/pods/0bc80e48-26f1-47c6-9fc2-95057bfd06b9/volumes" Jan 26 17:17:15 crc kubenswrapper[4754]: I0126 17:17:15.779205 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22b9c223-a293-4248-a32b-36498fa1a43a" path="/var/lib/kubelet/pods/22b9c223-a293-4248-a32b-36498fa1a43a/volumes" Jan 26 17:17:15 crc kubenswrapper[4754]: I0126 17:17:15.779773 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="83bdc38e-6bbc-4ea8-b57c-9942707b4071" path="/var/lib/kubelet/pods/83bdc38e-6bbc-4ea8-b57c-9942707b4071/volumes" Jan 26 17:17:15 crc kubenswrapper[4754]: I0126 17:17:15.780290 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d4102ba6-b5f4-40a7-8551-300924ddba03" path="/var/lib/kubelet/pods/d4102ba6-b5f4-40a7-8551-300924ddba03/volumes" Jan 26 17:17:16 crc kubenswrapper[4754]: I0126 17:17:16.036334 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-c8dc-account-create-update-w9qwp"] Jan 26 17:17:16 crc kubenswrapper[4754]: I0126 17:17:16.047993 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-c8dc-account-create-update-w9qwp"] Jan 26 17:17:16 crc kubenswrapper[4754]: I0126 17:17:16.061175 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-c7d7-account-create-update-qjftw"] Jan 26 17:17:16 crc kubenswrapper[4754]: I0126 17:17:16.073280 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-c7d7-account-create-update-qjftw"] Jan 26 17:17:17 crc kubenswrapper[4754]: I0126 17:17:17.781638 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2cb8378d-e281-4399-a82b-4fe351fcac16" path="/var/lib/kubelet/pods/2cb8378d-e281-4399-a82b-4fe351fcac16/volumes" Jan 26 17:17:17 crc kubenswrapper[4754]: I0126 17:17:17.782334 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fc67dbf2-5b6b-44cb-a5d4-8b7e32fd4a87" path="/var/lib/kubelet/pods/fc67dbf2-5b6b-44cb-a5d4-8b7e32fd4a87/volumes" Jan 26 17:17:19 crc kubenswrapper[4754]: I0126 17:17:19.035614 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-qbqp5"] Jan 26 17:17:19 crc kubenswrapper[4754]: I0126 17:17:19.044529 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-qbqp5"] Jan 26 17:17:19 crc kubenswrapper[4754]: I0126 17:17:19.779405 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bae8ecc7-8fac-4a63-9e99-4e7543eae5af" path="/var/lib/kubelet/pods/bae8ecc7-8fac-4a63-9e99-4e7543eae5af/volumes" Jan 26 17:17:20 crc kubenswrapper[4754]: I0126 17:17:20.767927 4754 scope.go:117] "RemoveContainer" containerID="19eac0e87afaeef93088a8556aa836389e2ebff93a96b482df571cb7b852f0b7" Jan 26 17:17:20 crc kubenswrapper[4754]: E0126 17:17:20.768498 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x65wv_openshift-machine-config-operator(8c3718a4-f354-4284-92e0-fdfb45a692bd)\"" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" Jan 26 17:17:23 crc kubenswrapper[4754]: I0126 17:17:23.827521 4754 scope.go:117] "RemoveContainer" containerID="fce952ed99ebf0ec3ce4c1388209744f2a30e0a41cc72ee2c1b854e6dd140583" Jan 26 17:17:25 crc kubenswrapper[4754]: I0126 17:17:25.098004 4754 scope.go:117] "RemoveContainer" containerID="f9f718438d30237bbc332cf9180c7bba7606646662f6c6d7426532bcfad633bf" Jan 26 17:17:25 crc kubenswrapper[4754]: I0126 17:17:25.364900 4754 scope.go:117] "RemoveContainer" containerID="b4c825db5f3f0d1ad8483b5b742fdd62df9527e36ac7253d696c7b31652165e7" Jan 26 17:17:25 crc kubenswrapper[4754]: I0126 17:17:25.390526 4754 scope.go:117] "RemoveContainer" containerID="189d6fed262dc3a4f55ece04db494b95855ed4513e2524c3e6835c1f6bca47b2" Jan 26 17:17:25 crc kubenswrapper[4754]: I0126 17:17:25.434144 4754 scope.go:117] "RemoveContainer" containerID="c63f8c2db5c0a10f443a10a4a6dca92401a4754f6725760e22c8417fc8de935b" Jan 26 17:17:25 crc kubenswrapper[4754]: I0126 17:17:25.495125 4754 scope.go:117] "RemoveContainer" containerID="2b1a9257ec9c280eab069d563f3604849bb5a428f499d0d234618db276339bc8" Jan 26 17:17:25 crc kubenswrapper[4754]: I0126 17:17:25.525745 4754 scope.go:117] "RemoveContainer" containerID="e84f43eb249c92fd72b6e9bc45e2e579ad9889dd0ac77ec6d06f9ec1235954ec" Jan 26 17:17:25 crc kubenswrapper[4754]: I0126 17:17:25.575806 4754 scope.go:117] "RemoveContainer" containerID="e330185b9b6205d6f4d57fde1a2bb14997226cc270954bfc37385ba8ca09676b" Jan 26 17:17:25 crc kubenswrapper[4754]: I0126 17:17:25.632974 4754 scope.go:117] "RemoveContainer" containerID="528723486d42beba27f3015599fc78a7ff5226695e282daaca10290a51dc0fbe" Jan 26 17:17:25 crc kubenswrapper[4754]: I0126 17:17:25.669924 4754 scope.go:117] "RemoveContainer" containerID="e6362fc04381ff9a7db06b0751f7ace044043ec05a105674da46640b67248fc0" Jan 26 17:17:25 crc kubenswrapper[4754]: I0126 17:17:25.735985 4754 scope.go:117] "RemoveContainer" containerID="f05e7698a2c643b6ab539c22a2c1983a4bfc697ed974bcb9c7ba86d8a596f3e9" Jan 26 17:17:25 crc kubenswrapper[4754]: I0126 17:17:25.760433 4754 scope.go:117] "RemoveContainer" containerID="88006dee24ddc93951251d2d1931aa8bd96e1a1ba2aef15158a5bd9c64ebd982" Jan 26 17:17:25 crc kubenswrapper[4754]: I0126 17:17:25.790601 4754 scope.go:117] "RemoveContainer" containerID="354866a3d7b6dedd7082b7847d6eb060b0999db466f2424c5454aebfe3938d6a" Jan 26 17:17:25 crc kubenswrapper[4754]: I0126 17:17:25.827662 4754 scope.go:117] "RemoveContainer" containerID="7b2c6f588352f73fbb3612f888d11cd5358bb977bdf45587a044dab856652069" Jan 26 17:17:27 crc kubenswrapper[4754]: I0126 17:17:27.034419 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-68nzr"] Jan 26 17:17:27 crc kubenswrapper[4754]: I0126 17:17:27.043386 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-68nzr"] Jan 26 17:17:27 crc kubenswrapper[4754]: I0126 17:17:27.778423 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="35a2e341-10fc-48ca-8c99-65f1664cbba5" path="/var/lib/kubelet/pods/35a2e341-10fc-48ca-8c99-65f1664cbba5/volumes" Jan 26 17:17:33 crc kubenswrapper[4754]: I0126 17:17:33.509240 4754 generic.go:334] "Generic (PLEG): container finished" podID="fc4da904-699d-44a4-995d-d5ac9b05695b" containerID="2d39c7847d06705a6d00d870977faf3d2436a164649c4f38d2d888a0460f6cb1" exitCode=0 Jan 26 17:17:33 crc kubenswrapper[4754]: I0126 17:17:33.509334 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-nng92" event={"ID":"fc4da904-699d-44a4-995d-d5ac9b05695b","Type":"ContainerDied","Data":"2d39c7847d06705a6d00d870977faf3d2436a164649c4f38d2d888a0460f6cb1"} Jan 26 17:17:33 crc kubenswrapper[4754]: I0126 17:17:33.774209 4754 scope.go:117] "RemoveContainer" containerID="19eac0e87afaeef93088a8556aa836389e2ebff93a96b482df571cb7b852f0b7" Jan 26 17:17:33 crc kubenswrapper[4754]: E0126 17:17:33.775145 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x65wv_openshift-machine-config-operator(8c3718a4-f354-4284-92e0-fdfb45a692bd)\"" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" Jan 26 17:17:35 crc kubenswrapper[4754]: I0126 17:17:35.001869 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-nng92" Jan 26 17:17:35 crc kubenswrapper[4754]: I0126 17:17:35.051867 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/root-account-create-update-hznmv"] Jan 26 17:17:35 crc kubenswrapper[4754]: I0126 17:17:35.059181 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/root-account-create-update-hznmv"] Jan 26 17:17:35 crc kubenswrapper[4754]: I0126 17:17:35.128173 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/fc4da904-699d-44a4-995d-d5ac9b05695b-ssh-key-openstack-edpm-ipam\") pod \"fc4da904-699d-44a4-995d-d5ac9b05695b\" (UID: \"fc4da904-699d-44a4-995d-d5ac9b05695b\") " Jan 26 17:17:35 crc kubenswrapper[4754]: I0126 17:17:35.128298 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc4da904-699d-44a4-995d-d5ac9b05695b-bootstrap-combined-ca-bundle\") pod \"fc4da904-699d-44a4-995d-d5ac9b05695b\" (UID: \"fc4da904-699d-44a4-995d-d5ac9b05695b\") " Jan 26 17:17:35 crc kubenswrapper[4754]: I0126 17:17:35.128390 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c5j77\" (UniqueName: \"kubernetes.io/projected/fc4da904-699d-44a4-995d-d5ac9b05695b-kube-api-access-c5j77\") pod \"fc4da904-699d-44a4-995d-d5ac9b05695b\" (UID: \"fc4da904-699d-44a4-995d-d5ac9b05695b\") " Jan 26 17:17:35 crc kubenswrapper[4754]: I0126 17:17:35.128462 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fc4da904-699d-44a4-995d-d5ac9b05695b-inventory\") pod \"fc4da904-699d-44a4-995d-d5ac9b05695b\" (UID: \"fc4da904-699d-44a4-995d-d5ac9b05695b\") " Jan 26 17:17:35 crc kubenswrapper[4754]: I0126 17:17:35.133366 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fc4da904-699d-44a4-995d-d5ac9b05695b-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "fc4da904-699d-44a4-995d-d5ac9b05695b" (UID: "fc4da904-699d-44a4-995d-d5ac9b05695b"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:17:35 crc kubenswrapper[4754]: I0126 17:17:35.133471 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fc4da904-699d-44a4-995d-d5ac9b05695b-kube-api-access-c5j77" (OuterVolumeSpecName: "kube-api-access-c5j77") pod "fc4da904-699d-44a4-995d-d5ac9b05695b" (UID: "fc4da904-699d-44a4-995d-d5ac9b05695b"). InnerVolumeSpecName "kube-api-access-c5j77". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:17:35 crc kubenswrapper[4754]: I0126 17:17:35.153726 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fc4da904-699d-44a4-995d-d5ac9b05695b-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "fc4da904-699d-44a4-995d-d5ac9b05695b" (UID: "fc4da904-699d-44a4-995d-d5ac9b05695b"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:17:35 crc kubenswrapper[4754]: I0126 17:17:35.154867 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fc4da904-699d-44a4-995d-d5ac9b05695b-inventory" (OuterVolumeSpecName: "inventory") pod "fc4da904-699d-44a4-995d-d5ac9b05695b" (UID: "fc4da904-699d-44a4-995d-d5ac9b05695b"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:17:35 crc kubenswrapper[4754]: I0126 17:17:35.231321 4754 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fc4da904-699d-44a4-995d-d5ac9b05695b-inventory\") on node \"crc\" DevicePath \"\"" Jan 26 17:17:35 crc kubenswrapper[4754]: I0126 17:17:35.231365 4754 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/fc4da904-699d-44a4-995d-d5ac9b05695b-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Jan 26 17:17:35 crc kubenswrapper[4754]: I0126 17:17:35.231379 4754 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc4da904-699d-44a4-995d-d5ac9b05695b-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 26 17:17:35 crc kubenswrapper[4754]: I0126 17:17:35.231392 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c5j77\" (UniqueName: \"kubernetes.io/projected/fc4da904-699d-44a4-995d-d5ac9b05695b-kube-api-access-c5j77\") on node \"crc\" DevicePath \"\"" Jan 26 17:17:35 crc kubenswrapper[4754]: I0126 17:17:35.528356 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-nng92" event={"ID":"fc4da904-699d-44a4-995d-d5ac9b05695b","Type":"ContainerDied","Data":"66160c7aaedd8a89bf0803297b66b87e65360a06988bcea8ba8c683a47b1700a"} Jan 26 17:17:35 crc kubenswrapper[4754]: I0126 17:17:35.528417 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="66160c7aaedd8a89bf0803297b66b87e65360a06988bcea8ba8c683a47b1700a" Jan 26 17:17:35 crc kubenswrapper[4754]: I0126 17:17:35.528687 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-nng92" Jan 26 17:17:35 crc kubenswrapper[4754]: I0126 17:17:35.629418 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-rrhnq"] Jan 26 17:17:35 crc kubenswrapper[4754]: E0126 17:17:35.629952 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc4da904-699d-44a4-995d-d5ac9b05695b" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Jan 26 17:17:35 crc kubenswrapper[4754]: I0126 17:17:35.629976 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc4da904-699d-44a4-995d-d5ac9b05695b" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Jan 26 17:17:35 crc kubenswrapper[4754]: E0126 17:17:35.630009 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="abd6ded6-5cbf-43d4-99d1-5e03586f1ee6" containerName="collect-profiles" Jan 26 17:17:35 crc kubenswrapper[4754]: I0126 17:17:35.630017 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="abd6ded6-5cbf-43d4-99d1-5e03586f1ee6" containerName="collect-profiles" Jan 26 17:17:35 crc kubenswrapper[4754]: I0126 17:17:35.630242 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="fc4da904-699d-44a4-995d-d5ac9b05695b" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Jan 26 17:17:35 crc kubenswrapper[4754]: I0126 17:17:35.630263 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="abd6ded6-5cbf-43d4-99d1-5e03586f1ee6" containerName="collect-profiles" Jan 26 17:17:35 crc kubenswrapper[4754]: I0126 17:17:35.630983 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-rrhnq" Jan 26 17:17:35 crc kubenswrapper[4754]: I0126 17:17:35.633141 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Jan 26 17:17:35 crc kubenswrapper[4754]: I0126 17:17:35.633431 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-fz7z8" Jan 26 17:17:35 crc kubenswrapper[4754]: I0126 17:17:35.636169 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Jan 26 17:17:35 crc kubenswrapper[4754]: I0126 17:17:35.637039 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Jan 26 17:17:35 crc kubenswrapper[4754]: I0126 17:17:35.639319 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-rrhnq"] Jan 26 17:17:35 crc kubenswrapper[4754]: I0126 17:17:35.739432 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/55b57fed-619a-44b7-af60-f6d7a43943f1-ssh-key-openstack-edpm-ipam\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-rrhnq\" (UID: \"55b57fed-619a-44b7-af60-f6d7a43943f1\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-rrhnq" Jan 26 17:17:35 crc kubenswrapper[4754]: I0126 17:17:35.739638 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fp5jn\" (UniqueName: \"kubernetes.io/projected/55b57fed-619a-44b7-af60-f6d7a43943f1-kube-api-access-fp5jn\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-rrhnq\" (UID: \"55b57fed-619a-44b7-af60-f6d7a43943f1\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-rrhnq" Jan 26 17:17:35 crc kubenswrapper[4754]: I0126 17:17:35.740111 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/55b57fed-619a-44b7-af60-f6d7a43943f1-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-rrhnq\" (UID: \"55b57fed-619a-44b7-af60-f6d7a43943f1\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-rrhnq" Jan 26 17:17:35 crc kubenswrapper[4754]: I0126 17:17:35.778449 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4d4b0b97-349d-45e4-884b-04c0dae58ccb" path="/var/lib/kubelet/pods/4d4b0b97-349d-45e4-884b-04c0dae58ccb/volumes" Jan 26 17:17:35 crc kubenswrapper[4754]: I0126 17:17:35.841873 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fp5jn\" (UniqueName: \"kubernetes.io/projected/55b57fed-619a-44b7-af60-f6d7a43943f1-kube-api-access-fp5jn\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-rrhnq\" (UID: \"55b57fed-619a-44b7-af60-f6d7a43943f1\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-rrhnq" Jan 26 17:17:35 crc kubenswrapper[4754]: I0126 17:17:35.842013 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/55b57fed-619a-44b7-af60-f6d7a43943f1-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-rrhnq\" (UID: \"55b57fed-619a-44b7-af60-f6d7a43943f1\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-rrhnq" Jan 26 17:17:35 crc kubenswrapper[4754]: I0126 17:17:35.842095 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/55b57fed-619a-44b7-af60-f6d7a43943f1-ssh-key-openstack-edpm-ipam\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-rrhnq\" (UID: \"55b57fed-619a-44b7-af60-f6d7a43943f1\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-rrhnq" Jan 26 17:17:35 crc kubenswrapper[4754]: I0126 17:17:35.846113 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/55b57fed-619a-44b7-af60-f6d7a43943f1-ssh-key-openstack-edpm-ipam\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-rrhnq\" (UID: \"55b57fed-619a-44b7-af60-f6d7a43943f1\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-rrhnq" Jan 26 17:17:35 crc kubenswrapper[4754]: I0126 17:17:35.855629 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/55b57fed-619a-44b7-af60-f6d7a43943f1-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-rrhnq\" (UID: \"55b57fed-619a-44b7-af60-f6d7a43943f1\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-rrhnq" Jan 26 17:17:35 crc kubenswrapper[4754]: I0126 17:17:35.863124 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fp5jn\" (UniqueName: \"kubernetes.io/projected/55b57fed-619a-44b7-af60-f6d7a43943f1-kube-api-access-fp5jn\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-rrhnq\" (UID: \"55b57fed-619a-44b7-af60-f6d7a43943f1\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-rrhnq" Jan 26 17:17:35 crc kubenswrapper[4754]: I0126 17:17:35.954377 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-rrhnq" Jan 26 17:17:36 crc kubenswrapper[4754]: I0126 17:17:36.498862 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-rrhnq"] Jan 26 17:17:36 crc kubenswrapper[4754]: I0126 17:17:36.501514 4754 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 26 17:17:36 crc kubenswrapper[4754]: I0126 17:17:36.538897 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-rrhnq" event={"ID":"55b57fed-619a-44b7-af60-f6d7a43943f1","Type":"ContainerStarted","Data":"4f88e24a27b0baf222d20554ed3ddd0d702d536ac920f7c3d425e7426f3b1f51"} Jan 26 17:17:37 crc kubenswrapper[4754]: I0126 17:17:37.549546 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-rrhnq" event={"ID":"55b57fed-619a-44b7-af60-f6d7a43943f1","Type":"ContainerStarted","Data":"c7081a1de193a8c59596a064551979f778ea21453c14cdfa6f8086ba8dcbd7de"} Jan 26 17:17:37 crc kubenswrapper[4754]: I0126 17:17:37.581475 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-rrhnq" podStartSLOduration=2.023895741 podStartE2EDuration="2.581455481s" podCreationTimestamp="2026-01-26 17:17:35 +0000 UTC" firstStartedPulling="2026-01-26 17:17:36.501282906 +0000 UTC m=+1823.025463340" lastFinishedPulling="2026-01-26 17:17:37.058842646 +0000 UTC m=+1823.583023080" observedRunningTime="2026-01-26 17:17:37.574305061 +0000 UTC m=+1824.098485515" watchObservedRunningTime="2026-01-26 17:17:37.581455481 +0000 UTC m=+1824.105635915" Jan 26 17:17:47 crc kubenswrapper[4754]: I0126 17:17:47.767244 4754 scope.go:117] "RemoveContainer" containerID="19eac0e87afaeef93088a8556aa836389e2ebff93a96b482df571cb7b852f0b7" Jan 26 17:17:47 crc kubenswrapper[4754]: E0126 17:17:47.768052 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x65wv_openshift-machine-config-operator(8c3718a4-f354-4284-92e0-fdfb45a692bd)\"" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" Jan 26 17:18:01 crc kubenswrapper[4754]: I0126 17:18:01.767390 4754 scope.go:117] "RemoveContainer" containerID="19eac0e87afaeef93088a8556aa836389e2ebff93a96b482df571cb7b852f0b7" Jan 26 17:18:01 crc kubenswrapper[4754]: E0126 17:18:01.768183 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x65wv_openshift-machine-config-operator(8c3718a4-f354-4284-92e0-fdfb45a692bd)\"" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" Jan 26 17:18:14 crc kubenswrapper[4754]: I0126 17:18:14.767163 4754 scope.go:117] "RemoveContainer" containerID="19eac0e87afaeef93088a8556aa836389e2ebff93a96b482df571cb7b852f0b7" Jan 26 17:18:15 crc kubenswrapper[4754]: I0126 17:18:15.913533 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" event={"ID":"8c3718a4-f354-4284-92e0-fdfb45a692bd","Type":"ContainerStarted","Data":"1f2fc5d329d9c26d0c14a843ed658e96c72f72ea44b0a9c9d8e76e987f4603de"} Jan 26 17:18:26 crc kubenswrapper[4754]: I0126 17:18:26.083091 4754 scope.go:117] "RemoveContainer" containerID="95b41b32fb093081378f984b4a590341fb635fd1ed3785ec87ad1db78080ff99" Jan 26 17:18:26 crc kubenswrapper[4754]: I0126 17:18:26.107304 4754 scope.go:117] "RemoveContainer" containerID="c553cf4f5c0479e11cee44a2fd1c1c574a2cb4a6536dec6cc0242f7317cbc1fb" Jan 26 17:18:30 crc kubenswrapper[4754]: I0126 17:18:30.056916 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-5xmfz"] Jan 26 17:18:30 crc kubenswrapper[4754]: I0126 17:18:30.087434 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-5xmfz"] Jan 26 17:18:31 crc kubenswrapper[4754]: I0126 17:18:31.781254 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e934228b-3d4e-4c15-baa6-5c03e0fceb23" path="/var/lib/kubelet/pods/e934228b-3d4e-4c15-baa6-5c03e0fceb23/volumes" Jan 26 17:18:39 crc kubenswrapper[4754]: I0126 17:18:39.041199 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-whnhn"] Jan 26 17:18:39 crc kubenswrapper[4754]: I0126 17:18:39.056690 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-whnhn"] Jan 26 17:18:39 crc kubenswrapper[4754]: I0126 17:18:39.779655 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01246f69-d6bc-4f70-a50f-4d7c8e4a7620" path="/var/lib/kubelet/pods/01246f69-d6bc-4f70-a50f-4d7c8e4a7620/volumes" Jan 26 17:18:55 crc kubenswrapper[4754]: I0126 17:18:55.037494 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-9spps"] Jan 26 17:18:55 crc kubenswrapper[4754]: I0126 17:18:55.050650 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-9spps"] Jan 26 17:18:55 crc kubenswrapper[4754]: I0126 17:18:55.778652 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ea094395-bb3c-4464-829f-313ab6b9e14c" path="/var/lib/kubelet/pods/ea094395-bb3c-4464-829f-313ab6b9e14c/volumes" Jan 26 17:19:15 crc kubenswrapper[4754]: I0126 17:19:15.042718 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-7hskr"] Jan 26 17:19:15 crc kubenswrapper[4754]: I0126 17:19:15.053434 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-7hskr"] Jan 26 17:19:15 crc kubenswrapper[4754]: I0126 17:19:15.777483 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ab3f008d-78ab-4b73-962d-4447f312c9fa" path="/var/lib/kubelet/pods/ab3f008d-78ab-4b73-962d-4447f312c9fa/volumes" Jan 26 17:19:26 crc kubenswrapper[4754]: I0126 17:19:26.202252 4754 scope.go:117] "RemoveContainer" containerID="eda7e173767be2f94247fe84877d1dc7a95556f64d1cc275cf4ccf7dd0a1ae64" Jan 26 17:19:26 crc kubenswrapper[4754]: I0126 17:19:26.244049 4754 scope.go:117] "RemoveContainer" containerID="3c17b0e81ba837680723ffaf2c7fb810b45ec801d7900bb2517aa72f884caa3d" Jan 26 17:19:26 crc kubenswrapper[4754]: I0126 17:19:26.307940 4754 scope.go:117] "RemoveContainer" containerID="fa8641d204e41cbb652b48277e0e1d4c609dfb5b9fb94837cbf80deac687370c" Jan 26 17:19:26 crc kubenswrapper[4754]: I0126 17:19:26.364631 4754 scope.go:117] "RemoveContainer" containerID="867f10be7d6d79bb9f1fbc8fcf04eedd21efb11b71b3248822ebec5b8840613a" Jan 26 17:19:26 crc kubenswrapper[4754]: I0126 17:19:26.384872 4754 scope.go:117] "RemoveContainer" containerID="a0aac0c2b2fa1e7140caf64c6f53502aa2cdbab58168b149446176ac7c4866bb" Jan 26 17:19:26 crc kubenswrapper[4754]: I0126 17:19:26.413284 4754 scope.go:117] "RemoveContainer" containerID="eecbbaf507684103d27ed11791648478ec4f0b4a963953c60909a7ac24b4e6db" Jan 26 17:19:41 crc kubenswrapper[4754]: I0126 17:19:41.040380 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-dzfm9"] Jan 26 17:19:41 crc kubenswrapper[4754]: I0126 17:19:41.051500 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-dzfm9"] Jan 26 17:19:41 crc kubenswrapper[4754]: I0126 17:19:41.777792 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7d5a6ee0-79d3-4112-a180-d2211fee09f6" path="/var/lib/kubelet/pods/7d5a6ee0-79d3-4112-a180-d2211fee09f6/volumes" Jan 26 17:19:42 crc kubenswrapper[4754]: I0126 17:19:42.644198 4754 generic.go:334] "Generic (PLEG): container finished" podID="55b57fed-619a-44b7-af60-f6d7a43943f1" containerID="c7081a1de193a8c59596a064551979f778ea21453c14cdfa6f8086ba8dcbd7de" exitCode=0 Jan 26 17:19:42 crc kubenswrapper[4754]: I0126 17:19:42.644271 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-rrhnq" event={"ID":"55b57fed-619a-44b7-af60-f6d7a43943f1","Type":"ContainerDied","Data":"c7081a1de193a8c59596a064551979f778ea21453c14cdfa6f8086ba8dcbd7de"} Jan 26 17:19:44 crc kubenswrapper[4754]: I0126 17:19:44.070747 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-rrhnq" Jan 26 17:19:44 crc kubenswrapper[4754]: I0126 17:19:44.129417 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/55b57fed-619a-44b7-af60-f6d7a43943f1-ssh-key-openstack-edpm-ipam\") pod \"55b57fed-619a-44b7-af60-f6d7a43943f1\" (UID: \"55b57fed-619a-44b7-af60-f6d7a43943f1\") " Jan 26 17:19:44 crc kubenswrapper[4754]: I0126 17:19:44.129488 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/55b57fed-619a-44b7-af60-f6d7a43943f1-inventory\") pod \"55b57fed-619a-44b7-af60-f6d7a43943f1\" (UID: \"55b57fed-619a-44b7-af60-f6d7a43943f1\") " Jan 26 17:19:44 crc kubenswrapper[4754]: I0126 17:19:44.129551 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fp5jn\" (UniqueName: \"kubernetes.io/projected/55b57fed-619a-44b7-af60-f6d7a43943f1-kube-api-access-fp5jn\") pod \"55b57fed-619a-44b7-af60-f6d7a43943f1\" (UID: \"55b57fed-619a-44b7-af60-f6d7a43943f1\") " Jan 26 17:19:44 crc kubenswrapper[4754]: I0126 17:19:44.135973 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/55b57fed-619a-44b7-af60-f6d7a43943f1-kube-api-access-fp5jn" (OuterVolumeSpecName: "kube-api-access-fp5jn") pod "55b57fed-619a-44b7-af60-f6d7a43943f1" (UID: "55b57fed-619a-44b7-af60-f6d7a43943f1"). InnerVolumeSpecName "kube-api-access-fp5jn". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:19:44 crc kubenswrapper[4754]: I0126 17:19:44.159527 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/55b57fed-619a-44b7-af60-f6d7a43943f1-inventory" (OuterVolumeSpecName: "inventory") pod "55b57fed-619a-44b7-af60-f6d7a43943f1" (UID: "55b57fed-619a-44b7-af60-f6d7a43943f1"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:19:44 crc kubenswrapper[4754]: I0126 17:19:44.166274 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/55b57fed-619a-44b7-af60-f6d7a43943f1-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "55b57fed-619a-44b7-af60-f6d7a43943f1" (UID: "55b57fed-619a-44b7-af60-f6d7a43943f1"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:19:44 crc kubenswrapper[4754]: I0126 17:19:44.231274 4754 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/55b57fed-619a-44b7-af60-f6d7a43943f1-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Jan 26 17:19:44 crc kubenswrapper[4754]: I0126 17:19:44.231316 4754 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/55b57fed-619a-44b7-af60-f6d7a43943f1-inventory\") on node \"crc\" DevicePath \"\"" Jan 26 17:19:44 crc kubenswrapper[4754]: I0126 17:19:44.231328 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fp5jn\" (UniqueName: \"kubernetes.io/projected/55b57fed-619a-44b7-af60-f6d7a43943f1-kube-api-access-fp5jn\") on node \"crc\" DevicePath \"\"" Jan 26 17:19:44 crc kubenswrapper[4754]: I0126 17:19:44.661865 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-rrhnq" event={"ID":"55b57fed-619a-44b7-af60-f6d7a43943f1","Type":"ContainerDied","Data":"4f88e24a27b0baf222d20554ed3ddd0d702d536ac920f7c3d425e7426f3b1f51"} Jan 26 17:19:44 crc kubenswrapper[4754]: I0126 17:19:44.661909 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4f88e24a27b0baf222d20554ed3ddd0d702d536ac920f7c3d425e7426f3b1f51" Jan 26 17:19:44 crc kubenswrapper[4754]: I0126 17:19:44.661924 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-rrhnq" Jan 26 17:19:44 crc kubenswrapper[4754]: I0126 17:19:44.780581 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-5d95m"] Jan 26 17:19:44 crc kubenswrapper[4754]: E0126 17:19:44.781089 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="55b57fed-619a-44b7-af60-f6d7a43943f1" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Jan 26 17:19:44 crc kubenswrapper[4754]: I0126 17:19:44.781111 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="55b57fed-619a-44b7-af60-f6d7a43943f1" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Jan 26 17:19:44 crc kubenswrapper[4754]: I0126 17:19:44.781346 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="55b57fed-619a-44b7-af60-f6d7a43943f1" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Jan 26 17:19:44 crc kubenswrapper[4754]: I0126 17:19:44.782197 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-5d95m" Jan 26 17:19:44 crc kubenswrapper[4754]: I0126 17:19:44.784350 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Jan 26 17:19:44 crc kubenswrapper[4754]: I0126 17:19:44.784482 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Jan 26 17:19:44 crc kubenswrapper[4754]: I0126 17:19:44.788303 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-fz7z8" Jan 26 17:19:44 crc kubenswrapper[4754]: I0126 17:19:44.788611 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Jan 26 17:19:44 crc kubenswrapper[4754]: I0126 17:19:44.795790 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-5d95m"] Jan 26 17:19:44 crc kubenswrapper[4754]: I0126 17:19:44.850784 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kplrj\" (UniqueName: \"kubernetes.io/projected/51a9ce2d-f224-449d-8988-950a60783ddb-kube-api-access-kplrj\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-5d95m\" (UID: \"51a9ce2d-f224-449d-8988-950a60783ddb\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-5d95m" Jan 26 17:19:44 crc kubenswrapper[4754]: I0126 17:19:44.852313 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/51a9ce2d-f224-449d-8988-950a60783ddb-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-5d95m\" (UID: \"51a9ce2d-f224-449d-8988-950a60783ddb\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-5d95m" Jan 26 17:19:44 crc kubenswrapper[4754]: I0126 17:19:44.853109 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/51a9ce2d-f224-449d-8988-950a60783ddb-ssh-key-openstack-edpm-ipam\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-5d95m\" (UID: \"51a9ce2d-f224-449d-8988-950a60783ddb\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-5d95m" Jan 26 17:19:44 crc kubenswrapper[4754]: I0126 17:19:44.955951 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/51a9ce2d-f224-449d-8988-950a60783ddb-ssh-key-openstack-edpm-ipam\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-5d95m\" (UID: \"51a9ce2d-f224-449d-8988-950a60783ddb\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-5d95m" Jan 26 17:19:44 crc kubenswrapper[4754]: I0126 17:19:44.956088 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kplrj\" (UniqueName: \"kubernetes.io/projected/51a9ce2d-f224-449d-8988-950a60783ddb-kube-api-access-kplrj\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-5d95m\" (UID: \"51a9ce2d-f224-449d-8988-950a60783ddb\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-5d95m" Jan 26 17:19:44 crc kubenswrapper[4754]: I0126 17:19:44.956254 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/51a9ce2d-f224-449d-8988-950a60783ddb-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-5d95m\" (UID: \"51a9ce2d-f224-449d-8988-950a60783ddb\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-5d95m" Jan 26 17:19:44 crc kubenswrapper[4754]: I0126 17:19:44.960279 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/51a9ce2d-f224-449d-8988-950a60783ddb-ssh-key-openstack-edpm-ipam\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-5d95m\" (UID: \"51a9ce2d-f224-449d-8988-950a60783ddb\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-5d95m" Jan 26 17:19:44 crc kubenswrapper[4754]: I0126 17:19:44.960291 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/51a9ce2d-f224-449d-8988-950a60783ddb-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-5d95m\" (UID: \"51a9ce2d-f224-449d-8988-950a60783ddb\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-5d95m" Jan 26 17:19:44 crc kubenswrapper[4754]: I0126 17:19:44.972570 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kplrj\" (UniqueName: \"kubernetes.io/projected/51a9ce2d-f224-449d-8988-950a60783ddb-kube-api-access-kplrj\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-5d95m\" (UID: \"51a9ce2d-f224-449d-8988-950a60783ddb\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-5d95m" Jan 26 17:19:45 crc kubenswrapper[4754]: I0126 17:19:45.107544 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-5d95m" Jan 26 17:19:45 crc kubenswrapper[4754]: I0126 17:19:45.621797 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-5d95m"] Jan 26 17:19:45 crc kubenswrapper[4754]: I0126 17:19:45.671741 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-5d95m" event={"ID":"51a9ce2d-f224-449d-8988-950a60783ddb","Type":"ContainerStarted","Data":"9af15fc57f567d98779b4557cc814580abee396e04f055eb4fd9a1745d2f71ce"} Jan 26 17:19:46 crc kubenswrapper[4754]: I0126 17:19:46.679935 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-5d95m" event={"ID":"51a9ce2d-f224-449d-8988-950a60783ddb","Type":"ContainerStarted","Data":"09a8646b8a9a63a4de0bdce4f0c3d1b29353ce3da1ddd4f0659ca8bba4007f41"} Jan 26 17:19:46 crc kubenswrapper[4754]: I0126 17:19:46.701741 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-5d95m" podStartSLOduration=2.008766861 podStartE2EDuration="2.70172058s" podCreationTimestamp="2026-01-26 17:19:44 +0000 UTC" firstStartedPulling="2026-01-26 17:19:45.628599935 +0000 UTC m=+1952.152780369" lastFinishedPulling="2026-01-26 17:19:46.321553654 +0000 UTC m=+1952.845734088" observedRunningTime="2026-01-26 17:19:46.695066594 +0000 UTC m=+1953.219247038" watchObservedRunningTime="2026-01-26 17:19:46.70172058 +0000 UTC m=+1953.225901004" Jan 26 17:20:05 crc kubenswrapper[4754]: I0126 17:20:05.041461 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-813d-account-create-update-wlvxd"] Jan 26 17:20:05 crc kubenswrapper[4754]: I0126 17:20:05.055997 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-2d04-account-create-update-lnbkh"] Jan 26 17:20:05 crc kubenswrapper[4754]: I0126 17:20:05.065857 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-e6d7-account-create-update-hs5pg"] Jan 26 17:20:05 crc kubenswrapper[4754]: I0126 17:20:05.073845 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-6qdbg"] Jan 26 17:20:05 crc kubenswrapper[4754]: I0126 17:20:05.097026 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-bjmmr"] Jan 26 17:20:05 crc kubenswrapper[4754]: I0126 17:20:05.097198 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-813d-account-create-update-wlvxd"] Jan 26 17:20:05 crc kubenswrapper[4754]: I0126 17:20:05.105959 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-6qdbg"] Jan 26 17:20:05 crc kubenswrapper[4754]: I0126 17:20:05.115291 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-bjmmr"] Jan 26 17:20:05 crc kubenswrapper[4754]: I0126 17:20:05.126138 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-2d04-account-create-update-lnbkh"] Jan 26 17:20:05 crc kubenswrapper[4754]: I0126 17:20:05.136293 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-j7p2d"] Jan 26 17:20:05 crc kubenswrapper[4754]: I0126 17:20:05.146821 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-e6d7-account-create-update-hs5pg"] Jan 26 17:20:05 crc kubenswrapper[4754]: I0126 17:20:05.154813 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-j7p2d"] Jan 26 17:20:05 crc kubenswrapper[4754]: I0126 17:20:05.672703 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-lwkkt"] Jan 26 17:20:05 crc kubenswrapper[4754]: I0126 17:20:05.674825 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lwkkt" Jan 26 17:20:05 crc kubenswrapper[4754]: I0126 17:20:05.691904 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-lwkkt"] Jan 26 17:20:05 crc kubenswrapper[4754]: I0126 17:20:05.733889 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e8118a9a-872d-4d9f-9d87-f45973ac75ef-utilities\") pod \"certified-operators-lwkkt\" (UID: \"e8118a9a-872d-4d9f-9d87-f45973ac75ef\") " pod="openshift-marketplace/certified-operators-lwkkt" Jan 26 17:20:05 crc kubenswrapper[4754]: I0126 17:20:05.734019 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-66gcf\" (UniqueName: \"kubernetes.io/projected/e8118a9a-872d-4d9f-9d87-f45973ac75ef-kube-api-access-66gcf\") pod \"certified-operators-lwkkt\" (UID: \"e8118a9a-872d-4d9f-9d87-f45973ac75ef\") " pod="openshift-marketplace/certified-operators-lwkkt" Jan 26 17:20:05 crc kubenswrapper[4754]: I0126 17:20:05.734089 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e8118a9a-872d-4d9f-9d87-f45973ac75ef-catalog-content\") pod \"certified-operators-lwkkt\" (UID: \"e8118a9a-872d-4d9f-9d87-f45973ac75ef\") " pod="openshift-marketplace/certified-operators-lwkkt" Jan 26 17:20:05 crc kubenswrapper[4754]: I0126 17:20:05.780137 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1597209c-abf0-4bd4-91b4-a8661e7e496b" path="/var/lib/kubelet/pods/1597209c-abf0-4bd4-91b4-a8661e7e496b/volumes" Jan 26 17:20:05 crc kubenswrapper[4754]: I0126 17:20:05.781121 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="755ea18f-9963-409b-a3a2-56d110120dc8" path="/var/lib/kubelet/pods/755ea18f-9963-409b-a3a2-56d110120dc8/volumes" Jan 26 17:20:05 crc kubenswrapper[4754]: I0126 17:20:05.781909 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="98a8c733-ba31-46a3-97ea-409b69050b02" path="/var/lib/kubelet/pods/98a8c733-ba31-46a3-97ea-409b69050b02/volumes" Jan 26 17:20:05 crc kubenswrapper[4754]: I0126 17:20:05.782530 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bef625bb-6bee-4c89-a9b4-4f695ab4ffd5" path="/var/lib/kubelet/pods/bef625bb-6bee-4c89-a9b4-4f695ab4ffd5/volumes" Jan 26 17:20:05 crc kubenswrapper[4754]: I0126 17:20:05.783861 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cf071f6e-289f-4f86-9a39-e4dc62335b14" path="/var/lib/kubelet/pods/cf071f6e-289f-4f86-9a39-e4dc62335b14/volumes" Jan 26 17:20:05 crc kubenswrapper[4754]: I0126 17:20:05.784753 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f5c367c8-74fd-4073-bbc6-2f6d4373f5ef" path="/var/lib/kubelet/pods/f5c367c8-74fd-4073-bbc6-2f6d4373f5ef/volumes" Jan 26 17:20:05 crc kubenswrapper[4754]: I0126 17:20:05.836053 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-66gcf\" (UniqueName: \"kubernetes.io/projected/e8118a9a-872d-4d9f-9d87-f45973ac75ef-kube-api-access-66gcf\") pod \"certified-operators-lwkkt\" (UID: \"e8118a9a-872d-4d9f-9d87-f45973ac75ef\") " pod="openshift-marketplace/certified-operators-lwkkt" Jan 26 17:20:05 crc kubenswrapper[4754]: I0126 17:20:05.836554 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e8118a9a-872d-4d9f-9d87-f45973ac75ef-catalog-content\") pod \"certified-operators-lwkkt\" (UID: \"e8118a9a-872d-4d9f-9d87-f45973ac75ef\") " pod="openshift-marketplace/certified-operators-lwkkt" Jan 26 17:20:05 crc kubenswrapper[4754]: I0126 17:20:05.836753 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e8118a9a-872d-4d9f-9d87-f45973ac75ef-utilities\") pod \"certified-operators-lwkkt\" (UID: \"e8118a9a-872d-4d9f-9d87-f45973ac75ef\") " pod="openshift-marketplace/certified-operators-lwkkt" Jan 26 17:20:05 crc kubenswrapper[4754]: I0126 17:20:05.837039 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e8118a9a-872d-4d9f-9d87-f45973ac75ef-catalog-content\") pod \"certified-operators-lwkkt\" (UID: \"e8118a9a-872d-4d9f-9d87-f45973ac75ef\") " pod="openshift-marketplace/certified-operators-lwkkt" Jan 26 17:20:05 crc kubenswrapper[4754]: I0126 17:20:05.837503 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e8118a9a-872d-4d9f-9d87-f45973ac75ef-utilities\") pod \"certified-operators-lwkkt\" (UID: \"e8118a9a-872d-4d9f-9d87-f45973ac75ef\") " pod="openshift-marketplace/certified-operators-lwkkt" Jan 26 17:20:05 crc kubenswrapper[4754]: I0126 17:20:05.862796 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-66gcf\" (UniqueName: \"kubernetes.io/projected/e8118a9a-872d-4d9f-9d87-f45973ac75ef-kube-api-access-66gcf\") pod \"certified-operators-lwkkt\" (UID: \"e8118a9a-872d-4d9f-9d87-f45973ac75ef\") " pod="openshift-marketplace/certified-operators-lwkkt" Jan 26 17:20:05 crc kubenswrapper[4754]: I0126 17:20:05.994541 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lwkkt" Jan 26 17:20:06 crc kubenswrapper[4754]: I0126 17:20:06.536377 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-lwkkt"] Jan 26 17:20:06 crc kubenswrapper[4754]: I0126 17:20:06.846811 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lwkkt" event={"ID":"e8118a9a-872d-4d9f-9d87-f45973ac75ef","Type":"ContainerStarted","Data":"030e1551fcdc5b1008454d9d8adc7ef7f3831db0fc0afe16be227a67e23d3fea"} Jan 26 17:20:07 crc kubenswrapper[4754]: I0126 17:20:07.858224 4754 generic.go:334] "Generic (PLEG): container finished" podID="e8118a9a-872d-4d9f-9d87-f45973ac75ef" containerID="8d15a16eeef87b8b1161a17da98a97f832b7594af68c17db16653d1d7d53e435" exitCode=0 Jan 26 17:20:07 crc kubenswrapper[4754]: I0126 17:20:07.858274 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lwkkt" event={"ID":"e8118a9a-872d-4d9f-9d87-f45973ac75ef","Type":"ContainerDied","Data":"8d15a16eeef87b8b1161a17da98a97f832b7594af68c17db16653d1d7d53e435"} Jan 26 17:20:08 crc kubenswrapper[4754]: I0126 17:20:08.869041 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lwkkt" event={"ID":"e8118a9a-872d-4d9f-9d87-f45973ac75ef","Type":"ContainerStarted","Data":"68aa7bfb8a8141cd01e396f6378e330afcc2c4e36a4eaf344b73bb90c0f6a160"} Jan 26 17:20:09 crc kubenswrapper[4754]: I0126 17:20:09.880745 4754 generic.go:334] "Generic (PLEG): container finished" podID="e8118a9a-872d-4d9f-9d87-f45973ac75ef" containerID="68aa7bfb8a8141cd01e396f6378e330afcc2c4e36a4eaf344b73bb90c0f6a160" exitCode=0 Jan 26 17:20:09 crc kubenswrapper[4754]: I0126 17:20:09.880825 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lwkkt" event={"ID":"e8118a9a-872d-4d9f-9d87-f45973ac75ef","Type":"ContainerDied","Data":"68aa7bfb8a8141cd01e396f6378e330afcc2c4e36a4eaf344b73bb90c0f6a160"} Jan 26 17:20:10 crc kubenswrapper[4754]: I0126 17:20:10.890782 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lwkkt" event={"ID":"e8118a9a-872d-4d9f-9d87-f45973ac75ef","Type":"ContainerStarted","Data":"3f86ddc7a24beee6b230d80171c4512302f02214e8f80156d7aa5fd165822593"} Jan 26 17:20:10 crc kubenswrapper[4754]: I0126 17:20:10.913229 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-lwkkt" podStartSLOduration=3.354847727 podStartE2EDuration="5.913208543s" podCreationTimestamp="2026-01-26 17:20:05 +0000 UTC" firstStartedPulling="2026-01-26 17:20:07.860248986 +0000 UTC m=+1974.384429420" lastFinishedPulling="2026-01-26 17:20:10.418609802 +0000 UTC m=+1976.942790236" observedRunningTime="2026-01-26 17:20:10.905814326 +0000 UTC m=+1977.429994760" watchObservedRunningTime="2026-01-26 17:20:10.913208543 +0000 UTC m=+1977.437388987" Jan 26 17:20:15 crc kubenswrapper[4754]: I0126 17:20:15.995446 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-lwkkt" Jan 26 17:20:15 crc kubenswrapper[4754]: I0126 17:20:15.995988 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-lwkkt" Jan 26 17:20:16 crc kubenswrapper[4754]: I0126 17:20:16.045076 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-lwkkt" Jan 26 17:20:16 crc kubenswrapper[4754]: I0126 17:20:16.996337 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-lwkkt" Jan 26 17:20:17 crc kubenswrapper[4754]: I0126 17:20:17.059106 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-lwkkt"] Jan 26 17:20:18 crc kubenswrapper[4754]: I0126 17:20:18.954277 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-lwkkt" podUID="e8118a9a-872d-4d9f-9d87-f45973ac75ef" containerName="registry-server" containerID="cri-o://3f86ddc7a24beee6b230d80171c4512302f02214e8f80156d7aa5fd165822593" gracePeriod=2 Jan 26 17:20:19 crc kubenswrapper[4754]: I0126 17:20:19.617169 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lwkkt" Jan 26 17:20:19 crc kubenswrapper[4754]: I0126 17:20:19.706039 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e8118a9a-872d-4d9f-9d87-f45973ac75ef-utilities\") pod \"e8118a9a-872d-4d9f-9d87-f45973ac75ef\" (UID: \"e8118a9a-872d-4d9f-9d87-f45973ac75ef\") " Jan 26 17:20:19 crc kubenswrapper[4754]: I0126 17:20:19.706134 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e8118a9a-872d-4d9f-9d87-f45973ac75ef-catalog-content\") pod \"e8118a9a-872d-4d9f-9d87-f45973ac75ef\" (UID: \"e8118a9a-872d-4d9f-9d87-f45973ac75ef\") " Jan 26 17:20:19 crc kubenswrapper[4754]: I0126 17:20:19.706239 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-66gcf\" (UniqueName: \"kubernetes.io/projected/e8118a9a-872d-4d9f-9d87-f45973ac75ef-kube-api-access-66gcf\") pod \"e8118a9a-872d-4d9f-9d87-f45973ac75ef\" (UID: \"e8118a9a-872d-4d9f-9d87-f45973ac75ef\") " Jan 26 17:20:19 crc kubenswrapper[4754]: I0126 17:20:19.707164 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e8118a9a-872d-4d9f-9d87-f45973ac75ef-utilities" (OuterVolumeSpecName: "utilities") pod "e8118a9a-872d-4d9f-9d87-f45973ac75ef" (UID: "e8118a9a-872d-4d9f-9d87-f45973ac75ef"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:20:19 crc kubenswrapper[4754]: I0126 17:20:19.712608 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e8118a9a-872d-4d9f-9d87-f45973ac75ef-kube-api-access-66gcf" (OuterVolumeSpecName: "kube-api-access-66gcf") pod "e8118a9a-872d-4d9f-9d87-f45973ac75ef" (UID: "e8118a9a-872d-4d9f-9d87-f45973ac75ef"). InnerVolumeSpecName "kube-api-access-66gcf". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:20:19 crc kubenswrapper[4754]: I0126 17:20:19.752903 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e8118a9a-872d-4d9f-9d87-f45973ac75ef-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e8118a9a-872d-4d9f-9d87-f45973ac75ef" (UID: "e8118a9a-872d-4d9f-9d87-f45973ac75ef"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:20:19 crc kubenswrapper[4754]: I0126 17:20:19.810200 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-66gcf\" (UniqueName: \"kubernetes.io/projected/e8118a9a-872d-4d9f-9d87-f45973ac75ef-kube-api-access-66gcf\") on node \"crc\" DevicePath \"\"" Jan 26 17:20:19 crc kubenswrapper[4754]: I0126 17:20:19.810631 4754 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e8118a9a-872d-4d9f-9d87-f45973ac75ef-utilities\") on node \"crc\" DevicePath \"\"" Jan 26 17:20:19 crc kubenswrapper[4754]: I0126 17:20:19.810646 4754 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e8118a9a-872d-4d9f-9d87-f45973ac75ef-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 26 17:20:19 crc kubenswrapper[4754]: I0126 17:20:19.965461 4754 generic.go:334] "Generic (PLEG): container finished" podID="e8118a9a-872d-4d9f-9d87-f45973ac75ef" containerID="3f86ddc7a24beee6b230d80171c4512302f02214e8f80156d7aa5fd165822593" exitCode=0 Jan 26 17:20:19 crc kubenswrapper[4754]: I0126 17:20:19.965518 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lwkkt" event={"ID":"e8118a9a-872d-4d9f-9d87-f45973ac75ef","Type":"ContainerDied","Data":"3f86ddc7a24beee6b230d80171c4512302f02214e8f80156d7aa5fd165822593"} Jan 26 17:20:19 crc kubenswrapper[4754]: I0126 17:20:19.965549 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lwkkt" event={"ID":"e8118a9a-872d-4d9f-9d87-f45973ac75ef","Type":"ContainerDied","Data":"030e1551fcdc5b1008454d9d8adc7ef7f3831db0fc0afe16be227a67e23d3fea"} Jan 26 17:20:19 crc kubenswrapper[4754]: I0126 17:20:19.965570 4754 scope.go:117] "RemoveContainer" containerID="3f86ddc7a24beee6b230d80171c4512302f02214e8f80156d7aa5fd165822593" Jan 26 17:20:19 crc kubenswrapper[4754]: I0126 17:20:19.965755 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lwkkt" Jan 26 17:20:19 crc kubenswrapper[4754]: I0126 17:20:19.989863 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-lwkkt"] Jan 26 17:20:19 crc kubenswrapper[4754]: I0126 17:20:19.997776 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-lwkkt"] Jan 26 17:20:19 crc kubenswrapper[4754]: I0126 17:20:19.999160 4754 scope.go:117] "RemoveContainer" containerID="68aa7bfb8a8141cd01e396f6378e330afcc2c4e36a4eaf344b73bb90c0f6a160" Jan 26 17:20:20 crc kubenswrapper[4754]: I0126 17:20:20.029989 4754 scope.go:117] "RemoveContainer" containerID="8d15a16eeef87b8b1161a17da98a97f832b7594af68c17db16653d1d7d53e435" Jan 26 17:20:20 crc kubenswrapper[4754]: I0126 17:20:20.066368 4754 scope.go:117] "RemoveContainer" containerID="3f86ddc7a24beee6b230d80171c4512302f02214e8f80156d7aa5fd165822593" Jan 26 17:20:20 crc kubenswrapper[4754]: E0126 17:20:20.066899 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3f86ddc7a24beee6b230d80171c4512302f02214e8f80156d7aa5fd165822593\": container with ID starting with 3f86ddc7a24beee6b230d80171c4512302f02214e8f80156d7aa5fd165822593 not found: ID does not exist" containerID="3f86ddc7a24beee6b230d80171c4512302f02214e8f80156d7aa5fd165822593" Jan 26 17:20:20 crc kubenswrapper[4754]: I0126 17:20:20.066940 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3f86ddc7a24beee6b230d80171c4512302f02214e8f80156d7aa5fd165822593"} err="failed to get container status \"3f86ddc7a24beee6b230d80171c4512302f02214e8f80156d7aa5fd165822593\": rpc error: code = NotFound desc = could not find container \"3f86ddc7a24beee6b230d80171c4512302f02214e8f80156d7aa5fd165822593\": container with ID starting with 3f86ddc7a24beee6b230d80171c4512302f02214e8f80156d7aa5fd165822593 not found: ID does not exist" Jan 26 17:20:20 crc kubenswrapper[4754]: I0126 17:20:20.066969 4754 scope.go:117] "RemoveContainer" containerID="68aa7bfb8a8141cd01e396f6378e330afcc2c4e36a4eaf344b73bb90c0f6a160" Jan 26 17:20:20 crc kubenswrapper[4754]: E0126 17:20:20.067527 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"68aa7bfb8a8141cd01e396f6378e330afcc2c4e36a4eaf344b73bb90c0f6a160\": container with ID starting with 68aa7bfb8a8141cd01e396f6378e330afcc2c4e36a4eaf344b73bb90c0f6a160 not found: ID does not exist" containerID="68aa7bfb8a8141cd01e396f6378e330afcc2c4e36a4eaf344b73bb90c0f6a160" Jan 26 17:20:20 crc kubenswrapper[4754]: I0126 17:20:20.067579 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"68aa7bfb8a8141cd01e396f6378e330afcc2c4e36a4eaf344b73bb90c0f6a160"} err="failed to get container status \"68aa7bfb8a8141cd01e396f6378e330afcc2c4e36a4eaf344b73bb90c0f6a160\": rpc error: code = NotFound desc = could not find container \"68aa7bfb8a8141cd01e396f6378e330afcc2c4e36a4eaf344b73bb90c0f6a160\": container with ID starting with 68aa7bfb8a8141cd01e396f6378e330afcc2c4e36a4eaf344b73bb90c0f6a160 not found: ID does not exist" Jan 26 17:20:20 crc kubenswrapper[4754]: I0126 17:20:20.067615 4754 scope.go:117] "RemoveContainer" containerID="8d15a16eeef87b8b1161a17da98a97f832b7594af68c17db16653d1d7d53e435" Jan 26 17:20:20 crc kubenswrapper[4754]: E0126 17:20:20.068024 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8d15a16eeef87b8b1161a17da98a97f832b7594af68c17db16653d1d7d53e435\": container with ID starting with 8d15a16eeef87b8b1161a17da98a97f832b7594af68c17db16653d1d7d53e435 not found: ID does not exist" containerID="8d15a16eeef87b8b1161a17da98a97f832b7594af68c17db16653d1d7d53e435" Jan 26 17:20:20 crc kubenswrapper[4754]: I0126 17:20:20.068060 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8d15a16eeef87b8b1161a17da98a97f832b7594af68c17db16653d1d7d53e435"} err="failed to get container status \"8d15a16eeef87b8b1161a17da98a97f832b7594af68c17db16653d1d7d53e435\": rpc error: code = NotFound desc = could not find container \"8d15a16eeef87b8b1161a17da98a97f832b7594af68c17db16653d1d7d53e435\": container with ID starting with 8d15a16eeef87b8b1161a17da98a97f832b7594af68c17db16653d1d7d53e435 not found: ID does not exist" Jan 26 17:20:21 crc kubenswrapper[4754]: I0126 17:20:21.778987 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e8118a9a-872d-4d9f-9d87-f45973ac75ef" path="/var/lib/kubelet/pods/e8118a9a-872d-4d9f-9d87-f45973ac75ef/volumes" Jan 26 17:20:26 crc kubenswrapper[4754]: I0126 17:20:26.572301 4754 scope.go:117] "RemoveContainer" containerID="2ac2874e0106d3797b0f67a75cd93ccea19413c4e35875f30d77331553081397" Jan 26 17:20:26 crc kubenswrapper[4754]: I0126 17:20:26.598352 4754 scope.go:117] "RemoveContainer" containerID="c502c69ad12ddc1b2b3981e832270d6541d21c9abdf23ec4e61ebfe2fd929ac9" Jan 26 17:20:26 crc kubenswrapper[4754]: I0126 17:20:26.661213 4754 scope.go:117] "RemoveContainer" containerID="f52878497a7ebd51d5218180b3f05165fc0ac52700a5ed47fb29b6523f6597cb" Jan 26 17:20:26 crc kubenswrapper[4754]: I0126 17:20:26.706457 4754 scope.go:117] "RemoveContainer" containerID="7520af6d81b045bab0b0552880d2a9323b1410e4bdbc087b69f3024f2836cebf" Jan 26 17:20:26 crc kubenswrapper[4754]: I0126 17:20:26.752638 4754 scope.go:117] "RemoveContainer" containerID="156feceeb22f23d1a2ff8accdb5d5abdf7fe3efad41579a732554fecce8bf645" Jan 26 17:20:26 crc kubenswrapper[4754]: I0126 17:20:26.815181 4754 scope.go:117] "RemoveContainer" containerID="d07947f249ed866995ee2e62e47cf9e3718926322c0dbaebe8e36d9013ee73f5" Jan 26 17:20:26 crc kubenswrapper[4754]: I0126 17:20:26.855718 4754 scope.go:117] "RemoveContainer" containerID="e3184d6ce076367c5bd58f3200fab9ce3aff97ec4c504666561622fa2dc4f696" Jan 26 17:20:37 crc kubenswrapper[4754]: I0126 17:20:37.129429 4754 patch_prober.go:28] interesting pod/machine-config-daemon-x65wv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 26 17:20:37 crc kubenswrapper[4754]: I0126 17:20:37.130022 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 26 17:20:42 crc kubenswrapper[4754]: I0126 17:20:42.046695 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-8lrb8"] Jan 26 17:20:42 crc kubenswrapper[4754]: I0126 17:20:42.054797 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-8lrb8"] Jan 26 17:20:43 crc kubenswrapper[4754]: I0126 17:20:43.779361 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="db76cab2-a25b-4072-a7c3-fcc0d6e5401a" path="/var/lib/kubelet/pods/db76cab2-a25b-4072-a7c3-fcc0d6e5401a/volumes" Jan 26 17:21:04 crc kubenswrapper[4754]: I0126 17:21:04.043262 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-njgdn"] Jan 26 17:21:04 crc kubenswrapper[4754]: I0126 17:21:04.049818 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-njgdn"] Jan 26 17:21:05 crc kubenswrapper[4754]: I0126 17:21:05.343196 4754 generic.go:334] "Generic (PLEG): container finished" podID="51a9ce2d-f224-449d-8988-950a60783ddb" containerID="09a8646b8a9a63a4de0bdce4f0c3d1b29353ce3da1ddd4f0659ca8bba4007f41" exitCode=0 Jan 26 17:21:05 crc kubenswrapper[4754]: I0126 17:21:05.343281 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-5d95m" event={"ID":"51a9ce2d-f224-449d-8988-950a60783ddb","Type":"ContainerDied","Data":"09a8646b8a9a63a4de0bdce4f0c3d1b29353ce3da1ddd4f0659ca8bba4007f41"} Jan 26 17:21:05 crc kubenswrapper[4754]: I0126 17:21:05.779353 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="54617228-f6c0-4db0-a468-2b9283bcae21" path="/var/lib/kubelet/pods/54617228-f6c0-4db0-a468-2b9283bcae21/volumes" Jan 26 17:21:06 crc kubenswrapper[4754]: I0126 17:21:06.732323 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-5d95m" Jan 26 17:21:06 crc kubenswrapper[4754]: I0126 17:21:06.802354 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/51a9ce2d-f224-449d-8988-950a60783ddb-ssh-key-openstack-edpm-ipam\") pod \"51a9ce2d-f224-449d-8988-950a60783ddb\" (UID: \"51a9ce2d-f224-449d-8988-950a60783ddb\") " Jan 26 17:21:06 crc kubenswrapper[4754]: I0126 17:21:06.802468 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kplrj\" (UniqueName: \"kubernetes.io/projected/51a9ce2d-f224-449d-8988-950a60783ddb-kube-api-access-kplrj\") pod \"51a9ce2d-f224-449d-8988-950a60783ddb\" (UID: \"51a9ce2d-f224-449d-8988-950a60783ddb\") " Jan 26 17:21:06 crc kubenswrapper[4754]: I0126 17:21:06.803095 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/51a9ce2d-f224-449d-8988-950a60783ddb-inventory\") pod \"51a9ce2d-f224-449d-8988-950a60783ddb\" (UID: \"51a9ce2d-f224-449d-8988-950a60783ddb\") " Jan 26 17:21:06 crc kubenswrapper[4754]: I0126 17:21:06.808327 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/51a9ce2d-f224-449d-8988-950a60783ddb-kube-api-access-kplrj" (OuterVolumeSpecName: "kube-api-access-kplrj") pod "51a9ce2d-f224-449d-8988-950a60783ddb" (UID: "51a9ce2d-f224-449d-8988-950a60783ddb"). InnerVolumeSpecName "kube-api-access-kplrj". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:21:06 crc kubenswrapper[4754]: I0126 17:21:06.833181 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/51a9ce2d-f224-449d-8988-950a60783ddb-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "51a9ce2d-f224-449d-8988-950a60783ddb" (UID: "51a9ce2d-f224-449d-8988-950a60783ddb"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:21:06 crc kubenswrapper[4754]: I0126 17:21:06.833948 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/51a9ce2d-f224-449d-8988-950a60783ddb-inventory" (OuterVolumeSpecName: "inventory") pod "51a9ce2d-f224-449d-8988-950a60783ddb" (UID: "51a9ce2d-f224-449d-8988-950a60783ddb"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:21:06 crc kubenswrapper[4754]: I0126 17:21:06.905299 4754 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/51a9ce2d-f224-449d-8988-950a60783ddb-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Jan 26 17:21:06 crc kubenswrapper[4754]: I0126 17:21:06.905340 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kplrj\" (UniqueName: \"kubernetes.io/projected/51a9ce2d-f224-449d-8988-950a60783ddb-kube-api-access-kplrj\") on node \"crc\" DevicePath \"\"" Jan 26 17:21:06 crc kubenswrapper[4754]: I0126 17:21:06.905360 4754 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/51a9ce2d-f224-449d-8988-950a60783ddb-inventory\") on node \"crc\" DevicePath \"\"" Jan 26 17:21:07 crc kubenswrapper[4754]: I0126 17:21:07.129679 4754 patch_prober.go:28] interesting pod/machine-config-daemon-x65wv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 26 17:21:07 crc kubenswrapper[4754]: I0126 17:21:07.129944 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 26 17:21:07 crc kubenswrapper[4754]: I0126 17:21:07.361202 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-5d95m" event={"ID":"51a9ce2d-f224-449d-8988-950a60783ddb","Type":"ContainerDied","Data":"9af15fc57f567d98779b4557cc814580abee396e04f055eb4fd9a1745d2f71ce"} Jan 26 17:21:07 crc kubenswrapper[4754]: I0126 17:21:07.361257 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9af15fc57f567d98779b4557cc814580abee396e04f055eb4fd9a1745d2f71ce" Jan 26 17:21:07 crc kubenswrapper[4754]: I0126 17:21:07.361327 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-5d95m" Jan 26 17:21:07 crc kubenswrapper[4754]: I0126 17:21:07.457067 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-8vzzg"] Jan 26 17:21:07 crc kubenswrapper[4754]: E0126 17:21:07.457438 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8118a9a-872d-4d9f-9d87-f45973ac75ef" containerName="extract-utilities" Jan 26 17:21:07 crc kubenswrapper[4754]: I0126 17:21:07.457455 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8118a9a-872d-4d9f-9d87-f45973ac75ef" containerName="extract-utilities" Jan 26 17:21:07 crc kubenswrapper[4754]: E0126 17:21:07.457466 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="51a9ce2d-f224-449d-8988-950a60783ddb" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Jan 26 17:21:07 crc kubenswrapper[4754]: I0126 17:21:07.457474 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="51a9ce2d-f224-449d-8988-950a60783ddb" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Jan 26 17:21:07 crc kubenswrapper[4754]: E0126 17:21:07.457502 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8118a9a-872d-4d9f-9d87-f45973ac75ef" containerName="registry-server" Jan 26 17:21:07 crc kubenswrapper[4754]: I0126 17:21:07.457509 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8118a9a-872d-4d9f-9d87-f45973ac75ef" containerName="registry-server" Jan 26 17:21:07 crc kubenswrapper[4754]: E0126 17:21:07.457520 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8118a9a-872d-4d9f-9d87-f45973ac75ef" containerName="extract-content" Jan 26 17:21:07 crc kubenswrapper[4754]: I0126 17:21:07.457526 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8118a9a-872d-4d9f-9d87-f45973ac75ef" containerName="extract-content" Jan 26 17:21:07 crc kubenswrapper[4754]: I0126 17:21:07.457715 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="51a9ce2d-f224-449d-8988-950a60783ddb" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Jan 26 17:21:07 crc kubenswrapper[4754]: I0126 17:21:07.457745 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="e8118a9a-872d-4d9f-9d87-f45973ac75ef" containerName="registry-server" Jan 26 17:21:07 crc kubenswrapper[4754]: I0126 17:21:07.458369 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-8vzzg" Jan 26 17:21:07 crc kubenswrapper[4754]: I0126 17:21:07.460350 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Jan 26 17:21:07 crc kubenswrapper[4754]: I0126 17:21:07.461048 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Jan 26 17:21:07 crc kubenswrapper[4754]: I0126 17:21:07.461281 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-fz7z8" Jan 26 17:21:07 crc kubenswrapper[4754]: I0126 17:21:07.461451 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Jan 26 17:21:07 crc kubenswrapper[4754]: I0126 17:21:07.469865 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-8vzzg"] Jan 26 17:21:07 crc kubenswrapper[4754]: I0126 17:21:07.516211 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/9d659728-5ede-4206-8137-8a6a62c5385c-ssh-key-openstack-edpm-ipam\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-8vzzg\" (UID: \"9d659728-5ede-4206-8137-8a6a62c5385c\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-8vzzg" Jan 26 17:21:07 crc kubenswrapper[4754]: I0126 17:21:07.516278 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vz876\" (UniqueName: \"kubernetes.io/projected/9d659728-5ede-4206-8137-8a6a62c5385c-kube-api-access-vz876\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-8vzzg\" (UID: \"9d659728-5ede-4206-8137-8a6a62c5385c\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-8vzzg" Jan 26 17:21:07 crc kubenswrapper[4754]: I0126 17:21:07.516400 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9d659728-5ede-4206-8137-8a6a62c5385c-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-8vzzg\" (UID: \"9d659728-5ede-4206-8137-8a6a62c5385c\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-8vzzg" Jan 26 17:21:07 crc kubenswrapper[4754]: I0126 17:21:07.618344 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9d659728-5ede-4206-8137-8a6a62c5385c-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-8vzzg\" (UID: \"9d659728-5ede-4206-8137-8a6a62c5385c\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-8vzzg" Jan 26 17:21:07 crc kubenswrapper[4754]: I0126 17:21:07.618802 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/9d659728-5ede-4206-8137-8a6a62c5385c-ssh-key-openstack-edpm-ipam\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-8vzzg\" (UID: \"9d659728-5ede-4206-8137-8a6a62c5385c\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-8vzzg" Jan 26 17:21:07 crc kubenswrapper[4754]: I0126 17:21:07.618866 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vz876\" (UniqueName: \"kubernetes.io/projected/9d659728-5ede-4206-8137-8a6a62c5385c-kube-api-access-vz876\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-8vzzg\" (UID: \"9d659728-5ede-4206-8137-8a6a62c5385c\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-8vzzg" Jan 26 17:21:07 crc kubenswrapper[4754]: I0126 17:21:07.622455 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9d659728-5ede-4206-8137-8a6a62c5385c-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-8vzzg\" (UID: \"9d659728-5ede-4206-8137-8a6a62c5385c\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-8vzzg" Jan 26 17:21:07 crc kubenswrapper[4754]: I0126 17:21:07.623100 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/9d659728-5ede-4206-8137-8a6a62c5385c-ssh-key-openstack-edpm-ipam\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-8vzzg\" (UID: \"9d659728-5ede-4206-8137-8a6a62c5385c\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-8vzzg" Jan 26 17:21:07 crc kubenswrapper[4754]: I0126 17:21:07.634034 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vz876\" (UniqueName: \"kubernetes.io/projected/9d659728-5ede-4206-8137-8a6a62c5385c-kube-api-access-vz876\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-8vzzg\" (UID: \"9d659728-5ede-4206-8137-8a6a62c5385c\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-8vzzg" Jan 26 17:21:07 crc kubenswrapper[4754]: I0126 17:21:07.781059 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-8vzzg" Jan 26 17:21:08 crc kubenswrapper[4754]: I0126 17:21:08.275979 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-8vzzg"] Jan 26 17:21:08 crc kubenswrapper[4754]: I0126 17:21:08.370687 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-8vzzg" event={"ID":"9d659728-5ede-4206-8137-8a6a62c5385c","Type":"ContainerStarted","Data":"d7f264db9a5f0fb2e601595fcc637f45885301f9b601c38251bf50652a01a420"} Jan 26 17:21:11 crc kubenswrapper[4754]: I0126 17:21:11.394973 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-8vzzg" event={"ID":"9d659728-5ede-4206-8137-8a6a62c5385c","Type":"ContainerStarted","Data":"474149cfe8ee5dac8935954a1676e80eba95c24e637499f0264876245483db6c"} Jan 26 17:21:11 crc kubenswrapper[4754]: I0126 17:21:11.417766 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-8vzzg" podStartSLOduration=2.227228155 podStartE2EDuration="4.4177451s" podCreationTimestamp="2026-01-26 17:21:07 +0000 UTC" firstStartedPulling="2026-01-26 17:21:08.280274753 +0000 UTC m=+2034.804455187" lastFinishedPulling="2026-01-26 17:21:10.470791698 +0000 UTC m=+2036.994972132" observedRunningTime="2026-01-26 17:21:11.409619503 +0000 UTC m=+2037.933799947" watchObservedRunningTime="2026-01-26 17:21:11.4177451 +0000 UTC m=+2037.941925534" Jan 26 17:21:14 crc kubenswrapper[4754]: I0126 17:21:14.033112 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-55c6p"] Jan 26 17:21:14 crc kubenswrapper[4754]: I0126 17:21:14.045527 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-55c6p"] Jan 26 17:21:15 crc kubenswrapper[4754]: I0126 17:21:15.777881 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d40200db-b145-4a60-b337-82ae5032cc51" path="/var/lib/kubelet/pods/d40200db-b145-4a60-b337-82ae5032cc51/volumes" Jan 26 17:21:16 crc kubenswrapper[4754]: I0126 17:21:16.434494 4754 generic.go:334] "Generic (PLEG): container finished" podID="9d659728-5ede-4206-8137-8a6a62c5385c" containerID="474149cfe8ee5dac8935954a1676e80eba95c24e637499f0264876245483db6c" exitCode=0 Jan 26 17:21:16 crc kubenswrapper[4754]: I0126 17:21:16.434593 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-8vzzg" event={"ID":"9d659728-5ede-4206-8137-8a6a62c5385c","Type":"ContainerDied","Data":"474149cfe8ee5dac8935954a1676e80eba95c24e637499f0264876245483db6c"} Jan 26 17:21:17 crc kubenswrapper[4754]: I0126 17:21:17.838604 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-8vzzg" Jan 26 17:21:17 crc kubenswrapper[4754]: I0126 17:21:17.905116 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/9d659728-5ede-4206-8137-8a6a62c5385c-ssh-key-openstack-edpm-ipam\") pod \"9d659728-5ede-4206-8137-8a6a62c5385c\" (UID: \"9d659728-5ede-4206-8137-8a6a62c5385c\") " Jan 26 17:21:17 crc kubenswrapper[4754]: I0126 17:21:17.905178 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9d659728-5ede-4206-8137-8a6a62c5385c-inventory\") pod \"9d659728-5ede-4206-8137-8a6a62c5385c\" (UID: \"9d659728-5ede-4206-8137-8a6a62c5385c\") " Jan 26 17:21:17 crc kubenswrapper[4754]: I0126 17:21:17.905325 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vz876\" (UniqueName: \"kubernetes.io/projected/9d659728-5ede-4206-8137-8a6a62c5385c-kube-api-access-vz876\") pod \"9d659728-5ede-4206-8137-8a6a62c5385c\" (UID: \"9d659728-5ede-4206-8137-8a6a62c5385c\") " Jan 26 17:21:17 crc kubenswrapper[4754]: I0126 17:21:17.911373 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d659728-5ede-4206-8137-8a6a62c5385c-kube-api-access-vz876" (OuterVolumeSpecName: "kube-api-access-vz876") pod "9d659728-5ede-4206-8137-8a6a62c5385c" (UID: "9d659728-5ede-4206-8137-8a6a62c5385c"). InnerVolumeSpecName "kube-api-access-vz876". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:21:17 crc kubenswrapper[4754]: I0126 17:21:17.936349 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d659728-5ede-4206-8137-8a6a62c5385c-inventory" (OuterVolumeSpecName: "inventory") pod "9d659728-5ede-4206-8137-8a6a62c5385c" (UID: "9d659728-5ede-4206-8137-8a6a62c5385c"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:21:17 crc kubenswrapper[4754]: I0126 17:21:17.936877 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d659728-5ede-4206-8137-8a6a62c5385c-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "9d659728-5ede-4206-8137-8a6a62c5385c" (UID: "9d659728-5ede-4206-8137-8a6a62c5385c"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:21:18 crc kubenswrapper[4754]: I0126 17:21:18.008263 4754 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9d659728-5ede-4206-8137-8a6a62c5385c-inventory\") on node \"crc\" DevicePath \"\"" Jan 26 17:21:18 crc kubenswrapper[4754]: I0126 17:21:18.008304 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vz876\" (UniqueName: \"kubernetes.io/projected/9d659728-5ede-4206-8137-8a6a62c5385c-kube-api-access-vz876\") on node \"crc\" DevicePath \"\"" Jan 26 17:21:18 crc kubenswrapper[4754]: I0126 17:21:18.008320 4754 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/9d659728-5ede-4206-8137-8a6a62c5385c-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Jan 26 17:21:18 crc kubenswrapper[4754]: I0126 17:21:18.451927 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-8vzzg" event={"ID":"9d659728-5ede-4206-8137-8a6a62c5385c","Type":"ContainerDied","Data":"d7f264db9a5f0fb2e601595fcc637f45885301f9b601c38251bf50652a01a420"} Jan 26 17:21:18 crc kubenswrapper[4754]: I0126 17:21:18.451965 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d7f264db9a5f0fb2e601595fcc637f45885301f9b601c38251bf50652a01a420" Jan 26 17:21:18 crc kubenswrapper[4754]: I0126 17:21:18.451981 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-8vzzg" Jan 26 17:21:18 crc kubenswrapper[4754]: I0126 17:21:18.515345 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-r92d9"] Jan 26 17:21:18 crc kubenswrapper[4754]: E0126 17:21:18.515821 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d659728-5ede-4206-8137-8a6a62c5385c" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Jan 26 17:21:18 crc kubenswrapper[4754]: I0126 17:21:18.515847 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d659728-5ede-4206-8137-8a6a62c5385c" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Jan 26 17:21:18 crc kubenswrapper[4754]: I0126 17:21:18.516081 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="9d659728-5ede-4206-8137-8a6a62c5385c" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Jan 26 17:21:18 crc kubenswrapper[4754]: I0126 17:21:18.516842 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-r92d9" Jan 26 17:21:18 crc kubenswrapper[4754]: I0126 17:21:18.519769 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Jan 26 17:21:18 crc kubenswrapper[4754]: I0126 17:21:18.519773 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Jan 26 17:21:18 crc kubenswrapper[4754]: I0126 17:21:18.520890 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Jan 26 17:21:18 crc kubenswrapper[4754]: I0126 17:21:18.521467 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-fz7z8" Jan 26 17:21:18 crc kubenswrapper[4754]: I0126 17:21:18.525136 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-r92d9"] Jan 26 17:21:18 crc kubenswrapper[4754]: I0126 17:21:18.617907 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/9bfc59f9-f503-4fec-9a31-240819fc3a52-ssh-key-openstack-edpm-ipam\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-r92d9\" (UID: \"9bfc59f9-f503-4fec-9a31-240819fc3a52\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-r92d9" Jan 26 17:21:18 crc kubenswrapper[4754]: I0126 17:21:18.618178 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9bfc59f9-f503-4fec-9a31-240819fc3a52-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-r92d9\" (UID: \"9bfc59f9-f503-4fec-9a31-240819fc3a52\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-r92d9" Jan 26 17:21:18 crc kubenswrapper[4754]: I0126 17:21:18.618336 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wn75n\" (UniqueName: \"kubernetes.io/projected/9bfc59f9-f503-4fec-9a31-240819fc3a52-kube-api-access-wn75n\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-r92d9\" (UID: \"9bfc59f9-f503-4fec-9a31-240819fc3a52\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-r92d9" Jan 26 17:21:18 crc kubenswrapper[4754]: I0126 17:21:18.719861 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/9bfc59f9-f503-4fec-9a31-240819fc3a52-ssh-key-openstack-edpm-ipam\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-r92d9\" (UID: \"9bfc59f9-f503-4fec-9a31-240819fc3a52\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-r92d9" Jan 26 17:21:18 crc kubenswrapper[4754]: I0126 17:21:18.719924 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9bfc59f9-f503-4fec-9a31-240819fc3a52-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-r92d9\" (UID: \"9bfc59f9-f503-4fec-9a31-240819fc3a52\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-r92d9" Jan 26 17:21:18 crc kubenswrapper[4754]: I0126 17:21:18.719969 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wn75n\" (UniqueName: \"kubernetes.io/projected/9bfc59f9-f503-4fec-9a31-240819fc3a52-kube-api-access-wn75n\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-r92d9\" (UID: \"9bfc59f9-f503-4fec-9a31-240819fc3a52\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-r92d9" Jan 26 17:21:18 crc kubenswrapper[4754]: I0126 17:21:18.724495 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9bfc59f9-f503-4fec-9a31-240819fc3a52-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-r92d9\" (UID: \"9bfc59f9-f503-4fec-9a31-240819fc3a52\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-r92d9" Jan 26 17:21:18 crc kubenswrapper[4754]: I0126 17:21:18.725760 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/9bfc59f9-f503-4fec-9a31-240819fc3a52-ssh-key-openstack-edpm-ipam\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-r92d9\" (UID: \"9bfc59f9-f503-4fec-9a31-240819fc3a52\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-r92d9" Jan 26 17:21:18 crc kubenswrapper[4754]: I0126 17:21:18.737302 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wn75n\" (UniqueName: \"kubernetes.io/projected/9bfc59f9-f503-4fec-9a31-240819fc3a52-kube-api-access-wn75n\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-r92d9\" (UID: \"9bfc59f9-f503-4fec-9a31-240819fc3a52\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-r92d9" Jan 26 17:21:18 crc kubenswrapper[4754]: I0126 17:21:18.835376 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-r92d9" Jan 26 17:21:19 crc kubenswrapper[4754]: I0126 17:21:19.485361 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-r92d9"] Jan 26 17:21:20 crc kubenswrapper[4754]: I0126 17:21:20.471329 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-r92d9" event={"ID":"9bfc59f9-f503-4fec-9a31-240819fc3a52","Type":"ContainerStarted","Data":"2163a77b4469ed7c768da44aff1316e81a8597c833c96692357158653fd66e5e"} Jan 26 17:21:20 crc kubenswrapper[4754]: I0126 17:21:20.472035 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-r92d9" event={"ID":"9bfc59f9-f503-4fec-9a31-240819fc3a52","Type":"ContainerStarted","Data":"977c724134dd94f45d3bbda9746f90ecea4293fa5eaf9dd957f79066060da912"} Jan 26 17:21:20 crc kubenswrapper[4754]: I0126 17:21:20.495081 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-r92d9" podStartSLOduration=2.080757714 podStartE2EDuration="2.495060203s" podCreationTimestamp="2026-01-26 17:21:18 +0000 UTC" firstStartedPulling="2026-01-26 17:21:19.483803976 +0000 UTC m=+2046.007984410" lastFinishedPulling="2026-01-26 17:21:19.898106465 +0000 UTC m=+2046.422286899" observedRunningTime="2026-01-26 17:21:20.485391484 +0000 UTC m=+2047.009571938" watchObservedRunningTime="2026-01-26 17:21:20.495060203 +0000 UTC m=+2047.019240647" Jan 26 17:21:20 crc kubenswrapper[4754]: I0126 17:21:20.663827 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-qdcgq"] Jan 26 17:21:20 crc kubenswrapper[4754]: I0126 17:21:20.665593 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qdcgq" Jan 26 17:21:20 crc kubenswrapper[4754]: I0126 17:21:20.686789 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-qdcgq"] Jan 26 17:21:20 crc kubenswrapper[4754]: I0126 17:21:20.754791 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/74000075-a282-49a3-ab32-94300363c549-utilities\") pod \"redhat-operators-qdcgq\" (UID: \"74000075-a282-49a3-ab32-94300363c549\") " pod="openshift-marketplace/redhat-operators-qdcgq" Jan 26 17:21:20 crc kubenswrapper[4754]: I0126 17:21:20.754859 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/74000075-a282-49a3-ab32-94300363c549-catalog-content\") pod \"redhat-operators-qdcgq\" (UID: \"74000075-a282-49a3-ab32-94300363c549\") " pod="openshift-marketplace/redhat-operators-qdcgq" Jan 26 17:21:20 crc kubenswrapper[4754]: I0126 17:21:20.754908 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-phcfr\" (UniqueName: \"kubernetes.io/projected/74000075-a282-49a3-ab32-94300363c549-kube-api-access-phcfr\") pod \"redhat-operators-qdcgq\" (UID: \"74000075-a282-49a3-ab32-94300363c549\") " pod="openshift-marketplace/redhat-operators-qdcgq" Jan 26 17:21:20 crc kubenswrapper[4754]: I0126 17:21:20.857003 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/74000075-a282-49a3-ab32-94300363c549-utilities\") pod \"redhat-operators-qdcgq\" (UID: \"74000075-a282-49a3-ab32-94300363c549\") " pod="openshift-marketplace/redhat-operators-qdcgq" Jan 26 17:21:20 crc kubenswrapper[4754]: I0126 17:21:20.857072 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/74000075-a282-49a3-ab32-94300363c549-catalog-content\") pod \"redhat-operators-qdcgq\" (UID: \"74000075-a282-49a3-ab32-94300363c549\") " pod="openshift-marketplace/redhat-operators-qdcgq" Jan 26 17:21:20 crc kubenswrapper[4754]: I0126 17:21:20.857111 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-phcfr\" (UniqueName: \"kubernetes.io/projected/74000075-a282-49a3-ab32-94300363c549-kube-api-access-phcfr\") pod \"redhat-operators-qdcgq\" (UID: \"74000075-a282-49a3-ab32-94300363c549\") " pod="openshift-marketplace/redhat-operators-qdcgq" Jan 26 17:21:20 crc kubenswrapper[4754]: I0126 17:21:20.857656 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/74000075-a282-49a3-ab32-94300363c549-utilities\") pod \"redhat-operators-qdcgq\" (UID: \"74000075-a282-49a3-ab32-94300363c549\") " pod="openshift-marketplace/redhat-operators-qdcgq" Jan 26 17:21:20 crc kubenswrapper[4754]: I0126 17:21:20.857779 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/74000075-a282-49a3-ab32-94300363c549-catalog-content\") pod \"redhat-operators-qdcgq\" (UID: \"74000075-a282-49a3-ab32-94300363c549\") " pod="openshift-marketplace/redhat-operators-qdcgq" Jan 26 17:21:20 crc kubenswrapper[4754]: I0126 17:21:20.877049 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-phcfr\" (UniqueName: \"kubernetes.io/projected/74000075-a282-49a3-ab32-94300363c549-kube-api-access-phcfr\") pod \"redhat-operators-qdcgq\" (UID: \"74000075-a282-49a3-ab32-94300363c549\") " pod="openshift-marketplace/redhat-operators-qdcgq" Jan 26 17:21:20 crc kubenswrapper[4754]: I0126 17:21:20.988379 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qdcgq" Jan 26 17:21:21 crc kubenswrapper[4754]: I0126 17:21:21.436443 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-qdcgq"] Jan 26 17:21:21 crc kubenswrapper[4754]: I0126 17:21:21.486390 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qdcgq" event={"ID":"74000075-a282-49a3-ab32-94300363c549","Type":"ContainerStarted","Data":"a9bbd2ff2ff2d13caab9937b3404cf91a14eb68aa0bf66daabcf13f5affa81b6"} Jan 26 17:21:22 crc kubenswrapper[4754]: I0126 17:21:22.495761 4754 generic.go:334] "Generic (PLEG): container finished" podID="74000075-a282-49a3-ab32-94300363c549" containerID="a76706cbb28536204294845e4727247a23d3ff29ea9bcf28659106e9cf288ecd" exitCode=0 Jan 26 17:21:22 crc kubenswrapper[4754]: I0126 17:21:22.495902 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qdcgq" event={"ID":"74000075-a282-49a3-ab32-94300363c549","Type":"ContainerDied","Data":"a76706cbb28536204294845e4727247a23d3ff29ea9bcf28659106e9cf288ecd"} Jan 26 17:21:23 crc kubenswrapper[4754]: I0126 17:21:23.505878 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qdcgq" event={"ID":"74000075-a282-49a3-ab32-94300363c549","Type":"ContainerStarted","Data":"50d6a87f4792ff9e786fbb218f05cbd6a3d189f57abcabb810a310b9c3ccb407"} Jan 26 17:21:25 crc kubenswrapper[4754]: I0126 17:21:25.523150 4754 generic.go:334] "Generic (PLEG): container finished" podID="74000075-a282-49a3-ab32-94300363c549" containerID="50d6a87f4792ff9e786fbb218f05cbd6a3d189f57abcabb810a310b9c3ccb407" exitCode=0 Jan 26 17:21:25 crc kubenswrapper[4754]: I0126 17:21:25.523243 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qdcgq" event={"ID":"74000075-a282-49a3-ab32-94300363c549","Type":"ContainerDied","Data":"50d6a87f4792ff9e786fbb218f05cbd6a3d189f57abcabb810a310b9c3ccb407"} Jan 26 17:21:26 crc kubenswrapper[4754]: I0126 17:21:26.534564 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qdcgq" event={"ID":"74000075-a282-49a3-ab32-94300363c549","Type":"ContainerStarted","Data":"eee2728661c359eaaa1b2ff4aa81e58c32425511a81af1c66c4ae004a9b79064"} Jan 26 17:21:26 crc kubenswrapper[4754]: I0126 17:21:26.557204 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-qdcgq" podStartSLOduration=3.05198542 podStartE2EDuration="6.557190145s" podCreationTimestamp="2026-01-26 17:21:20 +0000 UTC" firstStartedPulling="2026-01-26 17:21:22.497207689 +0000 UTC m=+2049.021388123" lastFinishedPulling="2026-01-26 17:21:26.002412414 +0000 UTC m=+2052.526592848" observedRunningTime="2026-01-26 17:21:26.554236342 +0000 UTC m=+2053.078416786" watchObservedRunningTime="2026-01-26 17:21:26.557190145 +0000 UTC m=+2053.081370579" Jan 26 17:21:26 crc kubenswrapper[4754]: I0126 17:21:26.981307 4754 scope.go:117] "RemoveContainer" containerID="62265a25a1e473fa628a59ad5be6830d0677bee3af7afe2a0e8f64ed52f241f3" Jan 26 17:21:27 crc kubenswrapper[4754]: I0126 17:21:27.031019 4754 scope.go:117] "RemoveContainer" containerID="f3fe6e1c3d0a565d5c2f44c4be05b0cb2ef674bd9aca961cfe8edcb5ff320e24" Jan 26 17:21:27 crc kubenswrapper[4754]: I0126 17:21:27.086590 4754 scope.go:117] "RemoveContainer" containerID="81c700b92383b930ceaf735cdacfcdd6280d44d2817b951b810ae6a363a3efb3" Jan 26 17:21:30 crc kubenswrapper[4754]: I0126 17:21:30.989983 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-qdcgq" Jan 26 17:21:30 crc kubenswrapper[4754]: I0126 17:21:30.990286 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-qdcgq" Jan 26 17:21:32 crc kubenswrapper[4754]: I0126 17:21:32.037188 4754 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-qdcgq" podUID="74000075-a282-49a3-ab32-94300363c549" containerName="registry-server" probeResult="failure" output=< Jan 26 17:21:32 crc kubenswrapper[4754]: timeout: failed to connect service ":50051" within 1s Jan 26 17:21:32 crc kubenswrapper[4754]: > Jan 26 17:21:37 crc kubenswrapper[4754]: I0126 17:21:37.129049 4754 patch_prober.go:28] interesting pod/machine-config-daemon-x65wv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 26 17:21:37 crc kubenswrapper[4754]: I0126 17:21:37.129632 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 26 17:21:37 crc kubenswrapper[4754]: I0126 17:21:37.129727 4754 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" Jan 26 17:21:37 crc kubenswrapper[4754]: I0126 17:21:37.130470 4754 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"1f2fc5d329d9c26d0c14a843ed658e96c72f72ea44b0a9c9d8e76e987f4603de"} pod="openshift-machine-config-operator/machine-config-daemon-x65wv" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 26 17:21:37 crc kubenswrapper[4754]: I0126 17:21:37.130536 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" containerName="machine-config-daemon" containerID="cri-o://1f2fc5d329d9c26d0c14a843ed658e96c72f72ea44b0a9c9d8e76e987f4603de" gracePeriod=600 Jan 26 17:21:37 crc kubenswrapper[4754]: I0126 17:21:37.634167 4754 generic.go:334] "Generic (PLEG): container finished" podID="8c3718a4-f354-4284-92e0-fdfb45a692bd" containerID="1f2fc5d329d9c26d0c14a843ed658e96c72f72ea44b0a9c9d8e76e987f4603de" exitCode=0 Jan 26 17:21:37 crc kubenswrapper[4754]: I0126 17:21:37.634206 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" event={"ID":"8c3718a4-f354-4284-92e0-fdfb45a692bd","Type":"ContainerDied","Data":"1f2fc5d329d9c26d0c14a843ed658e96c72f72ea44b0a9c9d8e76e987f4603de"} Jan 26 17:21:37 crc kubenswrapper[4754]: I0126 17:21:37.634619 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" event={"ID":"8c3718a4-f354-4284-92e0-fdfb45a692bd","Type":"ContainerStarted","Data":"00e15c82b01f7eb3c11968e396c0748604b39b71891170fde13a876ab0a851b3"} Jan 26 17:21:37 crc kubenswrapper[4754]: I0126 17:21:37.635165 4754 scope.go:117] "RemoveContainer" containerID="19eac0e87afaeef93088a8556aa836389e2ebff93a96b482df571cb7b852f0b7" Jan 26 17:21:41 crc kubenswrapper[4754]: I0126 17:21:41.043582 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-qdcgq" Jan 26 17:21:41 crc kubenswrapper[4754]: I0126 17:21:41.103348 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-qdcgq" Jan 26 17:21:41 crc kubenswrapper[4754]: I0126 17:21:41.282700 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-qdcgq"] Jan 26 17:21:42 crc kubenswrapper[4754]: I0126 17:21:42.679522 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-qdcgq" podUID="74000075-a282-49a3-ab32-94300363c549" containerName="registry-server" containerID="cri-o://eee2728661c359eaaa1b2ff4aa81e58c32425511a81af1c66c4ae004a9b79064" gracePeriod=2 Jan 26 17:21:43 crc kubenswrapper[4754]: I0126 17:21:43.617501 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qdcgq" Jan 26 17:21:43 crc kubenswrapper[4754]: I0126 17:21:43.683065 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/74000075-a282-49a3-ab32-94300363c549-catalog-content\") pod \"74000075-a282-49a3-ab32-94300363c549\" (UID: \"74000075-a282-49a3-ab32-94300363c549\") " Jan 26 17:21:43 crc kubenswrapper[4754]: I0126 17:21:43.683120 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-phcfr\" (UniqueName: \"kubernetes.io/projected/74000075-a282-49a3-ab32-94300363c549-kube-api-access-phcfr\") pod \"74000075-a282-49a3-ab32-94300363c549\" (UID: \"74000075-a282-49a3-ab32-94300363c549\") " Jan 26 17:21:43 crc kubenswrapper[4754]: I0126 17:21:43.683163 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/74000075-a282-49a3-ab32-94300363c549-utilities\") pod \"74000075-a282-49a3-ab32-94300363c549\" (UID: \"74000075-a282-49a3-ab32-94300363c549\") " Jan 26 17:21:43 crc kubenswrapper[4754]: I0126 17:21:43.684403 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/74000075-a282-49a3-ab32-94300363c549-utilities" (OuterVolumeSpecName: "utilities") pod "74000075-a282-49a3-ab32-94300363c549" (UID: "74000075-a282-49a3-ab32-94300363c549"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:21:43 crc kubenswrapper[4754]: I0126 17:21:43.690019 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/74000075-a282-49a3-ab32-94300363c549-kube-api-access-phcfr" (OuterVolumeSpecName: "kube-api-access-phcfr") pod "74000075-a282-49a3-ab32-94300363c549" (UID: "74000075-a282-49a3-ab32-94300363c549"). InnerVolumeSpecName "kube-api-access-phcfr". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:21:43 crc kubenswrapper[4754]: I0126 17:21:43.693903 4754 generic.go:334] "Generic (PLEG): container finished" podID="74000075-a282-49a3-ab32-94300363c549" containerID="eee2728661c359eaaa1b2ff4aa81e58c32425511a81af1c66c4ae004a9b79064" exitCode=0 Jan 26 17:21:43 crc kubenswrapper[4754]: I0126 17:21:43.694010 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qdcgq" event={"ID":"74000075-a282-49a3-ab32-94300363c549","Type":"ContainerDied","Data":"eee2728661c359eaaa1b2ff4aa81e58c32425511a81af1c66c4ae004a9b79064"} Jan 26 17:21:43 crc kubenswrapper[4754]: I0126 17:21:43.694064 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qdcgq" Jan 26 17:21:43 crc kubenswrapper[4754]: I0126 17:21:43.694310 4754 scope.go:117] "RemoveContainer" containerID="eee2728661c359eaaa1b2ff4aa81e58c32425511a81af1c66c4ae004a9b79064" Jan 26 17:21:43 crc kubenswrapper[4754]: I0126 17:21:43.694255 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qdcgq" event={"ID":"74000075-a282-49a3-ab32-94300363c549","Type":"ContainerDied","Data":"a9bbd2ff2ff2d13caab9937b3404cf91a14eb68aa0bf66daabcf13f5affa81b6"} Jan 26 17:21:43 crc kubenswrapper[4754]: I0126 17:21:43.748861 4754 scope.go:117] "RemoveContainer" containerID="50d6a87f4792ff9e786fbb218f05cbd6a3d189f57abcabb810a310b9c3ccb407" Jan 26 17:21:43 crc kubenswrapper[4754]: I0126 17:21:43.774354 4754 scope.go:117] "RemoveContainer" containerID="a76706cbb28536204294845e4727247a23d3ff29ea9bcf28659106e9cf288ecd" Jan 26 17:21:43 crc kubenswrapper[4754]: I0126 17:21:43.786281 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-phcfr\" (UniqueName: \"kubernetes.io/projected/74000075-a282-49a3-ab32-94300363c549-kube-api-access-phcfr\") on node \"crc\" DevicePath \"\"" Jan 26 17:21:43 crc kubenswrapper[4754]: I0126 17:21:43.786329 4754 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/74000075-a282-49a3-ab32-94300363c549-utilities\") on node \"crc\" DevicePath \"\"" Jan 26 17:21:43 crc kubenswrapper[4754]: I0126 17:21:43.816619 4754 scope.go:117] "RemoveContainer" containerID="eee2728661c359eaaa1b2ff4aa81e58c32425511a81af1c66c4ae004a9b79064" Jan 26 17:21:43 crc kubenswrapper[4754]: E0126 17:21:43.817253 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eee2728661c359eaaa1b2ff4aa81e58c32425511a81af1c66c4ae004a9b79064\": container with ID starting with eee2728661c359eaaa1b2ff4aa81e58c32425511a81af1c66c4ae004a9b79064 not found: ID does not exist" containerID="eee2728661c359eaaa1b2ff4aa81e58c32425511a81af1c66c4ae004a9b79064" Jan 26 17:21:43 crc kubenswrapper[4754]: I0126 17:21:43.817289 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eee2728661c359eaaa1b2ff4aa81e58c32425511a81af1c66c4ae004a9b79064"} err="failed to get container status \"eee2728661c359eaaa1b2ff4aa81e58c32425511a81af1c66c4ae004a9b79064\": rpc error: code = NotFound desc = could not find container \"eee2728661c359eaaa1b2ff4aa81e58c32425511a81af1c66c4ae004a9b79064\": container with ID starting with eee2728661c359eaaa1b2ff4aa81e58c32425511a81af1c66c4ae004a9b79064 not found: ID does not exist" Jan 26 17:21:43 crc kubenswrapper[4754]: I0126 17:21:43.817311 4754 scope.go:117] "RemoveContainer" containerID="50d6a87f4792ff9e786fbb218f05cbd6a3d189f57abcabb810a310b9c3ccb407" Jan 26 17:21:43 crc kubenswrapper[4754]: E0126 17:21:43.817623 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"50d6a87f4792ff9e786fbb218f05cbd6a3d189f57abcabb810a310b9c3ccb407\": container with ID starting with 50d6a87f4792ff9e786fbb218f05cbd6a3d189f57abcabb810a310b9c3ccb407 not found: ID does not exist" containerID="50d6a87f4792ff9e786fbb218f05cbd6a3d189f57abcabb810a310b9c3ccb407" Jan 26 17:21:43 crc kubenswrapper[4754]: I0126 17:21:43.817692 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"50d6a87f4792ff9e786fbb218f05cbd6a3d189f57abcabb810a310b9c3ccb407"} err="failed to get container status \"50d6a87f4792ff9e786fbb218f05cbd6a3d189f57abcabb810a310b9c3ccb407\": rpc error: code = NotFound desc = could not find container \"50d6a87f4792ff9e786fbb218f05cbd6a3d189f57abcabb810a310b9c3ccb407\": container with ID starting with 50d6a87f4792ff9e786fbb218f05cbd6a3d189f57abcabb810a310b9c3ccb407 not found: ID does not exist" Jan 26 17:21:43 crc kubenswrapper[4754]: I0126 17:21:43.817730 4754 scope.go:117] "RemoveContainer" containerID="a76706cbb28536204294845e4727247a23d3ff29ea9bcf28659106e9cf288ecd" Jan 26 17:21:43 crc kubenswrapper[4754]: E0126 17:21:43.818308 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a76706cbb28536204294845e4727247a23d3ff29ea9bcf28659106e9cf288ecd\": container with ID starting with a76706cbb28536204294845e4727247a23d3ff29ea9bcf28659106e9cf288ecd not found: ID does not exist" containerID="a76706cbb28536204294845e4727247a23d3ff29ea9bcf28659106e9cf288ecd" Jan 26 17:21:43 crc kubenswrapper[4754]: I0126 17:21:43.818332 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a76706cbb28536204294845e4727247a23d3ff29ea9bcf28659106e9cf288ecd"} err="failed to get container status \"a76706cbb28536204294845e4727247a23d3ff29ea9bcf28659106e9cf288ecd\": rpc error: code = NotFound desc = could not find container \"a76706cbb28536204294845e4727247a23d3ff29ea9bcf28659106e9cf288ecd\": container with ID starting with a76706cbb28536204294845e4727247a23d3ff29ea9bcf28659106e9cf288ecd not found: ID does not exist" Jan 26 17:21:43 crc kubenswrapper[4754]: I0126 17:21:43.829349 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/74000075-a282-49a3-ab32-94300363c549-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "74000075-a282-49a3-ab32-94300363c549" (UID: "74000075-a282-49a3-ab32-94300363c549"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:21:43 crc kubenswrapper[4754]: I0126 17:21:43.888190 4754 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/74000075-a282-49a3-ab32-94300363c549-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 26 17:21:44 crc kubenswrapper[4754]: I0126 17:21:44.040733 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-qdcgq"] Jan 26 17:21:44 crc kubenswrapper[4754]: I0126 17:21:44.051040 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-qdcgq"] Jan 26 17:21:45 crc kubenswrapper[4754]: I0126 17:21:45.779693 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="74000075-a282-49a3-ab32-94300363c549" path="/var/lib/kubelet/pods/74000075-a282-49a3-ab32-94300363c549/volumes" Jan 26 17:21:51 crc kubenswrapper[4754]: I0126 17:21:51.047104 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-bfp9x"] Jan 26 17:21:51 crc kubenswrapper[4754]: I0126 17:21:51.054073 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-bfp9x"] Jan 26 17:21:51 crc kubenswrapper[4754]: I0126 17:21:51.778835 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7f22d2aa-d1fd-4a87-b0bd-9e3fb661d254" path="/var/lib/kubelet/pods/7f22d2aa-d1fd-4a87-b0bd-9e3fb661d254/volumes" Jan 26 17:21:58 crc kubenswrapper[4754]: I0126 17:21:58.835777 4754 generic.go:334] "Generic (PLEG): container finished" podID="9bfc59f9-f503-4fec-9a31-240819fc3a52" containerID="2163a77b4469ed7c768da44aff1316e81a8597c833c96692357158653fd66e5e" exitCode=0 Jan 26 17:21:58 crc kubenswrapper[4754]: I0126 17:21:58.835873 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-r92d9" event={"ID":"9bfc59f9-f503-4fec-9a31-240819fc3a52","Type":"ContainerDied","Data":"2163a77b4469ed7c768da44aff1316e81a8597c833c96692357158653fd66e5e"} Jan 26 17:22:00 crc kubenswrapper[4754]: I0126 17:22:00.232914 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-r92d9" Jan 26 17:22:00 crc kubenswrapper[4754]: I0126 17:22:00.385834 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wn75n\" (UniqueName: \"kubernetes.io/projected/9bfc59f9-f503-4fec-9a31-240819fc3a52-kube-api-access-wn75n\") pod \"9bfc59f9-f503-4fec-9a31-240819fc3a52\" (UID: \"9bfc59f9-f503-4fec-9a31-240819fc3a52\") " Jan 26 17:22:00 crc kubenswrapper[4754]: I0126 17:22:00.386084 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/9bfc59f9-f503-4fec-9a31-240819fc3a52-ssh-key-openstack-edpm-ipam\") pod \"9bfc59f9-f503-4fec-9a31-240819fc3a52\" (UID: \"9bfc59f9-f503-4fec-9a31-240819fc3a52\") " Jan 26 17:22:00 crc kubenswrapper[4754]: I0126 17:22:00.386769 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9bfc59f9-f503-4fec-9a31-240819fc3a52-inventory\") pod \"9bfc59f9-f503-4fec-9a31-240819fc3a52\" (UID: \"9bfc59f9-f503-4fec-9a31-240819fc3a52\") " Jan 26 17:22:00 crc kubenswrapper[4754]: I0126 17:22:00.391362 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9bfc59f9-f503-4fec-9a31-240819fc3a52-kube-api-access-wn75n" (OuterVolumeSpecName: "kube-api-access-wn75n") pod "9bfc59f9-f503-4fec-9a31-240819fc3a52" (UID: "9bfc59f9-f503-4fec-9a31-240819fc3a52"). InnerVolumeSpecName "kube-api-access-wn75n". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:22:00 crc kubenswrapper[4754]: I0126 17:22:00.410851 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9bfc59f9-f503-4fec-9a31-240819fc3a52-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "9bfc59f9-f503-4fec-9a31-240819fc3a52" (UID: "9bfc59f9-f503-4fec-9a31-240819fc3a52"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:22:00 crc kubenswrapper[4754]: I0126 17:22:00.411233 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9bfc59f9-f503-4fec-9a31-240819fc3a52-inventory" (OuterVolumeSpecName: "inventory") pod "9bfc59f9-f503-4fec-9a31-240819fc3a52" (UID: "9bfc59f9-f503-4fec-9a31-240819fc3a52"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:22:00 crc kubenswrapper[4754]: I0126 17:22:00.487840 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wn75n\" (UniqueName: \"kubernetes.io/projected/9bfc59f9-f503-4fec-9a31-240819fc3a52-kube-api-access-wn75n\") on node \"crc\" DevicePath \"\"" Jan 26 17:22:00 crc kubenswrapper[4754]: I0126 17:22:00.487876 4754 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/9bfc59f9-f503-4fec-9a31-240819fc3a52-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Jan 26 17:22:00 crc kubenswrapper[4754]: I0126 17:22:00.487886 4754 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9bfc59f9-f503-4fec-9a31-240819fc3a52-inventory\") on node \"crc\" DevicePath \"\"" Jan 26 17:22:00 crc kubenswrapper[4754]: I0126 17:22:00.857813 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-r92d9" event={"ID":"9bfc59f9-f503-4fec-9a31-240819fc3a52","Type":"ContainerDied","Data":"977c724134dd94f45d3bbda9746f90ecea4293fa5eaf9dd957f79066060da912"} Jan 26 17:22:00 crc kubenswrapper[4754]: I0126 17:22:00.857872 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="977c724134dd94f45d3bbda9746f90ecea4293fa5eaf9dd957f79066060da912" Jan 26 17:22:00 crc kubenswrapper[4754]: I0126 17:22:00.857921 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-r92d9" Jan 26 17:22:00 crc kubenswrapper[4754]: I0126 17:22:00.944135 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-v9nvp"] Jan 26 17:22:00 crc kubenswrapper[4754]: E0126 17:22:00.944901 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74000075-a282-49a3-ab32-94300363c549" containerName="extract-utilities" Jan 26 17:22:00 crc kubenswrapper[4754]: I0126 17:22:00.944923 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="74000075-a282-49a3-ab32-94300363c549" containerName="extract-utilities" Jan 26 17:22:00 crc kubenswrapper[4754]: E0126 17:22:00.944946 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9bfc59f9-f503-4fec-9a31-240819fc3a52" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Jan 26 17:22:00 crc kubenswrapper[4754]: I0126 17:22:00.944958 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="9bfc59f9-f503-4fec-9a31-240819fc3a52" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Jan 26 17:22:00 crc kubenswrapper[4754]: E0126 17:22:00.944969 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74000075-a282-49a3-ab32-94300363c549" containerName="extract-content" Jan 26 17:22:00 crc kubenswrapper[4754]: I0126 17:22:00.944978 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="74000075-a282-49a3-ab32-94300363c549" containerName="extract-content" Jan 26 17:22:00 crc kubenswrapper[4754]: E0126 17:22:00.944989 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74000075-a282-49a3-ab32-94300363c549" containerName="registry-server" Jan 26 17:22:00 crc kubenswrapper[4754]: I0126 17:22:00.944998 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="74000075-a282-49a3-ab32-94300363c549" containerName="registry-server" Jan 26 17:22:00 crc kubenswrapper[4754]: I0126 17:22:00.945236 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="9bfc59f9-f503-4fec-9a31-240819fc3a52" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Jan 26 17:22:00 crc kubenswrapper[4754]: I0126 17:22:00.945265 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="74000075-a282-49a3-ab32-94300363c549" containerName="registry-server" Jan 26 17:22:00 crc kubenswrapper[4754]: I0126 17:22:00.946758 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-v9nvp" Jan 26 17:22:00 crc kubenswrapper[4754]: I0126 17:22:00.948695 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Jan 26 17:22:00 crc kubenswrapper[4754]: I0126 17:22:00.948898 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Jan 26 17:22:00 crc kubenswrapper[4754]: I0126 17:22:00.949299 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Jan 26 17:22:00 crc kubenswrapper[4754]: I0126 17:22:00.949966 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-fz7z8" Jan 26 17:22:00 crc kubenswrapper[4754]: I0126 17:22:00.961197 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-v9nvp"] Jan 26 17:22:01 crc kubenswrapper[4754]: I0126 17:22:01.100533 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-85bz6\" (UniqueName: \"kubernetes.io/projected/c31e5ca9-fa72-4350-bd79-b58eb9b0e7a4-kube-api-access-85bz6\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-v9nvp\" (UID: \"c31e5ca9-fa72-4350-bd79-b58eb9b0e7a4\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-v9nvp" Jan 26 17:22:01 crc kubenswrapper[4754]: I0126 17:22:01.100651 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c31e5ca9-fa72-4350-bd79-b58eb9b0e7a4-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-v9nvp\" (UID: \"c31e5ca9-fa72-4350-bd79-b58eb9b0e7a4\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-v9nvp" Jan 26 17:22:01 crc kubenswrapper[4754]: I0126 17:22:01.100779 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/c31e5ca9-fa72-4350-bd79-b58eb9b0e7a4-ssh-key-openstack-edpm-ipam\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-v9nvp\" (UID: \"c31e5ca9-fa72-4350-bd79-b58eb9b0e7a4\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-v9nvp" Jan 26 17:22:01 crc kubenswrapper[4754]: I0126 17:22:01.202731 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c31e5ca9-fa72-4350-bd79-b58eb9b0e7a4-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-v9nvp\" (UID: \"c31e5ca9-fa72-4350-bd79-b58eb9b0e7a4\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-v9nvp" Jan 26 17:22:01 crc kubenswrapper[4754]: I0126 17:22:01.202916 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/c31e5ca9-fa72-4350-bd79-b58eb9b0e7a4-ssh-key-openstack-edpm-ipam\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-v9nvp\" (UID: \"c31e5ca9-fa72-4350-bd79-b58eb9b0e7a4\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-v9nvp" Jan 26 17:22:01 crc kubenswrapper[4754]: I0126 17:22:01.203150 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-85bz6\" (UniqueName: \"kubernetes.io/projected/c31e5ca9-fa72-4350-bd79-b58eb9b0e7a4-kube-api-access-85bz6\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-v9nvp\" (UID: \"c31e5ca9-fa72-4350-bd79-b58eb9b0e7a4\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-v9nvp" Jan 26 17:22:01 crc kubenswrapper[4754]: I0126 17:22:01.208618 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c31e5ca9-fa72-4350-bd79-b58eb9b0e7a4-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-v9nvp\" (UID: \"c31e5ca9-fa72-4350-bd79-b58eb9b0e7a4\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-v9nvp" Jan 26 17:22:01 crc kubenswrapper[4754]: I0126 17:22:01.209386 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/c31e5ca9-fa72-4350-bd79-b58eb9b0e7a4-ssh-key-openstack-edpm-ipam\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-v9nvp\" (UID: \"c31e5ca9-fa72-4350-bd79-b58eb9b0e7a4\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-v9nvp" Jan 26 17:22:01 crc kubenswrapper[4754]: I0126 17:22:01.225946 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-85bz6\" (UniqueName: \"kubernetes.io/projected/c31e5ca9-fa72-4350-bd79-b58eb9b0e7a4-kube-api-access-85bz6\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-v9nvp\" (UID: \"c31e5ca9-fa72-4350-bd79-b58eb9b0e7a4\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-v9nvp" Jan 26 17:22:01 crc kubenswrapper[4754]: I0126 17:22:01.264816 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-v9nvp" Jan 26 17:22:01 crc kubenswrapper[4754]: I0126 17:22:01.788780 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-v9nvp"] Jan 26 17:22:01 crc kubenswrapper[4754]: W0126 17:22:01.793047 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc31e5ca9_fa72_4350_bd79_b58eb9b0e7a4.slice/crio-c424330780311a2857329edb962a913ad64a214de0d3fa6a0677af87188f3487 WatchSource:0}: Error finding container c424330780311a2857329edb962a913ad64a214de0d3fa6a0677af87188f3487: Status 404 returned error can't find the container with id c424330780311a2857329edb962a913ad64a214de0d3fa6a0677af87188f3487 Jan 26 17:22:01 crc kubenswrapper[4754]: I0126 17:22:01.866153 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-v9nvp" event={"ID":"c31e5ca9-fa72-4350-bd79-b58eb9b0e7a4","Type":"ContainerStarted","Data":"c424330780311a2857329edb962a913ad64a214de0d3fa6a0677af87188f3487"} Jan 26 17:22:02 crc kubenswrapper[4754]: I0126 17:22:02.875358 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-v9nvp" event={"ID":"c31e5ca9-fa72-4350-bd79-b58eb9b0e7a4","Type":"ContainerStarted","Data":"a7374f6a9b6ab24208f2a6ce777d8c8ddd9259b4cb281e10a31c17a1d589ead1"} Jan 26 17:22:02 crc kubenswrapper[4754]: I0126 17:22:02.896153 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-v9nvp" podStartSLOduration=2.361404789 podStartE2EDuration="2.896133261s" podCreationTimestamp="2026-01-26 17:22:00 +0000 UTC" firstStartedPulling="2026-01-26 17:22:01.795803536 +0000 UTC m=+2088.319983970" lastFinishedPulling="2026-01-26 17:22:02.330532008 +0000 UTC m=+2088.854712442" observedRunningTime="2026-01-26 17:22:02.888454336 +0000 UTC m=+2089.412634770" watchObservedRunningTime="2026-01-26 17:22:02.896133261 +0000 UTC m=+2089.420313695" Jan 26 17:22:27 crc kubenswrapper[4754]: I0126 17:22:27.182259 4754 scope.go:117] "RemoveContainer" containerID="1dc2a31a2a2541e5978333a908c08012c296165c761ff33581b1b2a9ea38bb40" Jan 26 17:22:34 crc kubenswrapper[4754]: I0126 17:22:34.664892 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-ptrzr"] Jan 26 17:22:34 crc kubenswrapper[4754]: I0126 17:22:34.667216 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ptrzr" Jan 26 17:22:34 crc kubenswrapper[4754]: I0126 17:22:34.676862 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-ptrzr"] Jan 26 17:22:34 crc kubenswrapper[4754]: I0126 17:22:34.818079 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fqprn\" (UniqueName: \"kubernetes.io/projected/6c24aa1c-ab07-4d9e-9fe9-95a6196c7cf7-kube-api-access-fqprn\") pod \"community-operators-ptrzr\" (UID: \"6c24aa1c-ab07-4d9e-9fe9-95a6196c7cf7\") " pod="openshift-marketplace/community-operators-ptrzr" Jan 26 17:22:34 crc kubenswrapper[4754]: I0126 17:22:34.818132 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6c24aa1c-ab07-4d9e-9fe9-95a6196c7cf7-utilities\") pod \"community-operators-ptrzr\" (UID: \"6c24aa1c-ab07-4d9e-9fe9-95a6196c7cf7\") " pod="openshift-marketplace/community-operators-ptrzr" Jan 26 17:22:34 crc kubenswrapper[4754]: I0126 17:22:34.818193 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6c24aa1c-ab07-4d9e-9fe9-95a6196c7cf7-catalog-content\") pod \"community-operators-ptrzr\" (UID: \"6c24aa1c-ab07-4d9e-9fe9-95a6196c7cf7\") " pod="openshift-marketplace/community-operators-ptrzr" Jan 26 17:22:34 crc kubenswrapper[4754]: I0126 17:22:34.928129 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fqprn\" (UniqueName: \"kubernetes.io/projected/6c24aa1c-ab07-4d9e-9fe9-95a6196c7cf7-kube-api-access-fqprn\") pod \"community-operators-ptrzr\" (UID: \"6c24aa1c-ab07-4d9e-9fe9-95a6196c7cf7\") " pod="openshift-marketplace/community-operators-ptrzr" Jan 26 17:22:34 crc kubenswrapper[4754]: I0126 17:22:34.928208 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6c24aa1c-ab07-4d9e-9fe9-95a6196c7cf7-utilities\") pod \"community-operators-ptrzr\" (UID: \"6c24aa1c-ab07-4d9e-9fe9-95a6196c7cf7\") " pod="openshift-marketplace/community-operators-ptrzr" Jan 26 17:22:34 crc kubenswrapper[4754]: I0126 17:22:34.928295 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6c24aa1c-ab07-4d9e-9fe9-95a6196c7cf7-catalog-content\") pod \"community-operators-ptrzr\" (UID: \"6c24aa1c-ab07-4d9e-9fe9-95a6196c7cf7\") " pod="openshift-marketplace/community-operators-ptrzr" Jan 26 17:22:34 crc kubenswrapper[4754]: I0126 17:22:34.929726 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6c24aa1c-ab07-4d9e-9fe9-95a6196c7cf7-utilities\") pod \"community-operators-ptrzr\" (UID: \"6c24aa1c-ab07-4d9e-9fe9-95a6196c7cf7\") " pod="openshift-marketplace/community-operators-ptrzr" Jan 26 17:22:34 crc kubenswrapper[4754]: I0126 17:22:34.929987 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6c24aa1c-ab07-4d9e-9fe9-95a6196c7cf7-catalog-content\") pod \"community-operators-ptrzr\" (UID: \"6c24aa1c-ab07-4d9e-9fe9-95a6196c7cf7\") " pod="openshift-marketplace/community-operators-ptrzr" Jan 26 17:22:34 crc kubenswrapper[4754]: I0126 17:22:34.949782 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fqprn\" (UniqueName: \"kubernetes.io/projected/6c24aa1c-ab07-4d9e-9fe9-95a6196c7cf7-kube-api-access-fqprn\") pod \"community-operators-ptrzr\" (UID: \"6c24aa1c-ab07-4d9e-9fe9-95a6196c7cf7\") " pod="openshift-marketplace/community-operators-ptrzr" Jan 26 17:22:34 crc kubenswrapper[4754]: I0126 17:22:34.987586 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ptrzr" Jan 26 17:22:35 crc kubenswrapper[4754]: I0126 17:22:35.545813 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-ptrzr"] Jan 26 17:22:36 crc kubenswrapper[4754]: I0126 17:22:36.195565 4754 generic.go:334] "Generic (PLEG): container finished" podID="6c24aa1c-ab07-4d9e-9fe9-95a6196c7cf7" containerID="d2c4377478edb87b8fc004b93c6db20a7569aab5e7bf8c00dff6dcff4bdfab79" exitCode=0 Jan 26 17:22:36 crc kubenswrapper[4754]: I0126 17:22:36.195646 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ptrzr" event={"ID":"6c24aa1c-ab07-4d9e-9fe9-95a6196c7cf7","Type":"ContainerDied","Data":"d2c4377478edb87b8fc004b93c6db20a7569aab5e7bf8c00dff6dcff4bdfab79"} Jan 26 17:22:36 crc kubenswrapper[4754]: I0126 17:22:36.195912 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ptrzr" event={"ID":"6c24aa1c-ab07-4d9e-9fe9-95a6196c7cf7","Type":"ContainerStarted","Data":"99ffa84b36990b9a03178c4e16655d4c219c6512fbac84fe5f43aafb4212f79e"} Jan 26 17:22:39 crc kubenswrapper[4754]: I0126 17:22:39.219794 4754 generic.go:334] "Generic (PLEG): container finished" podID="6c24aa1c-ab07-4d9e-9fe9-95a6196c7cf7" containerID="9055b47674b37695db60dfa2c00b74efa4a234365c2ba7b8035852dc49ddc27f" exitCode=0 Jan 26 17:22:39 crc kubenswrapper[4754]: I0126 17:22:39.219972 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ptrzr" event={"ID":"6c24aa1c-ab07-4d9e-9fe9-95a6196c7cf7","Type":"ContainerDied","Data":"9055b47674b37695db60dfa2c00b74efa4a234365c2ba7b8035852dc49ddc27f"} Jan 26 17:22:39 crc kubenswrapper[4754]: I0126 17:22:39.225288 4754 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 26 17:22:40 crc kubenswrapper[4754]: I0126 17:22:40.231586 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ptrzr" event={"ID":"6c24aa1c-ab07-4d9e-9fe9-95a6196c7cf7","Type":"ContainerStarted","Data":"5feaa070c7990612742be9edf823a8728b75f95160142769217d190a4d832c24"} Jan 26 17:22:40 crc kubenswrapper[4754]: I0126 17:22:40.253479 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-ptrzr" podStartSLOduration=2.580727279 podStartE2EDuration="6.253458802s" podCreationTimestamp="2026-01-26 17:22:34 +0000 UTC" firstStartedPulling="2026-01-26 17:22:36.197386645 +0000 UTC m=+2122.721567069" lastFinishedPulling="2026-01-26 17:22:39.870118158 +0000 UTC m=+2126.394298592" observedRunningTime="2026-01-26 17:22:40.249635096 +0000 UTC m=+2126.773815540" watchObservedRunningTime="2026-01-26 17:22:40.253458802 +0000 UTC m=+2126.777639256" Jan 26 17:22:44 crc kubenswrapper[4754]: I0126 17:22:44.988587 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-ptrzr" Jan 26 17:22:44 crc kubenswrapper[4754]: I0126 17:22:44.989201 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-ptrzr" Jan 26 17:22:45 crc kubenswrapper[4754]: I0126 17:22:45.037350 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-ptrzr" Jan 26 17:22:45 crc kubenswrapper[4754]: I0126 17:22:45.359125 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-ptrzr" Jan 26 17:22:45 crc kubenswrapper[4754]: I0126 17:22:45.419515 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-ptrzr"] Jan 26 17:22:47 crc kubenswrapper[4754]: I0126 17:22:47.304218 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-ptrzr" podUID="6c24aa1c-ab07-4d9e-9fe9-95a6196c7cf7" containerName="registry-server" containerID="cri-o://5feaa070c7990612742be9edf823a8728b75f95160142769217d190a4d832c24" gracePeriod=2 Jan 26 17:22:47 crc kubenswrapper[4754]: I0126 17:22:47.804720 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ptrzr" Jan 26 17:22:47 crc kubenswrapper[4754]: I0126 17:22:47.896062 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6c24aa1c-ab07-4d9e-9fe9-95a6196c7cf7-utilities\") pod \"6c24aa1c-ab07-4d9e-9fe9-95a6196c7cf7\" (UID: \"6c24aa1c-ab07-4d9e-9fe9-95a6196c7cf7\") " Jan 26 17:22:47 crc kubenswrapper[4754]: I0126 17:22:47.896112 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqprn\" (UniqueName: \"kubernetes.io/projected/6c24aa1c-ab07-4d9e-9fe9-95a6196c7cf7-kube-api-access-fqprn\") pod \"6c24aa1c-ab07-4d9e-9fe9-95a6196c7cf7\" (UID: \"6c24aa1c-ab07-4d9e-9fe9-95a6196c7cf7\") " Jan 26 17:22:47 crc kubenswrapper[4754]: I0126 17:22:47.896208 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6c24aa1c-ab07-4d9e-9fe9-95a6196c7cf7-catalog-content\") pod \"6c24aa1c-ab07-4d9e-9fe9-95a6196c7cf7\" (UID: \"6c24aa1c-ab07-4d9e-9fe9-95a6196c7cf7\") " Jan 26 17:22:47 crc kubenswrapper[4754]: I0126 17:22:47.897008 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6c24aa1c-ab07-4d9e-9fe9-95a6196c7cf7-utilities" (OuterVolumeSpecName: "utilities") pod "6c24aa1c-ab07-4d9e-9fe9-95a6196c7cf7" (UID: "6c24aa1c-ab07-4d9e-9fe9-95a6196c7cf7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:22:47 crc kubenswrapper[4754]: I0126 17:22:47.903069 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6c24aa1c-ab07-4d9e-9fe9-95a6196c7cf7-kube-api-access-fqprn" (OuterVolumeSpecName: "kube-api-access-fqprn") pod "6c24aa1c-ab07-4d9e-9fe9-95a6196c7cf7" (UID: "6c24aa1c-ab07-4d9e-9fe9-95a6196c7cf7"). InnerVolumeSpecName "kube-api-access-fqprn". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:22:47 crc kubenswrapper[4754]: I0126 17:22:47.955854 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6c24aa1c-ab07-4d9e-9fe9-95a6196c7cf7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6c24aa1c-ab07-4d9e-9fe9-95a6196c7cf7" (UID: "6c24aa1c-ab07-4d9e-9fe9-95a6196c7cf7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:22:47 crc kubenswrapper[4754]: I0126 17:22:47.998411 4754 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6c24aa1c-ab07-4d9e-9fe9-95a6196c7cf7-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 26 17:22:47 crc kubenswrapper[4754]: I0126 17:22:47.998463 4754 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6c24aa1c-ab07-4d9e-9fe9-95a6196c7cf7-utilities\") on node \"crc\" DevicePath \"\"" Jan 26 17:22:47 crc kubenswrapper[4754]: I0126 17:22:47.998476 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqprn\" (UniqueName: \"kubernetes.io/projected/6c24aa1c-ab07-4d9e-9fe9-95a6196c7cf7-kube-api-access-fqprn\") on node \"crc\" DevicePath \"\"" Jan 26 17:22:48 crc kubenswrapper[4754]: I0126 17:22:48.314295 4754 generic.go:334] "Generic (PLEG): container finished" podID="6c24aa1c-ab07-4d9e-9fe9-95a6196c7cf7" containerID="5feaa070c7990612742be9edf823a8728b75f95160142769217d190a4d832c24" exitCode=0 Jan 26 17:22:48 crc kubenswrapper[4754]: I0126 17:22:48.314338 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ptrzr" event={"ID":"6c24aa1c-ab07-4d9e-9fe9-95a6196c7cf7","Type":"ContainerDied","Data":"5feaa070c7990612742be9edf823a8728b75f95160142769217d190a4d832c24"} Jan 26 17:22:48 crc kubenswrapper[4754]: I0126 17:22:48.314364 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ptrzr" event={"ID":"6c24aa1c-ab07-4d9e-9fe9-95a6196c7cf7","Type":"ContainerDied","Data":"99ffa84b36990b9a03178c4e16655d4c219c6512fbac84fe5f43aafb4212f79e"} Jan 26 17:22:48 crc kubenswrapper[4754]: I0126 17:22:48.314358 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ptrzr" Jan 26 17:22:48 crc kubenswrapper[4754]: I0126 17:22:48.314400 4754 scope.go:117] "RemoveContainer" containerID="5feaa070c7990612742be9edf823a8728b75f95160142769217d190a4d832c24" Jan 26 17:22:48 crc kubenswrapper[4754]: I0126 17:22:48.337360 4754 scope.go:117] "RemoveContainer" containerID="9055b47674b37695db60dfa2c00b74efa4a234365c2ba7b8035852dc49ddc27f" Jan 26 17:22:48 crc kubenswrapper[4754]: I0126 17:22:48.356822 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-ptrzr"] Jan 26 17:22:48 crc kubenswrapper[4754]: I0126 17:22:48.371041 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-ptrzr"] Jan 26 17:22:48 crc kubenswrapper[4754]: I0126 17:22:48.376205 4754 scope.go:117] "RemoveContainer" containerID="d2c4377478edb87b8fc004b93c6db20a7569aab5e7bf8c00dff6dcff4bdfab79" Jan 26 17:22:48 crc kubenswrapper[4754]: I0126 17:22:48.413194 4754 scope.go:117] "RemoveContainer" containerID="5feaa070c7990612742be9edf823a8728b75f95160142769217d190a4d832c24" Jan 26 17:22:48 crc kubenswrapper[4754]: E0126 17:22:48.413646 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5feaa070c7990612742be9edf823a8728b75f95160142769217d190a4d832c24\": container with ID starting with 5feaa070c7990612742be9edf823a8728b75f95160142769217d190a4d832c24 not found: ID does not exist" containerID="5feaa070c7990612742be9edf823a8728b75f95160142769217d190a4d832c24" Jan 26 17:22:48 crc kubenswrapper[4754]: I0126 17:22:48.413781 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5feaa070c7990612742be9edf823a8728b75f95160142769217d190a4d832c24"} err="failed to get container status \"5feaa070c7990612742be9edf823a8728b75f95160142769217d190a4d832c24\": rpc error: code = NotFound desc = could not find container \"5feaa070c7990612742be9edf823a8728b75f95160142769217d190a4d832c24\": container with ID starting with 5feaa070c7990612742be9edf823a8728b75f95160142769217d190a4d832c24 not found: ID does not exist" Jan 26 17:22:48 crc kubenswrapper[4754]: I0126 17:22:48.413907 4754 scope.go:117] "RemoveContainer" containerID="9055b47674b37695db60dfa2c00b74efa4a234365c2ba7b8035852dc49ddc27f" Jan 26 17:22:48 crc kubenswrapper[4754]: E0126 17:22:48.414232 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9055b47674b37695db60dfa2c00b74efa4a234365c2ba7b8035852dc49ddc27f\": container with ID starting with 9055b47674b37695db60dfa2c00b74efa4a234365c2ba7b8035852dc49ddc27f not found: ID does not exist" containerID="9055b47674b37695db60dfa2c00b74efa4a234365c2ba7b8035852dc49ddc27f" Jan 26 17:22:48 crc kubenswrapper[4754]: I0126 17:22:48.414259 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9055b47674b37695db60dfa2c00b74efa4a234365c2ba7b8035852dc49ddc27f"} err="failed to get container status \"9055b47674b37695db60dfa2c00b74efa4a234365c2ba7b8035852dc49ddc27f\": rpc error: code = NotFound desc = could not find container \"9055b47674b37695db60dfa2c00b74efa4a234365c2ba7b8035852dc49ddc27f\": container with ID starting with 9055b47674b37695db60dfa2c00b74efa4a234365c2ba7b8035852dc49ddc27f not found: ID does not exist" Jan 26 17:22:48 crc kubenswrapper[4754]: I0126 17:22:48.414277 4754 scope.go:117] "RemoveContainer" containerID="d2c4377478edb87b8fc004b93c6db20a7569aab5e7bf8c00dff6dcff4bdfab79" Jan 26 17:22:48 crc kubenswrapper[4754]: E0126 17:22:48.414491 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d2c4377478edb87b8fc004b93c6db20a7569aab5e7bf8c00dff6dcff4bdfab79\": container with ID starting with d2c4377478edb87b8fc004b93c6db20a7569aab5e7bf8c00dff6dcff4bdfab79 not found: ID does not exist" containerID="d2c4377478edb87b8fc004b93c6db20a7569aab5e7bf8c00dff6dcff4bdfab79" Jan 26 17:22:48 crc kubenswrapper[4754]: I0126 17:22:48.414588 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d2c4377478edb87b8fc004b93c6db20a7569aab5e7bf8c00dff6dcff4bdfab79"} err="failed to get container status \"d2c4377478edb87b8fc004b93c6db20a7569aab5e7bf8c00dff6dcff4bdfab79\": rpc error: code = NotFound desc = could not find container \"d2c4377478edb87b8fc004b93c6db20a7569aab5e7bf8c00dff6dcff4bdfab79\": container with ID starting with d2c4377478edb87b8fc004b93c6db20a7569aab5e7bf8c00dff6dcff4bdfab79 not found: ID does not exist" Jan 26 17:22:49 crc kubenswrapper[4754]: I0126 17:22:49.777508 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6c24aa1c-ab07-4d9e-9fe9-95a6196c7cf7" path="/var/lib/kubelet/pods/6c24aa1c-ab07-4d9e-9fe9-95a6196c7cf7/volumes" Jan 26 17:22:53 crc kubenswrapper[4754]: I0126 17:22:53.362367 4754 generic.go:334] "Generic (PLEG): container finished" podID="c31e5ca9-fa72-4350-bd79-b58eb9b0e7a4" containerID="a7374f6a9b6ab24208f2a6ce777d8c8ddd9259b4cb281e10a31c17a1d589ead1" exitCode=0 Jan 26 17:22:53 crc kubenswrapper[4754]: I0126 17:22:53.362450 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-v9nvp" event={"ID":"c31e5ca9-fa72-4350-bd79-b58eb9b0e7a4","Type":"ContainerDied","Data":"a7374f6a9b6ab24208f2a6ce777d8c8ddd9259b4cb281e10a31c17a1d589ead1"} Jan 26 17:22:54 crc kubenswrapper[4754]: I0126 17:22:54.951300 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-v9nvp" Jan 26 17:22:55 crc kubenswrapper[4754]: I0126 17:22:55.026905 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/c31e5ca9-fa72-4350-bd79-b58eb9b0e7a4-ssh-key-openstack-edpm-ipam\") pod \"c31e5ca9-fa72-4350-bd79-b58eb9b0e7a4\" (UID: \"c31e5ca9-fa72-4350-bd79-b58eb9b0e7a4\") " Jan 26 17:22:55 crc kubenswrapper[4754]: I0126 17:22:55.026986 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-85bz6\" (UniqueName: \"kubernetes.io/projected/c31e5ca9-fa72-4350-bd79-b58eb9b0e7a4-kube-api-access-85bz6\") pod \"c31e5ca9-fa72-4350-bd79-b58eb9b0e7a4\" (UID: \"c31e5ca9-fa72-4350-bd79-b58eb9b0e7a4\") " Jan 26 17:22:55 crc kubenswrapper[4754]: I0126 17:22:55.027019 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c31e5ca9-fa72-4350-bd79-b58eb9b0e7a4-inventory\") pod \"c31e5ca9-fa72-4350-bd79-b58eb9b0e7a4\" (UID: \"c31e5ca9-fa72-4350-bd79-b58eb9b0e7a4\") " Jan 26 17:22:55 crc kubenswrapper[4754]: I0126 17:22:55.032403 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c31e5ca9-fa72-4350-bd79-b58eb9b0e7a4-kube-api-access-85bz6" (OuterVolumeSpecName: "kube-api-access-85bz6") pod "c31e5ca9-fa72-4350-bd79-b58eb9b0e7a4" (UID: "c31e5ca9-fa72-4350-bd79-b58eb9b0e7a4"). InnerVolumeSpecName "kube-api-access-85bz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:22:55 crc kubenswrapper[4754]: I0126 17:22:55.056123 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c31e5ca9-fa72-4350-bd79-b58eb9b0e7a4-inventory" (OuterVolumeSpecName: "inventory") pod "c31e5ca9-fa72-4350-bd79-b58eb9b0e7a4" (UID: "c31e5ca9-fa72-4350-bd79-b58eb9b0e7a4"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:22:55 crc kubenswrapper[4754]: I0126 17:22:55.057732 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c31e5ca9-fa72-4350-bd79-b58eb9b0e7a4-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "c31e5ca9-fa72-4350-bd79-b58eb9b0e7a4" (UID: "c31e5ca9-fa72-4350-bd79-b58eb9b0e7a4"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:22:55 crc kubenswrapper[4754]: I0126 17:22:55.128772 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-85bz6\" (UniqueName: \"kubernetes.io/projected/c31e5ca9-fa72-4350-bd79-b58eb9b0e7a4-kube-api-access-85bz6\") on node \"crc\" DevicePath \"\"" Jan 26 17:22:55 crc kubenswrapper[4754]: I0126 17:22:55.128803 4754 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c31e5ca9-fa72-4350-bd79-b58eb9b0e7a4-inventory\") on node \"crc\" DevicePath \"\"" Jan 26 17:22:55 crc kubenswrapper[4754]: I0126 17:22:55.128813 4754 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/c31e5ca9-fa72-4350-bd79-b58eb9b0e7a4-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Jan 26 17:22:55 crc kubenswrapper[4754]: I0126 17:22:55.390338 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-v9nvp" event={"ID":"c31e5ca9-fa72-4350-bd79-b58eb9b0e7a4","Type":"ContainerDied","Data":"c424330780311a2857329edb962a913ad64a214de0d3fa6a0677af87188f3487"} Jan 26 17:22:55 crc kubenswrapper[4754]: I0126 17:22:55.390408 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c424330780311a2857329edb962a913ad64a214de0d3fa6a0677af87188f3487" Jan 26 17:22:55 crc kubenswrapper[4754]: I0126 17:22:55.390463 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-v9nvp" Jan 26 17:22:55 crc kubenswrapper[4754]: I0126 17:22:55.464368 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-sqqxn"] Jan 26 17:22:55 crc kubenswrapper[4754]: E0126 17:22:55.464735 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6c24aa1c-ab07-4d9e-9fe9-95a6196c7cf7" containerName="extract-content" Jan 26 17:22:55 crc kubenswrapper[4754]: I0126 17:22:55.464746 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="6c24aa1c-ab07-4d9e-9fe9-95a6196c7cf7" containerName="extract-content" Jan 26 17:22:55 crc kubenswrapper[4754]: E0126 17:22:55.464771 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c31e5ca9-fa72-4350-bd79-b58eb9b0e7a4" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Jan 26 17:22:55 crc kubenswrapper[4754]: I0126 17:22:55.464778 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="c31e5ca9-fa72-4350-bd79-b58eb9b0e7a4" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Jan 26 17:22:55 crc kubenswrapper[4754]: E0126 17:22:55.464789 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6c24aa1c-ab07-4d9e-9fe9-95a6196c7cf7" containerName="registry-server" Jan 26 17:22:55 crc kubenswrapper[4754]: I0126 17:22:55.464795 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="6c24aa1c-ab07-4d9e-9fe9-95a6196c7cf7" containerName="registry-server" Jan 26 17:22:55 crc kubenswrapper[4754]: E0126 17:22:55.464809 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6c24aa1c-ab07-4d9e-9fe9-95a6196c7cf7" containerName="extract-utilities" Jan 26 17:22:55 crc kubenswrapper[4754]: I0126 17:22:55.464814 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="6c24aa1c-ab07-4d9e-9fe9-95a6196c7cf7" containerName="extract-utilities" Jan 26 17:22:55 crc kubenswrapper[4754]: I0126 17:22:55.464991 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="c31e5ca9-fa72-4350-bd79-b58eb9b0e7a4" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Jan 26 17:22:55 crc kubenswrapper[4754]: I0126 17:22:55.465003 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="6c24aa1c-ab07-4d9e-9fe9-95a6196c7cf7" containerName="registry-server" Jan 26 17:22:55 crc kubenswrapper[4754]: I0126 17:22:55.465580 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-sqqxn" Jan 26 17:22:55 crc kubenswrapper[4754]: I0126 17:22:55.468076 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Jan 26 17:22:55 crc kubenswrapper[4754]: I0126 17:22:55.468233 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Jan 26 17:22:55 crc kubenswrapper[4754]: I0126 17:22:55.468356 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-fz7z8" Jan 26 17:22:55 crc kubenswrapper[4754]: I0126 17:22:55.475391 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Jan 26 17:22:55 crc kubenswrapper[4754]: I0126 17:22:55.477756 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-sqqxn"] Jan 26 17:22:55 crc kubenswrapper[4754]: I0126 17:22:55.535471 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/8dcb5a12-a14f-4f95-8991-5a1b9b9bb431-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-sqqxn\" (UID: \"8dcb5a12-a14f-4f95-8991-5a1b9b9bb431\") " pod="openstack/ssh-known-hosts-edpm-deployment-sqqxn" Jan 26 17:22:55 crc kubenswrapper[4754]: I0126 17:22:55.535710 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/8dcb5a12-a14f-4f95-8991-5a1b9b9bb431-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-sqqxn\" (UID: \"8dcb5a12-a14f-4f95-8991-5a1b9b9bb431\") " pod="openstack/ssh-known-hosts-edpm-deployment-sqqxn" Jan 26 17:22:55 crc kubenswrapper[4754]: I0126 17:22:55.535758 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5dn9m\" (UniqueName: \"kubernetes.io/projected/8dcb5a12-a14f-4f95-8991-5a1b9b9bb431-kube-api-access-5dn9m\") pod \"ssh-known-hosts-edpm-deployment-sqqxn\" (UID: \"8dcb5a12-a14f-4f95-8991-5a1b9b9bb431\") " pod="openstack/ssh-known-hosts-edpm-deployment-sqqxn" Jan 26 17:22:55 crc kubenswrapper[4754]: I0126 17:22:55.637930 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/8dcb5a12-a14f-4f95-8991-5a1b9b9bb431-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-sqqxn\" (UID: \"8dcb5a12-a14f-4f95-8991-5a1b9b9bb431\") " pod="openstack/ssh-known-hosts-edpm-deployment-sqqxn" Jan 26 17:22:55 crc kubenswrapper[4754]: I0126 17:22:55.638134 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/8dcb5a12-a14f-4f95-8991-5a1b9b9bb431-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-sqqxn\" (UID: \"8dcb5a12-a14f-4f95-8991-5a1b9b9bb431\") " pod="openstack/ssh-known-hosts-edpm-deployment-sqqxn" Jan 26 17:22:55 crc kubenswrapper[4754]: I0126 17:22:55.638190 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5dn9m\" (UniqueName: \"kubernetes.io/projected/8dcb5a12-a14f-4f95-8991-5a1b9b9bb431-kube-api-access-5dn9m\") pod \"ssh-known-hosts-edpm-deployment-sqqxn\" (UID: \"8dcb5a12-a14f-4f95-8991-5a1b9b9bb431\") " pod="openstack/ssh-known-hosts-edpm-deployment-sqqxn" Jan 26 17:22:55 crc kubenswrapper[4754]: I0126 17:22:55.642492 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/8dcb5a12-a14f-4f95-8991-5a1b9b9bb431-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-sqqxn\" (UID: \"8dcb5a12-a14f-4f95-8991-5a1b9b9bb431\") " pod="openstack/ssh-known-hosts-edpm-deployment-sqqxn" Jan 26 17:22:55 crc kubenswrapper[4754]: I0126 17:22:55.643217 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/8dcb5a12-a14f-4f95-8991-5a1b9b9bb431-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-sqqxn\" (UID: \"8dcb5a12-a14f-4f95-8991-5a1b9b9bb431\") " pod="openstack/ssh-known-hosts-edpm-deployment-sqqxn" Jan 26 17:22:55 crc kubenswrapper[4754]: I0126 17:22:55.657023 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5dn9m\" (UniqueName: \"kubernetes.io/projected/8dcb5a12-a14f-4f95-8991-5a1b9b9bb431-kube-api-access-5dn9m\") pod \"ssh-known-hosts-edpm-deployment-sqqxn\" (UID: \"8dcb5a12-a14f-4f95-8991-5a1b9b9bb431\") " pod="openstack/ssh-known-hosts-edpm-deployment-sqqxn" Jan 26 17:22:55 crc kubenswrapper[4754]: I0126 17:22:55.788546 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-sqqxn" Jan 26 17:22:56 crc kubenswrapper[4754]: I0126 17:22:56.297533 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-sqqxn"] Jan 26 17:22:56 crc kubenswrapper[4754]: I0126 17:22:56.400969 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-sqqxn" event={"ID":"8dcb5a12-a14f-4f95-8991-5a1b9b9bb431","Type":"ContainerStarted","Data":"0099074ce4520faf4dcf564926f781e054f2c92679f343978917988ae9246689"} Jan 26 17:22:57 crc kubenswrapper[4754]: I0126 17:22:57.410915 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-sqqxn" event={"ID":"8dcb5a12-a14f-4f95-8991-5a1b9b9bb431","Type":"ContainerStarted","Data":"9a2b229d12aaf18c88fb0ec5a3248576d713af0f6e42fd47ed1f23797fb51a75"} Jan 26 17:22:57 crc kubenswrapper[4754]: I0126 17:22:57.426507 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-edpm-deployment-sqqxn" podStartSLOduration=1.7842559169999999 podStartE2EDuration="2.426486281s" podCreationTimestamp="2026-01-26 17:22:55 +0000 UTC" firstStartedPulling="2026-01-26 17:22:56.3107319 +0000 UTC m=+2142.834912334" lastFinishedPulling="2026-01-26 17:22:56.952962264 +0000 UTC m=+2143.477142698" observedRunningTime="2026-01-26 17:22:57.425186435 +0000 UTC m=+2143.949366889" watchObservedRunningTime="2026-01-26 17:22:57.426486281 +0000 UTC m=+2143.950666715" Jan 26 17:23:04 crc kubenswrapper[4754]: E0126 17:23:04.159824 4754 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8dcb5a12_a14f_4f95_8991_5a1b9b9bb431.slice/crio-conmon-9a2b229d12aaf18c88fb0ec5a3248576d713af0f6e42fd47ed1f23797fb51a75.scope\": RecentStats: unable to find data in memory cache]" Jan 26 17:23:04 crc kubenswrapper[4754]: I0126 17:23:04.473056 4754 generic.go:334] "Generic (PLEG): container finished" podID="8dcb5a12-a14f-4f95-8991-5a1b9b9bb431" containerID="9a2b229d12aaf18c88fb0ec5a3248576d713af0f6e42fd47ed1f23797fb51a75" exitCode=0 Jan 26 17:23:04 crc kubenswrapper[4754]: I0126 17:23:04.473124 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-sqqxn" event={"ID":"8dcb5a12-a14f-4f95-8991-5a1b9b9bb431","Type":"ContainerDied","Data":"9a2b229d12aaf18c88fb0ec5a3248576d713af0f6e42fd47ed1f23797fb51a75"} Jan 26 17:23:05 crc kubenswrapper[4754]: I0126 17:23:05.916949 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-sqqxn" Jan 26 17:23:05 crc kubenswrapper[4754]: I0126 17:23:05.928500 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5dn9m\" (UniqueName: \"kubernetes.io/projected/8dcb5a12-a14f-4f95-8991-5a1b9b9bb431-kube-api-access-5dn9m\") pod \"8dcb5a12-a14f-4f95-8991-5a1b9b9bb431\" (UID: \"8dcb5a12-a14f-4f95-8991-5a1b9b9bb431\") " Jan 26 17:23:05 crc kubenswrapper[4754]: I0126 17:23:05.928562 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/8dcb5a12-a14f-4f95-8991-5a1b9b9bb431-ssh-key-openstack-edpm-ipam\") pod \"8dcb5a12-a14f-4f95-8991-5a1b9b9bb431\" (UID: \"8dcb5a12-a14f-4f95-8991-5a1b9b9bb431\") " Jan 26 17:23:05 crc kubenswrapper[4754]: I0126 17:23:05.928644 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/8dcb5a12-a14f-4f95-8991-5a1b9b9bb431-inventory-0\") pod \"8dcb5a12-a14f-4f95-8991-5a1b9b9bb431\" (UID: \"8dcb5a12-a14f-4f95-8991-5a1b9b9bb431\") " Jan 26 17:23:05 crc kubenswrapper[4754]: I0126 17:23:05.938632 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8dcb5a12-a14f-4f95-8991-5a1b9b9bb431-kube-api-access-5dn9m" (OuterVolumeSpecName: "kube-api-access-5dn9m") pod "8dcb5a12-a14f-4f95-8991-5a1b9b9bb431" (UID: "8dcb5a12-a14f-4f95-8991-5a1b9b9bb431"). InnerVolumeSpecName "kube-api-access-5dn9m". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:23:05 crc kubenswrapper[4754]: I0126 17:23:05.966606 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8dcb5a12-a14f-4f95-8991-5a1b9b9bb431-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "8dcb5a12-a14f-4f95-8991-5a1b9b9bb431" (UID: "8dcb5a12-a14f-4f95-8991-5a1b9b9bb431"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:23:05 crc kubenswrapper[4754]: I0126 17:23:05.968139 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8dcb5a12-a14f-4f95-8991-5a1b9b9bb431-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "8dcb5a12-a14f-4f95-8991-5a1b9b9bb431" (UID: "8dcb5a12-a14f-4f95-8991-5a1b9b9bb431"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:23:06 crc kubenswrapper[4754]: I0126 17:23:06.030764 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5dn9m\" (UniqueName: \"kubernetes.io/projected/8dcb5a12-a14f-4f95-8991-5a1b9b9bb431-kube-api-access-5dn9m\") on node \"crc\" DevicePath \"\"" Jan 26 17:23:06 crc kubenswrapper[4754]: I0126 17:23:06.030941 4754 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/8dcb5a12-a14f-4f95-8991-5a1b9b9bb431-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Jan 26 17:23:06 crc kubenswrapper[4754]: I0126 17:23:06.031026 4754 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/8dcb5a12-a14f-4f95-8991-5a1b9b9bb431-inventory-0\") on node \"crc\" DevicePath \"\"" Jan 26 17:23:06 crc kubenswrapper[4754]: I0126 17:23:06.488952 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-sqqxn" event={"ID":"8dcb5a12-a14f-4f95-8991-5a1b9b9bb431","Type":"ContainerDied","Data":"0099074ce4520faf4dcf564926f781e054f2c92679f343978917988ae9246689"} Jan 26 17:23:06 crc kubenswrapper[4754]: I0126 17:23:06.488995 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0099074ce4520faf4dcf564926f781e054f2c92679f343978917988ae9246689" Jan 26 17:23:06 crc kubenswrapper[4754]: I0126 17:23:06.489023 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-sqqxn" Jan 26 17:23:06 crc kubenswrapper[4754]: I0126 17:23:06.564326 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-7rh8g"] Jan 26 17:23:06 crc kubenswrapper[4754]: E0126 17:23:06.564832 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8dcb5a12-a14f-4f95-8991-5a1b9b9bb431" containerName="ssh-known-hosts-edpm-deployment" Jan 26 17:23:06 crc kubenswrapper[4754]: I0126 17:23:06.564852 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="8dcb5a12-a14f-4f95-8991-5a1b9b9bb431" containerName="ssh-known-hosts-edpm-deployment" Jan 26 17:23:06 crc kubenswrapper[4754]: I0126 17:23:06.565092 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="8dcb5a12-a14f-4f95-8991-5a1b9b9bb431" containerName="ssh-known-hosts-edpm-deployment" Jan 26 17:23:06 crc kubenswrapper[4754]: I0126 17:23:06.566004 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-7rh8g" Jan 26 17:23:06 crc kubenswrapper[4754]: I0126 17:23:06.572197 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Jan 26 17:23:06 crc kubenswrapper[4754]: I0126 17:23:06.572450 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Jan 26 17:23:06 crc kubenswrapper[4754]: I0126 17:23:06.572901 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Jan 26 17:23:06 crc kubenswrapper[4754]: I0126 17:23:06.573040 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-fz7z8" Jan 26 17:23:06 crc kubenswrapper[4754]: I0126 17:23:06.577122 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-7rh8g"] Jan 26 17:23:06 crc kubenswrapper[4754]: I0126 17:23:06.745391 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v8wwl\" (UniqueName: \"kubernetes.io/projected/3dbed37c-3982-4625-929e-d99a8fd798e6-kube-api-access-v8wwl\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-7rh8g\" (UID: \"3dbed37c-3982-4625-929e-d99a8fd798e6\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-7rh8g" Jan 26 17:23:06 crc kubenswrapper[4754]: I0126 17:23:06.745904 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/3dbed37c-3982-4625-929e-d99a8fd798e6-ssh-key-openstack-edpm-ipam\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-7rh8g\" (UID: \"3dbed37c-3982-4625-929e-d99a8fd798e6\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-7rh8g" Jan 26 17:23:06 crc kubenswrapper[4754]: I0126 17:23:06.745946 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3dbed37c-3982-4625-929e-d99a8fd798e6-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-7rh8g\" (UID: \"3dbed37c-3982-4625-929e-d99a8fd798e6\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-7rh8g" Jan 26 17:23:06 crc kubenswrapper[4754]: I0126 17:23:06.847469 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/3dbed37c-3982-4625-929e-d99a8fd798e6-ssh-key-openstack-edpm-ipam\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-7rh8g\" (UID: \"3dbed37c-3982-4625-929e-d99a8fd798e6\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-7rh8g" Jan 26 17:23:06 crc kubenswrapper[4754]: I0126 17:23:06.847526 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3dbed37c-3982-4625-929e-d99a8fd798e6-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-7rh8g\" (UID: \"3dbed37c-3982-4625-929e-d99a8fd798e6\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-7rh8g" Jan 26 17:23:06 crc kubenswrapper[4754]: I0126 17:23:06.847631 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v8wwl\" (UniqueName: \"kubernetes.io/projected/3dbed37c-3982-4625-929e-d99a8fd798e6-kube-api-access-v8wwl\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-7rh8g\" (UID: \"3dbed37c-3982-4625-929e-d99a8fd798e6\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-7rh8g" Jan 26 17:23:06 crc kubenswrapper[4754]: I0126 17:23:06.852053 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/3dbed37c-3982-4625-929e-d99a8fd798e6-ssh-key-openstack-edpm-ipam\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-7rh8g\" (UID: \"3dbed37c-3982-4625-929e-d99a8fd798e6\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-7rh8g" Jan 26 17:23:06 crc kubenswrapper[4754]: I0126 17:23:06.852095 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3dbed37c-3982-4625-929e-d99a8fd798e6-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-7rh8g\" (UID: \"3dbed37c-3982-4625-929e-d99a8fd798e6\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-7rh8g" Jan 26 17:23:06 crc kubenswrapper[4754]: I0126 17:23:06.867854 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v8wwl\" (UniqueName: \"kubernetes.io/projected/3dbed37c-3982-4625-929e-d99a8fd798e6-kube-api-access-v8wwl\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-7rh8g\" (UID: \"3dbed37c-3982-4625-929e-d99a8fd798e6\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-7rh8g" Jan 26 17:23:06 crc kubenswrapper[4754]: I0126 17:23:06.919371 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-7rh8g" Jan 26 17:23:07 crc kubenswrapper[4754]: I0126 17:23:07.484650 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-7rh8g"] Jan 26 17:23:07 crc kubenswrapper[4754]: I0126 17:23:07.498562 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-7rh8g" event={"ID":"3dbed37c-3982-4625-929e-d99a8fd798e6","Type":"ContainerStarted","Data":"c88dd83e12721b9f5abf37671c01ee9a86e2924ad772ae9a3e110eb998cab186"} Jan 26 17:23:09 crc kubenswrapper[4754]: I0126 17:23:09.515253 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-7rh8g" event={"ID":"3dbed37c-3982-4625-929e-d99a8fd798e6","Type":"ContainerStarted","Data":"64444001df2b974e8b500bd448285fac9422cf71271e7eaa642ab60ab9e58c4d"} Jan 26 17:23:09 crc kubenswrapper[4754]: I0126 17:23:09.534608 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-7rh8g" podStartSLOduration=2.708353404 podStartE2EDuration="3.534588409s" podCreationTimestamp="2026-01-26 17:23:06 +0000 UTC" firstStartedPulling="2026-01-26 17:23:07.487079469 +0000 UTC m=+2154.011259903" lastFinishedPulling="2026-01-26 17:23:08.313314474 +0000 UTC m=+2154.837494908" observedRunningTime="2026-01-26 17:23:09.531190425 +0000 UTC m=+2156.055370879" watchObservedRunningTime="2026-01-26 17:23:09.534588409 +0000 UTC m=+2156.058768843" Jan 26 17:23:17 crc kubenswrapper[4754]: I0126 17:23:17.579900 4754 generic.go:334] "Generic (PLEG): container finished" podID="3dbed37c-3982-4625-929e-d99a8fd798e6" containerID="64444001df2b974e8b500bd448285fac9422cf71271e7eaa642ab60ab9e58c4d" exitCode=0 Jan 26 17:23:17 crc kubenswrapper[4754]: I0126 17:23:17.580356 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-7rh8g" event={"ID":"3dbed37c-3982-4625-929e-d99a8fd798e6","Type":"ContainerDied","Data":"64444001df2b974e8b500bd448285fac9422cf71271e7eaa642ab60ab9e58c4d"} Jan 26 17:23:19 crc kubenswrapper[4754]: I0126 17:23:19.012531 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-7rh8g" Jan 26 17:23:19 crc kubenswrapper[4754]: I0126 17:23:19.103119 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v8wwl\" (UniqueName: \"kubernetes.io/projected/3dbed37c-3982-4625-929e-d99a8fd798e6-kube-api-access-v8wwl\") pod \"3dbed37c-3982-4625-929e-d99a8fd798e6\" (UID: \"3dbed37c-3982-4625-929e-d99a8fd798e6\") " Jan 26 17:23:19 crc kubenswrapper[4754]: I0126 17:23:19.103181 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/3dbed37c-3982-4625-929e-d99a8fd798e6-ssh-key-openstack-edpm-ipam\") pod \"3dbed37c-3982-4625-929e-d99a8fd798e6\" (UID: \"3dbed37c-3982-4625-929e-d99a8fd798e6\") " Jan 26 17:23:19 crc kubenswrapper[4754]: I0126 17:23:19.103549 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3dbed37c-3982-4625-929e-d99a8fd798e6-inventory\") pod \"3dbed37c-3982-4625-929e-d99a8fd798e6\" (UID: \"3dbed37c-3982-4625-929e-d99a8fd798e6\") " Jan 26 17:23:19 crc kubenswrapper[4754]: I0126 17:23:19.111032 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3dbed37c-3982-4625-929e-d99a8fd798e6-kube-api-access-v8wwl" (OuterVolumeSpecName: "kube-api-access-v8wwl") pod "3dbed37c-3982-4625-929e-d99a8fd798e6" (UID: "3dbed37c-3982-4625-929e-d99a8fd798e6"). InnerVolumeSpecName "kube-api-access-v8wwl". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:23:19 crc kubenswrapper[4754]: I0126 17:23:19.140943 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3dbed37c-3982-4625-929e-d99a8fd798e6-inventory" (OuterVolumeSpecName: "inventory") pod "3dbed37c-3982-4625-929e-d99a8fd798e6" (UID: "3dbed37c-3982-4625-929e-d99a8fd798e6"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:23:19 crc kubenswrapper[4754]: I0126 17:23:19.141159 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3dbed37c-3982-4625-929e-d99a8fd798e6-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "3dbed37c-3982-4625-929e-d99a8fd798e6" (UID: "3dbed37c-3982-4625-929e-d99a8fd798e6"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:23:19 crc kubenswrapper[4754]: I0126 17:23:19.206754 4754 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3dbed37c-3982-4625-929e-d99a8fd798e6-inventory\") on node \"crc\" DevicePath \"\"" Jan 26 17:23:19 crc kubenswrapper[4754]: I0126 17:23:19.206795 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v8wwl\" (UniqueName: \"kubernetes.io/projected/3dbed37c-3982-4625-929e-d99a8fd798e6-kube-api-access-v8wwl\") on node \"crc\" DevicePath \"\"" Jan 26 17:23:19 crc kubenswrapper[4754]: I0126 17:23:19.206809 4754 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/3dbed37c-3982-4625-929e-d99a8fd798e6-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Jan 26 17:23:19 crc kubenswrapper[4754]: I0126 17:23:19.598010 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-7rh8g" event={"ID":"3dbed37c-3982-4625-929e-d99a8fd798e6","Type":"ContainerDied","Data":"c88dd83e12721b9f5abf37671c01ee9a86e2924ad772ae9a3e110eb998cab186"} Jan 26 17:23:19 crc kubenswrapper[4754]: I0126 17:23:19.598090 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c88dd83e12721b9f5abf37671c01ee9a86e2924ad772ae9a3e110eb998cab186" Jan 26 17:23:19 crc kubenswrapper[4754]: I0126 17:23:19.598143 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-7rh8g" Jan 26 17:23:19 crc kubenswrapper[4754]: I0126 17:23:19.676138 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-p4jpg"] Jan 26 17:23:19 crc kubenswrapper[4754]: E0126 17:23:19.676502 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3dbed37c-3982-4625-929e-d99a8fd798e6" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Jan 26 17:23:19 crc kubenswrapper[4754]: I0126 17:23:19.676520 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="3dbed37c-3982-4625-929e-d99a8fd798e6" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Jan 26 17:23:19 crc kubenswrapper[4754]: I0126 17:23:19.676725 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="3dbed37c-3982-4625-929e-d99a8fd798e6" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Jan 26 17:23:19 crc kubenswrapper[4754]: I0126 17:23:19.677303 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-p4jpg" Jan 26 17:23:19 crc kubenswrapper[4754]: I0126 17:23:19.679755 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Jan 26 17:23:19 crc kubenswrapper[4754]: I0126 17:23:19.680018 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Jan 26 17:23:19 crc kubenswrapper[4754]: I0126 17:23:19.680726 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-fz7z8" Jan 26 17:23:19 crc kubenswrapper[4754]: I0126 17:23:19.682098 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Jan 26 17:23:19 crc kubenswrapper[4754]: I0126 17:23:19.685633 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-p4jpg"] Jan 26 17:23:19 crc kubenswrapper[4754]: I0126 17:23:19.729056 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/e4cbbee8-ed91-4ad6-983b-569cff60fa07-ssh-key-openstack-edpm-ipam\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-p4jpg\" (UID: \"e4cbbee8-ed91-4ad6-983b-569cff60fa07\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-p4jpg" Jan 26 17:23:19 crc kubenswrapper[4754]: I0126 17:23:19.729360 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e4cbbee8-ed91-4ad6-983b-569cff60fa07-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-p4jpg\" (UID: \"e4cbbee8-ed91-4ad6-983b-569cff60fa07\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-p4jpg" Jan 26 17:23:19 crc kubenswrapper[4754]: I0126 17:23:19.729650 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r4sjk\" (UniqueName: \"kubernetes.io/projected/e4cbbee8-ed91-4ad6-983b-569cff60fa07-kube-api-access-r4sjk\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-p4jpg\" (UID: \"e4cbbee8-ed91-4ad6-983b-569cff60fa07\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-p4jpg" Jan 26 17:23:19 crc kubenswrapper[4754]: I0126 17:23:19.831351 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/e4cbbee8-ed91-4ad6-983b-569cff60fa07-ssh-key-openstack-edpm-ipam\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-p4jpg\" (UID: \"e4cbbee8-ed91-4ad6-983b-569cff60fa07\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-p4jpg" Jan 26 17:23:19 crc kubenswrapper[4754]: I0126 17:23:19.831675 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e4cbbee8-ed91-4ad6-983b-569cff60fa07-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-p4jpg\" (UID: \"e4cbbee8-ed91-4ad6-983b-569cff60fa07\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-p4jpg" Jan 26 17:23:19 crc kubenswrapper[4754]: I0126 17:23:19.831718 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r4sjk\" (UniqueName: \"kubernetes.io/projected/e4cbbee8-ed91-4ad6-983b-569cff60fa07-kube-api-access-r4sjk\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-p4jpg\" (UID: \"e4cbbee8-ed91-4ad6-983b-569cff60fa07\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-p4jpg" Jan 26 17:23:19 crc kubenswrapper[4754]: I0126 17:23:19.836555 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/e4cbbee8-ed91-4ad6-983b-569cff60fa07-ssh-key-openstack-edpm-ipam\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-p4jpg\" (UID: \"e4cbbee8-ed91-4ad6-983b-569cff60fa07\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-p4jpg" Jan 26 17:23:19 crc kubenswrapper[4754]: I0126 17:23:19.856523 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r4sjk\" (UniqueName: \"kubernetes.io/projected/e4cbbee8-ed91-4ad6-983b-569cff60fa07-kube-api-access-r4sjk\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-p4jpg\" (UID: \"e4cbbee8-ed91-4ad6-983b-569cff60fa07\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-p4jpg" Jan 26 17:23:19 crc kubenswrapper[4754]: I0126 17:23:19.859822 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e4cbbee8-ed91-4ad6-983b-569cff60fa07-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-p4jpg\" (UID: \"e4cbbee8-ed91-4ad6-983b-569cff60fa07\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-p4jpg" Jan 26 17:23:19 crc kubenswrapper[4754]: I0126 17:23:19.993784 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-p4jpg" Jan 26 17:23:20 crc kubenswrapper[4754]: I0126 17:23:20.525340 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-p4jpg"] Jan 26 17:23:20 crc kubenswrapper[4754]: I0126 17:23:20.608426 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-p4jpg" event={"ID":"e4cbbee8-ed91-4ad6-983b-569cff60fa07","Type":"ContainerStarted","Data":"f6c481d77ba4a65a4296f825ee66391cde9a57725f1e6718da1f76ceb5f788a7"} Jan 26 17:23:21 crc kubenswrapper[4754]: I0126 17:23:21.621963 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-p4jpg" event={"ID":"e4cbbee8-ed91-4ad6-983b-569cff60fa07","Type":"ContainerStarted","Data":"8961bdbb2378031e1a436a62eb157b65a386b5cee2e2c2503910cf91541c4e70"} Jan 26 17:23:21 crc kubenswrapper[4754]: I0126 17:23:21.644426 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-p4jpg" podStartSLOduration=2.116503262 podStartE2EDuration="2.644410327s" podCreationTimestamp="2026-01-26 17:23:19 +0000 UTC" firstStartedPulling="2026-01-26 17:23:20.529589262 +0000 UTC m=+2167.053769696" lastFinishedPulling="2026-01-26 17:23:21.057496327 +0000 UTC m=+2167.581676761" observedRunningTime="2026-01-26 17:23:21.642109113 +0000 UTC m=+2168.166289537" watchObservedRunningTime="2026-01-26 17:23:21.644410327 +0000 UTC m=+2168.168590761" Jan 26 17:23:31 crc kubenswrapper[4754]: I0126 17:23:31.715507 4754 generic.go:334] "Generic (PLEG): container finished" podID="e4cbbee8-ed91-4ad6-983b-569cff60fa07" containerID="8961bdbb2378031e1a436a62eb157b65a386b5cee2e2c2503910cf91541c4e70" exitCode=0 Jan 26 17:23:31 crc kubenswrapper[4754]: I0126 17:23:31.715559 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-p4jpg" event={"ID":"e4cbbee8-ed91-4ad6-983b-569cff60fa07","Type":"ContainerDied","Data":"8961bdbb2378031e1a436a62eb157b65a386b5cee2e2c2503910cf91541c4e70"} Jan 26 17:23:33 crc kubenswrapper[4754]: I0126 17:23:33.161552 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-p4jpg" Jan 26 17:23:33 crc kubenswrapper[4754]: I0126 17:23:33.207828 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r4sjk\" (UniqueName: \"kubernetes.io/projected/e4cbbee8-ed91-4ad6-983b-569cff60fa07-kube-api-access-r4sjk\") pod \"e4cbbee8-ed91-4ad6-983b-569cff60fa07\" (UID: \"e4cbbee8-ed91-4ad6-983b-569cff60fa07\") " Jan 26 17:23:33 crc kubenswrapper[4754]: I0126 17:23:33.208095 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e4cbbee8-ed91-4ad6-983b-569cff60fa07-inventory\") pod \"e4cbbee8-ed91-4ad6-983b-569cff60fa07\" (UID: \"e4cbbee8-ed91-4ad6-983b-569cff60fa07\") " Jan 26 17:23:33 crc kubenswrapper[4754]: I0126 17:23:33.208121 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/e4cbbee8-ed91-4ad6-983b-569cff60fa07-ssh-key-openstack-edpm-ipam\") pod \"e4cbbee8-ed91-4ad6-983b-569cff60fa07\" (UID: \"e4cbbee8-ed91-4ad6-983b-569cff60fa07\") " Jan 26 17:23:33 crc kubenswrapper[4754]: I0126 17:23:33.214847 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e4cbbee8-ed91-4ad6-983b-569cff60fa07-kube-api-access-r4sjk" (OuterVolumeSpecName: "kube-api-access-r4sjk") pod "e4cbbee8-ed91-4ad6-983b-569cff60fa07" (UID: "e4cbbee8-ed91-4ad6-983b-569cff60fa07"). InnerVolumeSpecName "kube-api-access-r4sjk". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:23:33 crc kubenswrapper[4754]: I0126 17:23:33.257632 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e4cbbee8-ed91-4ad6-983b-569cff60fa07-inventory" (OuterVolumeSpecName: "inventory") pod "e4cbbee8-ed91-4ad6-983b-569cff60fa07" (UID: "e4cbbee8-ed91-4ad6-983b-569cff60fa07"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:23:33 crc kubenswrapper[4754]: I0126 17:23:33.260037 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e4cbbee8-ed91-4ad6-983b-569cff60fa07-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "e4cbbee8-ed91-4ad6-983b-569cff60fa07" (UID: "e4cbbee8-ed91-4ad6-983b-569cff60fa07"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:23:33 crc kubenswrapper[4754]: I0126 17:23:33.312297 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r4sjk\" (UniqueName: \"kubernetes.io/projected/e4cbbee8-ed91-4ad6-983b-569cff60fa07-kube-api-access-r4sjk\") on node \"crc\" DevicePath \"\"" Jan 26 17:23:33 crc kubenswrapper[4754]: I0126 17:23:33.312626 4754 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e4cbbee8-ed91-4ad6-983b-569cff60fa07-inventory\") on node \"crc\" DevicePath \"\"" Jan 26 17:23:33 crc kubenswrapper[4754]: I0126 17:23:33.312723 4754 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/e4cbbee8-ed91-4ad6-983b-569cff60fa07-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Jan 26 17:23:33 crc kubenswrapper[4754]: I0126 17:23:33.736733 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-p4jpg" event={"ID":"e4cbbee8-ed91-4ad6-983b-569cff60fa07","Type":"ContainerDied","Data":"f6c481d77ba4a65a4296f825ee66391cde9a57725f1e6718da1f76ceb5f788a7"} Jan 26 17:23:33 crc kubenswrapper[4754]: I0126 17:23:33.736782 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f6c481d77ba4a65a4296f825ee66391cde9a57725f1e6718da1f76ceb5f788a7" Jan 26 17:23:33 crc kubenswrapper[4754]: I0126 17:23:33.736874 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-p4jpg" Jan 26 17:23:33 crc kubenswrapper[4754]: I0126 17:23:33.842475 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-qpc2q"] Jan 26 17:23:33 crc kubenswrapper[4754]: E0126 17:23:33.843767 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e4cbbee8-ed91-4ad6-983b-569cff60fa07" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Jan 26 17:23:33 crc kubenswrapper[4754]: I0126 17:23:33.843816 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="e4cbbee8-ed91-4ad6-983b-569cff60fa07" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Jan 26 17:23:33 crc kubenswrapper[4754]: I0126 17:23:33.844079 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="e4cbbee8-ed91-4ad6-983b-569cff60fa07" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Jan 26 17:23:33 crc kubenswrapper[4754]: I0126 17:23:33.844890 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-qpc2q" Jan 26 17:23:33 crc kubenswrapper[4754]: I0126 17:23:33.849224 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-ovn-default-certs-0" Jan 26 17:23:33 crc kubenswrapper[4754]: I0126 17:23:33.849445 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Jan 26 17:23:33 crc kubenswrapper[4754]: I0126 17:23:33.849595 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-neutron-metadata-default-certs-0" Jan 26 17:23:33 crc kubenswrapper[4754]: I0126 17:23:33.851140 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Jan 26 17:23:33 crc kubenswrapper[4754]: I0126 17:23:33.851355 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-telemetry-default-certs-0" Jan 26 17:23:33 crc kubenswrapper[4754]: I0126 17:23:33.851477 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-qpc2q"] Jan 26 17:23:33 crc kubenswrapper[4754]: I0126 17:23:33.851508 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-libvirt-default-certs-0" Jan 26 17:23:33 crc kubenswrapper[4754]: I0126 17:23:33.851553 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Jan 26 17:23:33 crc kubenswrapper[4754]: I0126 17:23:33.851835 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-fz7z8" Jan 26 17:23:33 crc kubenswrapper[4754]: I0126 17:23:33.924824 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/70a69e93-9d4a-4d51-b0c9-4b986172e2ac-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-qpc2q\" (UID: \"70a69e93-9d4a-4d51-b0c9-4b986172e2ac\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-qpc2q" Jan 26 17:23:33 crc kubenswrapper[4754]: I0126 17:23:33.925184 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70a69e93-9d4a-4d51-b0c9-4b986172e2ac-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-qpc2q\" (UID: \"70a69e93-9d4a-4d51-b0c9-4b986172e2ac\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-qpc2q" Jan 26 17:23:33 crc kubenswrapper[4754]: I0126 17:23:33.925245 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/70a69e93-9d4a-4d51-b0c9-4b986172e2ac-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-qpc2q\" (UID: \"70a69e93-9d4a-4d51-b0c9-4b986172e2ac\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-qpc2q" Jan 26 17:23:33 crc kubenswrapper[4754]: I0126 17:23:33.925272 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70a69e93-9d4a-4d51-b0c9-4b986172e2ac-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-qpc2q\" (UID: \"70a69e93-9d4a-4d51-b0c9-4b986172e2ac\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-qpc2q" Jan 26 17:23:33 crc kubenswrapper[4754]: I0126 17:23:33.925375 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bwgr2\" (UniqueName: \"kubernetes.io/projected/70a69e93-9d4a-4d51-b0c9-4b986172e2ac-kube-api-access-bwgr2\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-qpc2q\" (UID: \"70a69e93-9d4a-4d51-b0c9-4b986172e2ac\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-qpc2q" Jan 26 17:23:33 crc kubenswrapper[4754]: I0126 17:23:33.925398 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/70a69e93-9d4a-4d51-b0c9-4b986172e2ac-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-qpc2q\" (UID: \"70a69e93-9d4a-4d51-b0c9-4b986172e2ac\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-qpc2q" Jan 26 17:23:33 crc kubenswrapper[4754]: I0126 17:23:33.925415 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70a69e93-9d4a-4d51-b0c9-4b986172e2ac-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-qpc2q\" (UID: \"70a69e93-9d4a-4d51-b0c9-4b986172e2ac\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-qpc2q" Jan 26 17:23:33 crc kubenswrapper[4754]: I0126 17:23:33.925442 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70a69e93-9d4a-4d51-b0c9-4b986172e2ac-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-qpc2q\" (UID: \"70a69e93-9d4a-4d51-b0c9-4b986172e2ac\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-qpc2q" Jan 26 17:23:33 crc kubenswrapper[4754]: I0126 17:23:33.925462 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70a69e93-9d4a-4d51-b0c9-4b986172e2ac-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-qpc2q\" (UID: \"70a69e93-9d4a-4d51-b0c9-4b986172e2ac\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-qpc2q" Jan 26 17:23:33 crc kubenswrapper[4754]: I0126 17:23:33.925477 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/70a69e93-9d4a-4d51-b0c9-4b986172e2ac-ssh-key-openstack-edpm-ipam\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-qpc2q\" (UID: \"70a69e93-9d4a-4d51-b0c9-4b986172e2ac\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-qpc2q" Jan 26 17:23:33 crc kubenswrapper[4754]: I0126 17:23:33.925501 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/70a69e93-9d4a-4d51-b0c9-4b986172e2ac-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-qpc2q\" (UID: \"70a69e93-9d4a-4d51-b0c9-4b986172e2ac\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-qpc2q" Jan 26 17:23:33 crc kubenswrapper[4754]: I0126 17:23:33.925531 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70a69e93-9d4a-4d51-b0c9-4b986172e2ac-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-qpc2q\" (UID: \"70a69e93-9d4a-4d51-b0c9-4b986172e2ac\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-qpc2q" Jan 26 17:23:33 crc kubenswrapper[4754]: I0126 17:23:33.925553 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70a69e93-9d4a-4d51-b0c9-4b986172e2ac-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-qpc2q\" (UID: \"70a69e93-9d4a-4d51-b0c9-4b986172e2ac\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-qpc2q" Jan 26 17:23:33 crc kubenswrapper[4754]: I0126 17:23:33.925571 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/70a69e93-9d4a-4d51-b0c9-4b986172e2ac-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-qpc2q\" (UID: \"70a69e93-9d4a-4d51-b0c9-4b986172e2ac\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-qpc2q" Jan 26 17:23:34 crc kubenswrapper[4754]: I0126 17:23:34.027197 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/70a69e93-9d4a-4d51-b0c9-4b986172e2ac-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-qpc2q\" (UID: \"70a69e93-9d4a-4d51-b0c9-4b986172e2ac\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-qpc2q" Jan 26 17:23:34 crc kubenswrapper[4754]: I0126 17:23:34.027498 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70a69e93-9d4a-4d51-b0c9-4b986172e2ac-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-qpc2q\" (UID: \"70a69e93-9d4a-4d51-b0c9-4b986172e2ac\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-qpc2q" Jan 26 17:23:34 crc kubenswrapper[4754]: I0126 17:23:34.027613 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70a69e93-9d4a-4d51-b0c9-4b986172e2ac-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-qpc2q\" (UID: \"70a69e93-9d4a-4d51-b0c9-4b986172e2ac\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-qpc2q" Jan 26 17:23:34 crc kubenswrapper[4754]: I0126 17:23:34.027737 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/70a69e93-9d4a-4d51-b0c9-4b986172e2ac-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-qpc2q\" (UID: \"70a69e93-9d4a-4d51-b0c9-4b986172e2ac\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-qpc2q" Jan 26 17:23:34 crc kubenswrapper[4754]: I0126 17:23:34.027865 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/70a69e93-9d4a-4d51-b0c9-4b986172e2ac-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-qpc2q\" (UID: \"70a69e93-9d4a-4d51-b0c9-4b986172e2ac\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-qpc2q" Jan 26 17:23:34 crc kubenswrapper[4754]: I0126 17:23:34.027987 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70a69e93-9d4a-4d51-b0c9-4b986172e2ac-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-qpc2q\" (UID: \"70a69e93-9d4a-4d51-b0c9-4b986172e2ac\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-qpc2q" Jan 26 17:23:34 crc kubenswrapper[4754]: I0126 17:23:34.028149 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/70a69e93-9d4a-4d51-b0c9-4b986172e2ac-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-qpc2q\" (UID: \"70a69e93-9d4a-4d51-b0c9-4b986172e2ac\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-qpc2q" Jan 26 17:23:34 crc kubenswrapper[4754]: I0126 17:23:34.028258 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70a69e93-9d4a-4d51-b0c9-4b986172e2ac-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-qpc2q\" (UID: \"70a69e93-9d4a-4d51-b0c9-4b986172e2ac\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-qpc2q" Jan 26 17:23:34 crc kubenswrapper[4754]: I0126 17:23:34.028454 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bwgr2\" (UniqueName: \"kubernetes.io/projected/70a69e93-9d4a-4d51-b0c9-4b986172e2ac-kube-api-access-bwgr2\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-qpc2q\" (UID: \"70a69e93-9d4a-4d51-b0c9-4b986172e2ac\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-qpc2q" Jan 26 17:23:34 crc kubenswrapper[4754]: I0126 17:23:34.029064 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/70a69e93-9d4a-4d51-b0c9-4b986172e2ac-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-qpc2q\" (UID: \"70a69e93-9d4a-4d51-b0c9-4b986172e2ac\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-qpc2q" Jan 26 17:23:34 crc kubenswrapper[4754]: I0126 17:23:34.029541 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70a69e93-9d4a-4d51-b0c9-4b986172e2ac-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-qpc2q\" (UID: \"70a69e93-9d4a-4d51-b0c9-4b986172e2ac\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-qpc2q" Jan 26 17:23:34 crc kubenswrapper[4754]: I0126 17:23:34.029738 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70a69e93-9d4a-4d51-b0c9-4b986172e2ac-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-qpc2q\" (UID: \"70a69e93-9d4a-4d51-b0c9-4b986172e2ac\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-qpc2q" Jan 26 17:23:34 crc kubenswrapper[4754]: I0126 17:23:34.029859 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70a69e93-9d4a-4d51-b0c9-4b986172e2ac-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-qpc2q\" (UID: \"70a69e93-9d4a-4d51-b0c9-4b986172e2ac\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-qpc2q" Jan 26 17:23:34 crc kubenswrapper[4754]: I0126 17:23:34.029953 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/70a69e93-9d4a-4d51-b0c9-4b986172e2ac-ssh-key-openstack-edpm-ipam\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-qpc2q\" (UID: \"70a69e93-9d4a-4d51-b0c9-4b986172e2ac\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-qpc2q" Jan 26 17:23:34 crc kubenswrapper[4754]: I0126 17:23:34.032217 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/70a69e93-9d4a-4d51-b0c9-4b986172e2ac-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-qpc2q\" (UID: \"70a69e93-9d4a-4d51-b0c9-4b986172e2ac\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-qpc2q" Jan 26 17:23:34 crc kubenswrapper[4754]: I0126 17:23:34.032318 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70a69e93-9d4a-4d51-b0c9-4b986172e2ac-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-qpc2q\" (UID: \"70a69e93-9d4a-4d51-b0c9-4b986172e2ac\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-qpc2q" Jan 26 17:23:34 crc kubenswrapper[4754]: I0126 17:23:34.032396 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70a69e93-9d4a-4d51-b0c9-4b986172e2ac-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-qpc2q\" (UID: \"70a69e93-9d4a-4d51-b0c9-4b986172e2ac\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-qpc2q" Jan 26 17:23:34 crc kubenswrapper[4754]: I0126 17:23:34.033234 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/70a69e93-9d4a-4d51-b0c9-4b986172e2ac-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-qpc2q\" (UID: \"70a69e93-9d4a-4d51-b0c9-4b986172e2ac\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-qpc2q" Jan 26 17:23:34 crc kubenswrapper[4754]: I0126 17:23:34.034340 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70a69e93-9d4a-4d51-b0c9-4b986172e2ac-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-qpc2q\" (UID: \"70a69e93-9d4a-4d51-b0c9-4b986172e2ac\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-qpc2q" Jan 26 17:23:34 crc kubenswrapper[4754]: I0126 17:23:34.034725 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/70a69e93-9d4a-4d51-b0c9-4b986172e2ac-ssh-key-openstack-edpm-ipam\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-qpc2q\" (UID: \"70a69e93-9d4a-4d51-b0c9-4b986172e2ac\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-qpc2q" Jan 26 17:23:34 crc kubenswrapper[4754]: I0126 17:23:34.034738 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/70a69e93-9d4a-4d51-b0c9-4b986172e2ac-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-qpc2q\" (UID: \"70a69e93-9d4a-4d51-b0c9-4b986172e2ac\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-qpc2q" Jan 26 17:23:34 crc kubenswrapper[4754]: I0126 17:23:34.035319 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70a69e93-9d4a-4d51-b0c9-4b986172e2ac-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-qpc2q\" (UID: \"70a69e93-9d4a-4d51-b0c9-4b986172e2ac\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-qpc2q" Jan 26 17:23:34 crc kubenswrapper[4754]: I0126 17:23:34.035625 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70a69e93-9d4a-4d51-b0c9-4b986172e2ac-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-qpc2q\" (UID: \"70a69e93-9d4a-4d51-b0c9-4b986172e2ac\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-qpc2q" Jan 26 17:23:34 crc kubenswrapper[4754]: I0126 17:23:34.036819 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/70a69e93-9d4a-4d51-b0c9-4b986172e2ac-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-qpc2q\" (UID: \"70a69e93-9d4a-4d51-b0c9-4b986172e2ac\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-qpc2q" Jan 26 17:23:34 crc kubenswrapper[4754]: I0126 17:23:34.036955 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70a69e93-9d4a-4d51-b0c9-4b986172e2ac-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-qpc2q\" (UID: \"70a69e93-9d4a-4d51-b0c9-4b986172e2ac\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-qpc2q" Jan 26 17:23:34 crc kubenswrapper[4754]: I0126 17:23:34.038035 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/70a69e93-9d4a-4d51-b0c9-4b986172e2ac-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-qpc2q\" (UID: \"70a69e93-9d4a-4d51-b0c9-4b986172e2ac\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-qpc2q" Jan 26 17:23:34 crc kubenswrapper[4754]: I0126 17:23:34.040459 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70a69e93-9d4a-4d51-b0c9-4b986172e2ac-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-qpc2q\" (UID: \"70a69e93-9d4a-4d51-b0c9-4b986172e2ac\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-qpc2q" Jan 26 17:23:34 crc kubenswrapper[4754]: I0126 17:23:34.047868 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bwgr2\" (UniqueName: \"kubernetes.io/projected/70a69e93-9d4a-4d51-b0c9-4b986172e2ac-kube-api-access-bwgr2\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-qpc2q\" (UID: \"70a69e93-9d4a-4d51-b0c9-4b986172e2ac\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-qpc2q" Jan 26 17:23:34 crc kubenswrapper[4754]: I0126 17:23:34.169859 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-qpc2q" Jan 26 17:23:34 crc kubenswrapper[4754]: I0126 17:23:34.718217 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-qpc2q"] Jan 26 17:23:34 crc kubenswrapper[4754]: I0126 17:23:34.744034 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-qpc2q" event={"ID":"70a69e93-9d4a-4d51-b0c9-4b986172e2ac","Type":"ContainerStarted","Data":"eff17dc73c46f83f57beb0a851a55b7f7eff7f00da9ca1716c10b954067a253e"} Jan 26 17:23:35 crc kubenswrapper[4754]: I0126 17:23:35.761160 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-qpc2q" event={"ID":"70a69e93-9d4a-4d51-b0c9-4b986172e2ac","Type":"ContainerStarted","Data":"95992d6a519cd0d86f7488d1abbb6984f63e60f036b22056c6f0ae31d440c67d"} Jan 26 17:23:35 crc kubenswrapper[4754]: I0126 17:23:35.777513 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-qpc2q" podStartSLOduration=2.169860984 podStartE2EDuration="2.777495409s" podCreationTimestamp="2026-01-26 17:23:33 +0000 UTC" firstStartedPulling="2026-01-26 17:23:34.720381414 +0000 UTC m=+2181.244561848" lastFinishedPulling="2026-01-26 17:23:35.328015839 +0000 UTC m=+2181.852196273" observedRunningTime="2026-01-26 17:23:35.776581354 +0000 UTC m=+2182.300761798" watchObservedRunningTime="2026-01-26 17:23:35.777495409 +0000 UTC m=+2182.301675843" Jan 26 17:23:37 crc kubenswrapper[4754]: I0126 17:23:37.129536 4754 patch_prober.go:28] interesting pod/machine-config-daemon-x65wv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 26 17:23:37 crc kubenswrapper[4754]: I0126 17:23:37.129904 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 26 17:24:07 crc kubenswrapper[4754]: I0126 17:24:07.129599 4754 patch_prober.go:28] interesting pod/machine-config-daemon-x65wv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 26 17:24:07 crc kubenswrapper[4754]: I0126 17:24:07.131876 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 26 17:24:12 crc kubenswrapper[4754]: I0126 17:24:12.343141 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-8dzsr"] Jan 26 17:24:12 crc kubenswrapper[4754]: I0126 17:24:12.346186 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8dzsr" Jan 26 17:24:12 crc kubenswrapper[4754]: I0126 17:24:12.358239 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-8dzsr"] Jan 26 17:24:12 crc kubenswrapper[4754]: I0126 17:24:12.479472 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b3b67190-caf6-444e-8c6f-bf2878f37421-utilities\") pod \"redhat-marketplace-8dzsr\" (UID: \"b3b67190-caf6-444e-8c6f-bf2878f37421\") " pod="openshift-marketplace/redhat-marketplace-8dzsr" Jan 26 17:24:12 crc kubenswrapper[4754]: I0126 17:24:12.479600 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b3b67190-caf6-444e-8c6f-bf2878f37421-catalog-content\") pod \"redhat-marketplace-8dzsr\" (UID: \"b3b67190-caf6-444e-8c6f-bf2878f37421\") " pod="openshift-marketplace/redhat-marketplace-8dzsr" Jan 26 17:24:12 crc kubenswrapper[4754]: I0126 17:24:12.479630 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h9kkw\" (UniqueName: \"kubernetes.io/projected/b3b67190-caf6-444e-8c6f-bf2878f37421-kube-api-access-h9kkw\") pod \"redhat-marketplace-8dzsr\" (UID: \"b3b67190-caf6-444e-8c6f-bf2878f37421\") " pod="openshift-marketplace/redhat-marketplace-8dzsr" Jan 26 17:24:12 crc kubenswrapper[4754]: I0126 17:24:12.581176 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b3b67190-caf6-444e-8c6f-bf2878f37421-catalog-content\") pod \"redhat-marketplace-8dzsr\" (UID: \"b3b67190-caf6-444e-8c6f-bf2878f37421\") " pod="openshift-marketplace/redhat-marketplace-8dzsr" Jan 26 17:24:12 crc kubenswrapper[4754]: I0126 17:24:12.581231 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h9kkw\" (UniqueName: \"kubernetes.io/projected/b3b67190-caf6-444e-8c6f-bf2878f37421-kube-api-access-h9kkw\") pod \"redhat-marketplace-8dzsr\" (UID: \"b3b67190-caf6-444e-8c6f-bf2878f37421\") " pod="openshift-marketplace/redhat-marketplace-8dzsr" Jan 26 17:24:12 crc kubenswrapper[4754]: I0126 17:24:12.581331 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b3b67190-caf6-444e-8c6f-bf2878f37421-utilities\") pod \"redhat-marketplace-8dzsr\" (UID: \"b3b67190-caf6-444e-8c6f-bf2878f37421\") " pod="openshift-marketplace/redhat-marketplace-8dzsr" Jan 26 17:24:12 crc kubenswrapper[4754]: I0126 17:24:12.581771 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b3b67190-caf6-444e-8c6f-bf2878f37421-catalog-content\") pod \"redhat-marketplace-8dzsr\" (UID: \"b3b67190-caf6-444e-8c6f-bf2878f37421\") " pod="openshift-marketplace/redhat-marketplace-8dzsr" Jan 26 17:24:12 crc kubenswrapper[4754]: I0126 17:24:12.581902 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b3b67190-caf6-444e-8c6f-bf2878f37421-utilities\") pod \"redhat-marketplace-8dzsr\" (UID: \"b3b67190-caf6-444e-8c6f-bf2878f37421\") " pod="openshift-marketplace/redhat-marketplace-8dzsr" Jan 26 17:24:12 crc kubenswrapper[4754]: I0126 17:24:12.600315 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h9kkw\" (UniqueName: \"kubernetes.io/projected/b3b67190-caf6-444e-8c6f-bf2878f37421-kube-api-access-h9kkw\") pod \"redhat-marketplace-8dzsr\" (UID: \"b3b67190-caf6-444e-8c6f-bf2878f37421\") " pod="openshift-marketplace/redhat-marketplace-8dzsr" Jan 26 17:24:12 crc kubenswrapper[4754]: I0126 17:24:12.670075 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8dzsr" Jan 26 17:24:13 crc kubenswrapper[4754]: I0126 17:24:13.134475 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-8dzsr"] Jan 26 17:24:14 crc kubenswrapper[4754]: I0126 17:24:14.105024 4754 generic.go:334] "Generic (PLEG): container finished" podID="b3b67190-caf6-444e-8c6f-bf2878f37421" containerID="0c1e9334cca604a6378d6354f3a5844bdf909694b48bdf7667a035164923f3f1" exitCode=0 Jan 26 17:24:14 crc kubenswrapper[4754]: I0126 17:24:14.105090 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8dzsr" event={"ID":"b3b67190-caf6-444e-8c6f-bf2878f37421","Type":"ContainerDied","Data":"0c1e9334cca604a6378d6354f3a5844bdf909694b48bdf7667a035164923f3f1"} Jan 26 17:24:14 crc kubenswrapper[4754]: I0126 17:24:14.105154 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8dzsr" event={"ID":"b3b67190-caf6-444e-8c6f-bf2878f37421","Type":"ContainerStarted","Data":"75e8fff32dff7f9e577931c33dd477e4b11fc1685612e4a7cb47df631cc8e538"} Jan 26 17:24:15 crc kubenswrapper[4754]: I0126 17:24:15.127356 4754 generic.go:334] "Generic (PLEG): container finished" podID="70a69e93-9d4a-4d51-b0c9-4b986172e2ac" containerID="95992d6a519cd0d86f7488d1abbb6984f63e60f036b22056c6f0ae31d440c67d" exitCode=0 Jan 26 17:24:15 crc kubenswrapper[4754]: I0126 17:24:15.127468 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-qpc2q" event={"ID":"70a69e93-9d4a-4d51-b0c9-4b986172e2ac","Type":"ContainerDied","Data":"95992d6a519cd0d86f7488d1abbb6984f63e60f036b22056c6f0ae31d440c67d"} Jan 26 17:24:15 crc kubenswrapper[4754]: E0126 17:24:15.908972 4754 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb3b67190_caf6_444e_8c6f_bf2878f37421.slice/crio-conmon-82d46e7ae399a5312b75aeea432e3fc1fb28917c134d936b9a47421a510c4406.scope\": RecentStats: unable to find data in memory cache]" Jan 26 17:24:16 crc kubenswrapper[4754]: I0126 17:24:16.138268 4754 generic.go:334] "Generic (PLEG): container finished" podID="b3b67190-caf6-444e-8c6f-bf2878f37421" containerID="82d46e7ae399a5312b75aeea432e3fc1fb28917c134d936b9a47421a510c4406" exitCode=0 Jan 26 17:24:16 crc kubenswrapper[4754]: I0126 17:24:16.138328 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8dzsr" event={"ID":"b3b67190-caf6-444e-8c6f-bf2878f37421","Type":"ContainerDied","Data":"82d46e7ae399a5312b75aeea432e3fc1fb28917c134d936b9a47421a510c4406"} Jan 26 17:24:16 crc kubenswrapper[4754]: I0126 17:24:16.538973 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-qpc2q" Jan 26 17:24:16 crc kubenswrapper[4754]: I0126 17:24:16.675512 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70a69e93-9d4a-4d51-b0c9-4b986172e2ac-telemetry-combined-ca-bundle\") pod \"70a69e93-9d4a-4d51-b0c9-4b986172e2ac\" (UID: \"70a69e93-9d4a-4d51-b0c9-4b986172e2ac\") " Jan 26 17:24:16 crc kubenswrapper[4754]: I0126 17:24:16.676043 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/70a69e93-9d4a-4d51-b0c9-4b986172e2ac-ssh-key-openstack-edpm-ipam\") pod \"70a69e93-9d4a-4d51-b0c9-4b986172e2ac\" (UID: \"70a69e93-9d4a-4d51-b0c9-4b986172e2ac\") " Jan 26 17:24:16 crc kubenswrapper[4754]: I0126 17:24:16.676105 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/70a69e93-9d4a-4d51-b0c9-4b986172e2ac-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"70a69e93-9d4a-4d51-b0c9-4b986172e2ac\" (UID: \"70a69e93-9d4a-4d51-b0c9-4b986172e2ac\") " Jan 26 17:24:16 crc kubenswrapper[4754]: I0126 17:24:16.676132 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/70a69e93-9d4a-4d51-b0c9-4b986172e2ac-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"70a69e93-9d4a-4d51-b0c9-4b986172e2ac\" (UID: \"70a69e93-9d4a-4d51-b0c9-4b986172e2ac\") " Jan 26 17:24:16 crc kubenswrapper[4754]: I0126 17:24:16.676592 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/70a69e93-9d4a-4d51-b0c9-4b986172e2ac-inventory\") pod \"70a69e93-9d4a-4d51-b0c9-4b986172e2ac\" (UID: \"70a69e93-9d4a-4d51-b0c9-4b986172e2ac\") " Jan 26 17:24:16 crc kubenswrapper[4754]: I0126 17:24:16.676648 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70a69e93-9d4a-4d51-b0c9-4b986172e2ac-ovn-combined-ca-bundle\") pod \"70a69e93-9d4a-4d51-b0c9-4b986172e2ac\" (UID: \"70a69e93-9d4a-4d51-b0c9-4b986172e2ac\") " Jan 26 17:24:16 crc kubenswrapper[4754]: I0126 17:24:16.676682 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/70a69e93-9d4a-4d51-b0c9-4b986172e2ac-openstack-edpm-ipam-ovn-default-certs-0\") pod \"70a69e93-9d4a-4d51-b0c9-4b986172e2ac\" (UID: \"70a69e93-9d4a-4d51-b0c9-4b986172e2ac\") " Jan 26 17:24:16 crc kubenswrapper[4754]: I0126 17:24:16.676707 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70a69e93-9d4a-4d51-b0c9-4b986172e2ac-nova-combined-ca-bundle\") pod \"70a69e93-9d4a-4d51-b0c9-4b986172e2ac\" (UID: \"70a69e93-9d4a-4d51-b0c9-4b986172e2ac\") " Jan 26 17:24:16 crc kubenswrapper[4754]: I0126 17:24:16.676757 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70a69e93-9d4a-4d51-b0c9-4b986172e2ac-libvirt-combined-ca-bundle\") pod \"70a69e93-9d4a-4d51-b0c9-4b986172e2ac\" (UID: \"70a69e93-9d4a-4d51-b0c9-4b986172e2ac\") " Jan 26 17:24:16 crc kubenswrapper[4754]: I0126 17:24:16.676779 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70a69e93-9d4a-4d51-b0c9-4b986172e2ac-neutron-metadata-combined-ca-bundle\") pod \"70a69e93-9d4a-4d51-b0c9-4b986172e2ac\" (UID: \"70a69e93-9d4a-4d51-b0c9-4b986172e2ac\") " Jan 26 17:24:16 crc kubenswrapper[4754]: I0126 17:24:16.676853 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70a69e93-9d4a-4d51-b0c9-4b986172e2ac-bootstrap-combined-ca-bundle\") pod \"70a69e93-9d4a-4d51-b0c9-4b986172e2ac\" (UID: \"70a69e93-9d4a-4d51-b0c9-4b986172e2ac\") " Jan 26 17:24:16 crc kubenswrapper[4754]: I0126 17:24:16.676896 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/70a69e93-9d4a-4d51-b0c9-4b986172e2ac-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"70a69e93-9d4a-4d51-b0c9-4b986172e2ac\" (UID: \"70a69e93-9d4a-4d51-b0c9-4b986172e2ac\") " Jan 26 17:24:16 crc kubenswrapper[4754]: I0126 17:24:16.676914 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bwgr2\" (UniqueName: \"kubernetes.io/projected/70a69e93-9d4a-4d51-b0c9-4b986172e2ac-kube-api-access-bwgr2\") pod \"70a69e93-9d4a-4d51-b0c9-4b986172e2ac\" (UID: \"70a69e93-9d4a-4d51-b0c9-4b986172e2ac\") " Jan 26 17:24:16 crc kubenswrapper[4754]: I0126 17:24:16.676941 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70a69e93-9d4a-4d51-b0c9-4b986172e2ac-repo-setup-combined-ca-bundle\") pod \"70a69e93-9d4a-4d51-b0c9-4b986172e2ac\" (UID: \"70a69e93-9d4a-4d51-b0c9-4b986172e2ac\") " Jan 26 17:24:16 crc kubenswrapper[4754]: I0126 17:24:16.682985 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/70a69e93-9d4a-4d51-b0c9-4b986172e2ac-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "70a69e93-9d4a-4d51-b0c9-4b986172e2ac" (UID: "70a69e93-9d4a-4d51-b0c9-4b986172e2ac"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:24:16 crc kubenswrapper[4754]: I0126 17:24:16.683307 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/70a69e93-9d4a-4d51-b0c9-4b986172e2ac-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "70a69e93-9d4a-4d51-b0c9-4b986172e2ac" (UID: "70a69e93-9d4a-4d51-b0c9-4b986172e2ac"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:24:16 crc kubenswrapper[4754]: I0126 17:24:16.685254 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/70a69e93-9d4a-4d51-b0c9-4b986172e2ac-openstack-edpm-ipam-libvirt-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-libvirt-default-certs-0") pod "70a69e93-9d4a-4d51-b0c9-4b986172e2ac" (UID: "70a69e93-9d4a-4d51-b0c9-4b986172e2ac"). InnerVolumeSpecName "openstack-edpm-ipam-libvirt-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:24:16 crc kubenswrapper[4754]: I0126 17:24:16.685845 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/70a69e93-9d4a-4d51-b0c9-4b986172e2ac-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "70a69e93-9d4a-4d51-b0c9-4b986172e2ac" (UID: "70a69e93-9d4a-4d51-b0c9-4b986172e2ac"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:24:16 crc kubenswrapper[4754]: I0126 17:24:16.685880 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/70a69e93-9d4a-4d51-b0c9-4b986172e2ac-openstack-edpm-ipam-ovn-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-ovn-default-certs-0") pod "70a69e93-9d4a-4d51-b0c9-4b986172e2ac" (UID: "70a69e93-9d4a-4d51-b0c9-4b986172e2ac"). InnerVolumeSpecName "openstack-edpm-ipam-ovn-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:24:16 crc kubenswrapper[4754]: I0126 17:24:16.685914 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/70a69e93-9d4a-4d51-b0c9-4b986172e2ac-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "70a69e93-9d4a-4d51-b0c9-4b986172e2ac" (UID: "70a69e93-9d4a-4d51-b0c9-4b986172e2ac"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:24:16 crc kubenswrapper[4754]: I0126 17:24:16.685938 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/70a69e93-9d4a-4d51-b0c9-4b986172e2ac-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "70a69e93-9d4a-4d51-b0c9-4b986172e2ac" (UID: "70a69e93-9d4a-4d51-b0c9-4b986172e2ac"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:24:16 crc kubenswrapper[4754]: I0126 17:24:16.686938 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/70a69e93-9d4a-4d51-b0c9-4b986172e2ac-openstack-edpm-ipam-neutron-metadata-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-neutron-metadata-default-certs-0") pod "70a69e93-9d4a-4d51-b0c9-4b986172e2ac" (UID: "70a69e93-9d4a-4d51-b0c9-4b986172e2ac"). InnerVolumeSpecName "openstack-edpm-ipam-neutron-metadata-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:24:16 crc kubenswrapper[4754]: I0126 17:24:16.687083 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/70a69e93-9d4a-4d51-b0c9-4b986172e2ac-openstack-edpm-ipam-telemetry-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-telemetry-default-certs-0") pod "70a69e93-9d4a-4d51-b0c9-4b986172e2ac" (UID: "70a69e93-9d4a-4d51-b0c9-4b986172e2ac"). InnerVolumeSpecName "openstack-edpm-ipam-telemetry-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:24:16 crc kubenswrapper[4754]: I0126 17:24:16.689640 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/70a69e93-9d4a-4d51-b0c9-4b986172e2ac-kube-api-access-bwgr2" (OuterVolumeSpecName: "kube-api-access-bwgr2") pod "70a69e93-9d4a-4d51-b0c9-4b986172e2ac" (UID: "70a69e93-9d4a-4d51-b0c9-4b986172e2ac"). InnerVolumeSpecName "kube-api-access-bwgr2". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:24:16 crc kubenswrapper[4754]: I0126 17:24:16.690706 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/70a69e93-9d4a-4d51-b0c9-4b986172e2ac-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "70a69e93-9d4a-4d51-b0c9-4b986172e2ac" (UID: "70a69e93-9d4a-4d51-b0c9-4b986172e2ac"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:24:16 crc kubenswrapper[4754]: I0126 17:24:16.695186 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/70a69e93-9d4a-4d51-b0c9-4b986172e2ac-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "70a69e93-9d4a-4d51-b0c9-4b986172e2ac" (UID: "70a69e93-9d4a-4d51-b0c9-4b986172e2ac"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:24:16 crc kubenswrapper[4754]: I0126 17:24:16.727212 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/70a69e93-9d4a-4d51-b0c9-4b986172e2ac-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "70a69e93-9d4a-4d51-b0c9-4b986172e2ac" (UID: "70a69e93-9d4a-4d51-b0c9-4b986172e2ac"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:24:16 crc kubenswrapper[4754]: I0126 17:24:16.730567 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/70a69e93-9d4a-4d51-b0c9-4b986172e2ac-inventory" (OuterVolumeSpecName: "inventory") pod "70a69e93-9d4a-4d51-b0c9-4b986172e2ac" (UID: "70a69e93-9d4a-4d51-b0c9-4b986172e2ac"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:24:16 crc kubenswrapper[4754]: I0126 17:24:16.780886 4754 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70a69e93-9d4a-4d51-b0c9-4b986172e2ac-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 26 17:24:16 crc kubenswrapper[4754]: I0126 17:24:16.780949 4754 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/70a69e93-9d4a-4d51-b0c9-4b986172e2ac-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Jan 26 17:24:16 crc kubenswrapper[4754]: I0126 17:24:16.780966 4754 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/70a69e93-9d4a-4d51-b0c9-4b986172e2ac-openstack-edpm-ipam-libvirt-default-certs-0\") on node \"crc\" DevicePath \"\"" Jan 26 17:24:16 crc kubenswrapper[4754]: I0126 17:24:16.780984 4754 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/70a69e93-9d4a-4d51-b0c9-4b986172e2ac-openstack-edpm-ipam-neutron-metadata-default-certs-0\") on node \"crc\" DevicePath \"\"" Jan 26 17:24:16 crc kubenswrapper[4754]: I0126 17:24:16.781001 4754 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/70a69e93-9d4a-4d51-b0c9-4b986172e2ac-inventory\") on node \"crc\" DevicePath \"\"" Jan 26 17:24:16 crc kubenswrapper[4754]: I0126 17:24:16.781015 4754 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70a69e93-9d4a-4d51-b0c9-4b986172e2ac-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 26 17:24:16 crc kubenswrapper[4754]: I0126 17:24:16.781030 4754 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/70a69e93-9d4a-4d51-b0c9-4b986172e2ac-openstack-edpm-ipam-ovn-default-certs-0\") on node \"crc\" DevicePath \"\"" Jan 26 17:24:16 crc kubenswrapper[4754]: I0126 17:24:16.781044 4754 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70a69e93-9d4a-4d51-b0c9-4b986172e2ac-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 26 17:24:16 crc kubenswrapper[4754]: I0126 17:24:16.781060 4754 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70a69e93-9d4a-4d51-b0c9-4b986172e2ac-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 26 17:24:16 crc kubenswrapper[4754]: I0126 17:24:16.781072 4754 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70a69e93-9d4a-4d51-b0c9-4b986172e2ac-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 26 17:24:16 crc kubenswrapper[4754]: I0126 17:24:16.781085 4754 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70a69e93-9d4a-4d51-b0c9-4b986172e2ac-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 26 17:24:16 crc kubenswrapper[4754]: I0126 17:24:16.781102 4754 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/70a69e93-9d4a-4d51-b0c9-4b986172e2ac-openstack-edpm-ipam-telemetry-default-certs-0\") on node \"crc\" DevicePath \"\"" Jan 26 17:24:16 crc kubenswrapper[4754]: I0126 17:24:16.781116 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bwgr2\" (UniqueName: \"kubernetes.io/projected/70a69e93-9d4a-4d51-b0c9-4b986172e2ac-kube-api-access-bwgr2\") on node \"crc\" DevicePath \"\"" Jan 26 17:24:16 crc kubenswrapper[4754]: I0126 17:24:16.781129 4754 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70a69e93-9d4a-4d51-b0c9-4b986172e2ac-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 26 17:24:17 crc kubenswrapper[4754]: I0126 17:24:17.149002 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8dzsr" event={"ID":"b3b67190-caf6-444e-8c6f-bf2878f37421","Type":"ContainerStarted","Data":"2ccee32755120faf442927650f3394d0e30f0905e041377faaeb986eb455d14b"} Jan 26 17:24:17 crc kubenswrapper[4754]: I0126 17:24:17.150564 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-qpc2q" event={"ID":"70a69e93-9d4a-4d51-b0c9-4b986172e2ac","Type":"ContainerDied","Data":"eff17dc73c46f83f57beb0a851a55b7f7eff7f00da9ca1716c10b954067a253e"} Jan 26 17:24:17 crc kubenswrapper[4754]: I0126 17:24:17.150656 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="eff17dc73c46f83f57beb0a851a55b7f7eff7f00da9ca1716c10b954067a253e" Jan 26 17:24:17 crc kubenswrapper[4754]: I0126 17:24:17.150774 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-qpc2q" Jan 26 17:24:17 crc kubenswrapper[4754]: I0126 17:24:17.166624 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-8dzsr" podStartSLOduration=2.587288826 podStartE2EDuration="5.1665959s" podCreationTimestamp="2026-01-26 17:24:12 +0000 UTC" firstStartedPulling="2026-01-26 17:24:14.10764326 +0000 UTC m=+2220.631823694" lastFinishedPulling="2026-01-26 17:24:16.686950334 +0000 UTC m=+2223.211130768" observedRunningTime="2026-01-26 17:24:17.165190181 +0000 UTC m=+2223.689370625" watchObservedRunningTime="2026-01-26 17:24:17.1665959 +0000 UTC m=+2223.690776334" Jan 26 17:24:17 crc kubenswrapper[4754]: I0126 17:24:17.285477 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-xsq5z"] Jan 26 17:24:17 crc kubenswrapper[4754]: E0126 17:24:17.286155 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="70a69e93-9d4a-4d51-b0c9-4b986172e2ac" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Jan 26 17:24:17 crc kubenswrapper[4754]: I0126 17:24:17.286191 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="70a69e93-9d4a-4d51-b0c9-4b986172e2ac" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Jan 26 17:24:17 crc kubenswrapper[4754]: I0126 17:24:17.286477 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="70a69e93-9d4a-4d51-b0c9-4b986172e2ac" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Jan 26 17:24:17 crc kubenswrapper[4754]: I0126 17:24:17.287800 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xsq5z" Jan 26 17:24:17 crc kubenswrapper[4754]: I0126 17:24:17.291332 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Jan 26 17:24:17 crc kubenswrapper[4754]: I0126 17:24:17.291604 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Jan 26 17:24:17 crc kubenswrapper[4754]: I0126 17:24:17.291400 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-fz7z8" Jan 26 17:24:17 crc kubenswrapper[4754]: I0126 17:24:17.291471 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Jan 26 17:24:17 crc kubenswrapper[4754]: I0126 17:24:17.292087 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Jan 26 17:24:17 crc kubenswrapper[4754]: I0126 17:24:17.313505 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-xsq5z"] Jan 26 17:24:17 crc kubenswrapper[4754]: I0126 17:24:17.398898 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9057d2fb-d9b9-4ce9-b219-9f15c5b7f051-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-xsq5z\" (UID: \"9057d2fb-d9b9-4ce9-b219-9f15c5b7f051\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xsq5z" Jan 26 17:24:17 crc kubenswrapper[4754]: I0126 17:24:17.399467 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/9057d2fb-d9b9-4ce9-b219-9f15c5b7f051-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-xsq5z\" (UID: \"9057d2fb-d9b9-4ce9-b219-9f15c5b7f051\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xsq5z" Jan 26 17:24:17 crc kubenswrapper[4754]: I0126 17:24:17.399552 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/9057d2fb-d9b9-4ce9-b219-9f15c5b7f051-ssh-key-openstack-edpm-ipam\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-xsq5z\" (UID: \"9057d2fb-d9b9-4ce9-b219-9f15c5b7f051\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xsq5z" Jan 26 17:24:17 crc kubenswrapper[4754]: I0126 17:24:17.399603 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9057d2fb-d9b9-4ce9-b219-9f15c5b7f051-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-xsq5z\" (UID: \"9057d2fb-d9b9-4ce9-b219-9f15c5b7f051\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xsq5z" Jan 26 17:24:17 crc kubenswrapper[4754]: I0126 17:24:17.399694 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qbcpj\" (UniqueName: \"kubernetes.io/projected/9057d2fb-d9b9-4ce9-b219-9f15c5b7f051-kube-api-access-qbcpj\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-xsq5z\" (UID: \"9057d2fb-d9b9-4ce9-b219-9f15c5b7f051\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xsq5z" Jan 26 17:24:17 crc kubenswrapper[4754]: I0126 17:24:17.501507 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qbcpj\" (UniqueName: \"kubernetes.io/projected/9057d2fb-d9b9-4ce9-b219-9f15c5b7f051-kube-api-access-qbcpj\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-xsq5z\" (UID: \"9057d2fb-d9b9-4ce9-b219-9f15c5b7f051\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xsq5z" Jan 26 17:24:17 crc kubenswrapper[4754]: I0126 17:24:17.501575 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9057d2fb-d9b9-4ce9-b219-9f15c5b7f051-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-xsq5z\" (UID: \"9057d2fb-d9b9-4ce9-b219-9f15c5b7f051\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xsq5z" Jan 26 17:24:17 crc kubenswrapper[4754]: I0126 17:24:17.501641 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/9057d2fb-d9b9-4ce9-b219-9f15c5b7f051-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-xsq5z\" (UID: \"9057d2fb-d9b9-4ce9-b219-9f15c5b7f051\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xsq5z" Jan 26 17:24:17 crc kubenswrapper[4754]: I0126 17:24:17.501736 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/9057d2fb-d9b9-4ce9-b219-9f15c5b7f051-ssh-key-openstack-edpm-ipam\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-xsq5z\" (UID: \"9057d2fb-d9b9-4ce9-b219-9f15c5b7f051\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xsq5z" Jan 26 17:24:17 crc kubenswrapper[4754]: I0126 17:24:17.501780 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9057d2fb-d9b9-4ce9-b219-9f15c5b7f051-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-xsq5z\" (UID: \"9057d2fb-d9b9-4ce9-b219-9f15c5b7f051\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xsq5z" Jan 26 17:24:17 crc kubenswrapper[4754]: I0126 17:24:17.503009 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/9057d2fb-d9b9-4ce9-b219-9f15c5b7f051-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-xsq5z\" (UID: \"9057d2fb-d9b9-4ce9-b219-9f15c5b7f051\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xsq5z" Jan 26 17:24:17 crc kubenswrapper[4754]: I0126 17:24:17.506099 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9057d2fb-d9b9-4ce9-b219-9f15c5b7f051-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-xsq5z\" (UID: \"9057d2fb-d9b9-4ce9-b219-9f15c5b7f051\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xsq5z" Jan 26 17:24:17 crc kubenswrapper[4754]: I0126 17:24:17.508612 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9057d2fb-d9b9-4ce9-b219-9f15c5b7f051-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-xsq5z\" (UID: \"9057d2fb-d9b9-4ce9-b219-9f15c5b7f051\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xsq5z" Jan 26 17:24:17 crc kubenswrapper[4754]: I0126 17:24:17.509122 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/9057d2fb-d9b9-4ce9-b219-9f15c5b7f051-ssh-key-openstack-edpm-ipam\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-xsq5z\" (UID: \"9057d2fb-d9b9-4ce9-b219-9f15c5b7f051\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xsq5z" Jan 26 17:24:17 crc kubenswrapper[4754]: I0126 17:24:17.522525 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qbcpj\" (UniqueName: \"kubernetes.io/projected/9057d2fb-d9b9-4ce9-b219-9f15c5b7f051-kube-api-access-qbcpj\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-xsq5z\" (UID: \"9057d2fb-d9b9-4ce9-b219-9f15c5b7f051\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xsq5z" Jan 26 17:24:17 crc kubenswrapper[4754]: I0126 17:24:17.609332 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xsq5z" Jan 26 17:24:17 crc kubenswrapper[4754]: I0126 17:24:17.968974 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-xsq5z"] Jan 26 17:24:17 crc kubenswrapper[4754]: W0126 17:24:17.969176 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9057d2fb_d9b9_4ce9_b219_9f15c5b7f051.slice/crio-dc7b9b7e7a388fce90892d09010bca4614d835109cdd85d6e487b54a5448cb46 WatchSource:0}: Error finding container dc7b9b7e7a388fce90892d09010bca4614d835109cdd85d6e487b54a5448cb46: Status 404 returned error can't find the container with id dc7b9b7e7a388fce90892d09010bca4614d835109cdd85d6e487b54a5448cb46 Jan 26 17:24:18 crc kubenswrapper[4754]: I0126 17:24:18.160109 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xsq5z" event={"ID":"9057d2fb-d9b9-4ce9-b219-9f15c5b7f051","Type":"ContainerStarted","Data":"dc7b9b7e7a388fce90892d09010bca4614d835109cdd85d6e487b54a5448cb46"} Jan 26 17:24:19 crc kubenswrapper[4754]: I0126 17:24:19.169489 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xsq5z" event={"ID":"9057d2fb-d9b9-4ce9-b219-9f15c5b7f051","Type":"ContainerStarted","Data":"656a4684ee0fb7d7304b08c0d09c1a7fac9361eb19875310d86548d902b985b9"} Jan 26 17:24:19 crc kubenswrapper[4754]: I0126 17:24:19.194140 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xsq5z" podStartSLOduration=1.696805291 podStartE2EDuration="2.194119907s" podCreationTimestamp="2026-01-26 17:24:17 +0000 UTC" firstStartedPulling="2026-01-26 17:24:17.972978125 +0000 UTC m=+2224.497158559" lastFinishedPulling="2026-01-26 17:24:18.470292741 +0000 UTC m=+2224.994473175" observedRunningTime="2026-01-26 17:24:19.190729492 +0000 UTC m=+2225.714909926" watchObservedRunningTime="2026-01-26 17:24:19.194119907 +0000 UTC m=+2225.718300351" Jan 26 17:24:22 crc kubenswrapper[4754]: I0126 17:24:22.670632 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-8dzsr" Jan 26 17:24:22 crc kubenswrapper[4754]: I0126 17:24:22.671647 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-8dzsr" Jan 26 17:24:22 crc kubenswrapper[4754]: I0126 17:24:22.718233 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-8dzsr" Jan 26 17:24:23 crc kubenswrapper[4754]: I0126 17:24:23.249137 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-8dzsr" Jan 26 17:24:23 crc kubenswrapper[4754]: I0126 17:24:23.312284 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-8dzsr"] Jan 26 17:24:25 crc kubenswrapper[4754]: I0126 17:24:25.219496 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-8dzsr" podUID="b3b67190-caf6-444e-8c6f-bf2878f37421" containerName="registry-server" containerID="cri-o://2ccee32755120faf442927650f3394d0e30f0905e041377faaeb986eb455d14b" gracePeriod=2 Jan 26 17:24:25 crc kubenswrapper[4754]: I0126 17:24:25.660078 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8dzsr" Jan 26 17:24:25 crc kubenswrapper[4754]: I0126 17:24:25.781335 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b3b67190-caf6-444e-8c6f-bf2878f37421-catalog-content\") pod \"b3b67190-caf6-444e-8c6f-bf2878f37421\" (UID: \"b3b67190-caf6-444e-8c6f-bf2878f37421\") " Jan 26 17:24:25 crc kubenswrapper[4754]: I0126 17:24:25.781637 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h9kkw\" (UniqueName: \"kubernetes.io/projected/b3b67190-caf6-444e-8c6f-bf2878f37421-kube-api-access-h9kkw\") pod \"b3b67190-caf6-444e-8c6f-bf2878f37421\" (UID: \"b3b67190-caf6-444e-8c6f-bf2878f37421\") " Jan 26 17:24:25 crc kubenswrapper[4754]: I0126 17:24:25.781787 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b3b67190-caf6-444e-8c6f-bf2878f37421-utilities\") pod \"b3b67190-caf6-444e-8c6f-bf2878f37421\" (UID: \"b3b67190-caf6-444e-8c6f-bf2878f37421\") " Jan 26 17:24:25 crc kubenswrapper[4754]: I0126 17:24:25.782860 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b3b67190-caf6-444e-8c6f-bf2878f37421-utilities" (OuterVolumeSpecName: "utilities") pod "b3b67190-caf6-444e-8c6f-bf2878f37421" (UID: "b3b67190-caf6-444e-8c6f-bf2878f37421"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:24:25 crc kubenswrapper[4754]: I0126 17:24:25.787686 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b3b67190-caf6-444e-8c6f-bf2878f37421-kube-api-access-h9kkw" (OuterVolumeSpecName: "kube-api-access-h9kkw") pod "b3b67190-caf6-444e-8c6f-bf2878f37421" (UID: "b3b67190-caf6-444e-8c6f-bf2878f37421"). InnerVolumeSpecName "kube-api-access-h9kkw". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:24:25 crc kubenswrapper[4754]: I0126 17:24:25.815271 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b3b67190-caf6-444e-8c6f-bf2878f37421-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b3b67190-caf6-444e-8c6f-bf2878f37421" (UID: "b3b67190-caf6-444e-8c6f-bf2878f37421"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:24:25 crc kubenswrapper[4754]: I0126 17:24:25.884270 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h9kkw\" (UniqueName: \"kubernetes.io/projected/b3b67190-caf6-444e-8c6f-bf2878f37421-kube-api-access-h9kkw\") on node \"crc\" DevicePath \"\"" Jan 26 17:24:25 crc kubenswrapper[4754]: I0126 17:24:25.884318 4754 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b3b67190-caf6-444e-8c6f-bf2878f37421-utilities\") on node \"crc\" DevicePath \"\"" Jan 26 17:24:25 crc kubenswrapper[4754]: I0126 17:24:25.884334 4754 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b3b67190-caf6-444e-8c6f-bf2878f37421-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 26 17:24:26 crc kubenswrapper[4754]: I0126 17:24:26.229797 4754 generic.go:334] "Generic (PLEG): container finished" podID="b3b67190-caf6-444e-8c6f-bf2878f37421" containerID="2ccee32755120faf442927650f3394d0e30f0905e041377faaeb986eb455d14b" exitCode=0 Jan 26 17:24:26 crc kubenswrapper[4754]: I0126 17:24:26.229854 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8dzsr" event={"ID":"b3b67190-caf6-444e-8c6f-bf2878f37421","Type":"ContainerDied","Data":"2ccee32755120faf442927650f3394d0e30f0905e041377faaeb986eb455d14b"} Jan 26 17:24:26 crc kubenswrapper[4754]: I0126 17:24:26.229880 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8dzsr" event={"ID":"b3b67190-caf6-444e-8c6f-bf2878f37421","Type":"ContainerDied","Data":"75e8fff32dff7f9e577931c33dd477e4b11fc1685612e4a7cb47df631cc8e538"} Jan 26 17:24:26 crc kubenswrapper[4754]: I0126 17:24:26.229897 4754 scope.go:117] "RemoveContainer" containerID="2ccee32755120faf442927650f3394d0e30f0905e041377faaeb986eb455d14b" Jan 26 17:24:26 crc kubenswrapper[4754]: I0126 17:24:26.230018 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8dzsr" Jan 26 17:24:26 crc kubenswrapper[4754]: I0126 17:24:26.272474 4754 scope.go:117] "RemoveContainer" containerID="82d46e7ae399a5312b75aeea432e3fc1fb28917c134d936b9a47421a510c4406" Jan 26 17:24:26 crc kubenswrapper[4754]: I0126 17:24:26.278731 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-8dzsr"] Jan 26 17:24:26 crc kubenswrapper[4754]: I0126 17:24:26.287902 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-8dzsr"] Jan 26 17:24:26 crc kubenswrapper[4754]: I0126 17:24:26.301278 4754 scope.go:117] "RemoveContainer" containerID="0c1e9334cca604a6378d6354f3a5844bdf909694b48bdf7667a035164923f3f1" Jan 26 17:24:26 crc kubenswrapper[4754]: I0126 17:24:26.342377 4754 scope.go:117] "RemoveContainer" containerID="2ccee32755120faf442927650f3394d0e30f0905e041377faaeb986eb455d14b" Jan 26 17:24:26 crc kubenswrapper[4754]: E0126 17:24:26.342944 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2ccee32755120faf442927650f3394d0e30f0905e041377faaeb986eb455d14b\": container with ID starting with 2ccee32755120faf442927650f3394d0e30f0905e041377faaeb986eb455d14b not found: ID does not exist" containerID="2ccee32755120faf442927650f3394d0e30f0905e041377faaeb986eb455d14b" Jan 26 17:24:26 crc kubenswrapper[4754]: I0126 17:24:26.343002 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2ccee32755120faf442927650f3394d0e30f0905e041377faaeb986eb455d14b"} err="failed to get container status \"2ccee32755120faf442927650f3394d0e30f0905e041377faaeb986eb455d14b\": rpc error: code = NotFound desc = could not find container \"2ccee32755120faf442927650f3394d0e30f0905e041377faaeb986eb455d14b\": container with ID starting with 2ccee32755120faf442927650f3394d0e30f0905e041377faaeb986eb455d14b not found: ID does not exist" Jan 26 17:24:26 crc kubenswrapper[4754]: I0126 17:24:26.343030 4754 scope.go:117] "RemoveContainer" containerID="82d46e7ae399a5312b75aeea432e3fc1fb28917c134d936b9a47421a510c4406" Jan 26 17:24:26 crc kubenswrapper[4754]: E0126 17:24:26.343346 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"82d46e7ae399a5312b75aeea432e3fc1fb28917c134d936b9a47421a510c4406\": container with ID starting with 82d46e7ae399a5312b75aeea432e3fc1fb28917c134d936b9a47421a510c4406 not found: ID does not exist" containerID="82d46e7ae399a5312b75aeea432e3fc1fb28917c134d936b9a47421a510c4406" Jan 26 17:24:26 crc kubenswrapper[4754]: I0126 17:24:26.343367 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"82d46e7ae399a5312b75aeea432e3fc1fb28917c134d936b9a47421a510c4406"} err="failed to get container status \"82d46e7ae399a5312b75aeea432e3fc1fb28917c134d936b9a47421a510c4406\": rpc error: code = NotFound desc = could not find container \"82d46e7ae399a5312b75aeea432e3fc1fb28917c134d936b9a47421a510c4406\": container with ID starting with 82d46e7ae399a5312b75aeea432e3fc1fb28917c134d936b9a47421a510c4406 not found: ID does not exist" Jan 26 17:24:26 crc kubenswrapper[4754]: I0126 17:24:26.343379 4754 scope.go:117] "RemoveContainer" containerID="0c1e9334cca604a6378d6354f3a5844bdf909694b48bdf7667a035164923f3f1" Jan 26 17:24:26 crc kubenswrapper[4754]: E0126 17:24:26.343744 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0c1e9334cca604a6378d6354f3a5844bdf909694b48bdf7667a035164923f3f1\": container with ID starting with 0c1e9334cca604a6378d6354f3a5844bdf909694b48bdf7667a035164923f3f1 not found: ID does not exist" containerID="0c1e9334cca604a6378d6354f3a5844bdf909694b48bdf7667a035164923f3f1" Jan 26 17:24:26 crc kubenswrapper[4754]: I0126 17:24:26.343762 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0c1e9334cca604a6378d6354f3a5844bdf909694b48bdf7667a035164923f3f1"} err="failed to get container status \"0c1e9334cca604a6378d6354f3a5844bdf909694b48bdf7667a035164923f3f1\": rpc error: code = NotFound desc = could not find container \"0c1e9334cca604a6378d6354f3a5844bdf909694b48bdf7667a035164923f3f1\": container with ID starting with 0c1e9334cca604a6378d6354f3a5844bdf909694b48bdf7667a035164923f3f1 not found: ID does not exist" Jan 26 17:24:27 crc kubenswrapper[4754]: I0126 17:24:27.778710 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b3b67190-caf6-444e-8c6f-bf2878f37421" path="/var/lib/kubelet/pods/b3b67190-caf6-444e-8c6f-bf2878f37421/volumes" Jan 26 17:24:37 crc kubenswrapper[4754]: I0126 17:24:37.129382 4754 patch_prober.go:28] interesting pod/machine-config-daemon-x65wv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 26 17:24:37 crc kubenswrapper[4754]: I0126 17:24:37.129946 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 26 17:24:37 crc kubenswrapper[4754]: I0126 17:24:37.129987 4754 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" Jan 26 17:24:37 crc kubenswrapper[4754]: I0126 17:24:37.130649 4754 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"00e15c82b01f7eb3c11968e396c0748604b39b71891170fde13a876ab0a851b3"} pod="openshift-machine-config-operator/machine-config-daemon-x65wv" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 26 17:24:37 crc kubenswrapper[4754]: I0126 17:24:37.130729 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" containerName="machine-config-daemon" containerID="cri-o://00e15c82b01f7eb3c11968e396c0748604b39b71891170fde13a876ab0a851b3" gracePeriod=600 Jan 26 17:24:37 crc kubenswrapper[4754]: E0126 17:24:37.757404 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x65wv_openshift-machine-config-operator(8c3718a4-f354-4284-92e0-fdfb45a692bd)\"" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" Jan 26 17:24:38 crc kubenswrapper[4754]: I0126 17:24:38.338097 4754 generic.go:334] "Generic (PLEG): container finished" podID="8c3718a4-f354-4284-92e0-fdfb45a692bd" containerID="00e15c82b01f7eb3c11968e396c0748604b39b71891170fde13a876ab0a851b3" exitCode=0 Jan 26 17:24:38 crc kubenswrapper[4754]: I0126 17:24:38.338144 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" event={"ID":"8c3718a4-f354-4284-92e0-fdfb45a692bd","Type":"ContainerDied","Data":"00e15c82b01f7eb3c11968e396c0748604b39b71891170fde13a876ab0a851b3"} Jan 26 17:24:38 crc kubenswrapper[4754]: I0126 17:24:38.338180 4754 scope.go:117] "RemoveContainer" containerID="1f2fc5d329d9c26d0c14a843ed658e96c72f72ea44b0a9c9d8e76e987f4603de" Jan 26 17:24:38 crc kubenswrapper[4754]: I0126 17:24:38.338921 4754 scope.go:117] "RemoveContainer" containerID="00e15c82b01f7eb3c11968e396c0748604b39b71891170fde13a876ab0a851b3" Jan 26 17:24:38 crc kubenswrapper[4754]: E0126 17:24:38.339204 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x65wv_openshift-machine-config-operator(8c3718a4-f354-4284-92e0-fdfb45a692bd)\"" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" Jan 26 17:24:51 crc kubenswrapper[4754]: I0126 17:24:51.768092 4754 scope.go:117] "RemoveContainer" containerID="00e15c82b01f7eb3c11968e396c0748604b39b71891170fde13a876ab0a851b3" Jan 26 17:24:51 crc kubenswrapper[4754]: E0126 17:24:51.769502 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x65wv_openshift-machine-config-operator(8c3718a4-f354-4284-92e0-fdfb45a692bd)\"" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" Jan 26 17:25:05 crc kubenswrapper[4754]: I0126 17:25:05.767607 4754 scope.go:117] "RemoveContainer" containerID="00e15c82b01f7eb3c11968e396c0748604b39b71891170fde13a876ab0a851b3" Jan 26 17:25:05 crc kubenswrapper[4754]: E0126 17:25:05.768519 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x65wv_openshift-machine-config-operator(8c3718a4-f354-4284-92e0-fdfb45a692bd)\"" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" Jan 26 17:25:20 crc kubenswrapper[4754]: I0126 17:25:20.767258 4754 scope.go:117] "RemoveContainer" containerID="00e15c82b01f7eb3c11968e396c0748604b39b71891170fde13a876ab0a851b3" Jan 26 17:25:20 crc kubenswrapper[4754]: E0126 17:25:20.768067 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x65wv_openshift-machine-config-operator(8c3718a4-f354-4284-92e0-fdfb45a692bd)\"" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" Jan 26 17:25:23 crc kubenswrapper[4754]: I0126 17:25:23.730771 4754 generic.go:334] "Generic (PLEG): container finished" podID="9057d2fb-d9b9-4ce9-b219-9f15c5b7f051" containerID="656a4684ee0fb7d7304b08c0d09c1a7fac9361eb19875310d86548d902b985b9" exitCode=0 Jan 26 17:25:23 crc kubenswrapper[4754]: I0126 17:25:23.730857 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xsq5z" event={"ID":"9057d2fb-d9b9-4ce9-b219-9f15c5b7f051","Type":"ContainerDied","Data":"656a4684ee0fb7d7304b08c0d09c1a7fac9361eb19875310d86548d902b985b9"} Jan 26 17:25:25 crc kubenswrapper[4754]: I0126 17:25:25.150922 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xsq5z" Jan 26 17:25:25 crc kubenswrapper[4754]: I0126 17:25:25.249279 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9057d2fb-d9b9-4ce9-b219-9f15c5b7f051-ovn-combined-ca-bundle\") pod \"9057d2fb-d9b9-4ce9-b219-9f15c5b7f051\" (UID: \"9057d2fb-d9b9-4ce9-b219-9f15c5b7f051\") " Jan 26 17:25:25 crc kubenswrapper[4754]: I0126 17:25:25.249354 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/9057d2fb-d9b9-4ce9-b219-9f15c5b7f051-ssh-key-openstack-edpm-ipam\") pod \"9057d2fb-d9b9-4ce9-b219-9f15c5b7f051\" (UID: \"9057d2fb-d9b9-4ce9-b219-9f15c5b7f051\") " Jan 26 17:25:25 crc kubenswrapper[4754]: I0126 17:25:25.249390 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/9057d2fb-d9b9-4ce9-b219-9f15c5b7f051-ovncontroller-config-0\") pod \"9057d2fb-d9b9-4ce9-b219-9f15c5b7f051\" (UID: \"9057d2fb-d9b9-4ce9-b219-9f15c5b7f051\") " Jan 26 17:25:25 crc kubenswrapper[4754]: I0126 17:25:25.249469 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9057d2fb-d9b9-4ce9-b219-9f15c5b7f051-inventory\") pod \"9057d2fb-d9b9-4ce9-b219-9f15c5b7f051\" (UID: \"9057d2fb-d9b9-4ce9-b219-9f15c5b7f051\") " Jan 26 17:25:25 crc kubenswrapper[4754]: I0126 17:25:25.249502 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qbcpj\" (UniqueName: \"kubernetes.io/projected/9057d2fb-d9b9-4ce9-b219-9f15c5b7f051-kube-api-access-qbcpj\") pod \"9057d2fb-d9b9-4ce9-b219-9f15c5b7f051\" (UID: \"9057d2fb-d9b9-4ce9-b219-9f15c5b7f051\") " Jan 26 17:25:25 crc kubenswrapper[4754]: I0126 17:25:25.255209 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9057d2fb-d9b9-4ce9-b219-9f15c5b7f051-kube-api-access-qbcpj" (OuterVolumeSpecName: "kube-api-access-qbcpj") pod "9057d2fb-d9b9-4ce9-b219-9f15c5b7f051" (UID: "9057d2fb-d9b9-4ce9-b219-9f15c5b7f051"). InnerVolumeSpecName "kube-api-access-qbcpj". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:25:25 crc kubenswrapper[4754]: I0126 17:25:25.255345 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9057d2fb-d9b9-4ce9-b219-9f15c5b7f051-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "9057d2fb-d9b9-4ce9-b219-9f15c5b7f051" (UID: "9057d2fb-d9b9-4ce9-b219-9f15c5b7f051"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:25:25 crc kubenswrapper[4754]: I0126 17:25:25.277074 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9057d2fb-d9b9-4ce9-b219-9f15c5b7f051-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "9057d2fb-d9b9-4ce9-b219-9f15c5b7f051" (UID: "9057d2fb-d9b9-4ce9-b219-9f15c5b7f051"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:25:25 crc kubenswrapper[4754]: I0126 17:25:25.279565 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9057d2fb-d9b9-4ce9-b219-9f15c5b7f051-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "9057d2fb-d9b9-4ce9-b219-9f15c5b7f051" (UID: "9057d2fb-d9b9-4ce9-b219-9f15c5b7f051"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:25:25 crc kubenswrapper[4754]: I0126 17:25:25.280241 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9057d2fb-d9b9-4ce9-b219-9f15c5b7f051-inventory" (OuterVolumeSpecName: "inventory") pod "9057d2fb-d9b9-4ce9-b219-9f15c5b7f051" (UID: "9057d2fb-d9b9-4ce9-b219-9f15c5b7f051"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:25:25 crc kubenswrapper[4754]: I0126 17:25:25.351536 4754 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/9057d2fb-d9b9-4ce9-b219-9f15c5b7f051-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Jan 26 17:25:25 crc kubenswrapper[4754]: I0126 17:25:25.351578 4754 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9057d2fb-d9b9-4ce9-b219-9f15c5b7f051-inventory\") on node \"crc\" DevicePath \"\"" Jan 26 17:25:25 crc kubenswrapper[4754]: I0126 17:25:25.351586 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qbcpj\" (UniqueName: \"kubernetes.io/projected/9057d2fb-d9b9-4ce9-b219-9f15c5b7f051-kube-api-access-qbcpj\") on node \"crc\" DevicePath \"\"" Jan 26 17:25:25 crc kubenswrapper[4754]: I0126 17:25:25.351595 4754 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9057d2fb-d9b9-4ce9-b219-9f15c5b7f051-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 26 17:25:25 crc kubenswrapper[4754]: I0126 17:25:25.351604 4754 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/9057d2fb-d9b9-4ce9-b219-9f15c5b7f051-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Jan 26 17:25:25 crc kubenswrapper[4754]: I0126 17:25:25.751903 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xsq5z" event={"ID":"9057d2fb-d9b9-4ce9-b219-9f15c5b7f051","Type":"ContainerDied","Data":"dc7b9b7e7a388fce90892d09010bca4614d835109cdd85d6e487b54a5448cb46"} Jan 26 17:25:25 crc kubenswrapper[4754]: I0126 17:25:25.751950 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dc7b9b7e7a388fce90892d09010bca4614d835109cdd85d6e487b54a5448cb46" Jan 26 17:25:25 crc kubenswrapper[4754]: I0126 17:25:25.751958 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xsq5z" Jan 26 17:25:25 crc kubenswrapper[4754]: I0126 17:25:25.855073 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wljzd"] Jan 26 17:25:25 crc kubenswrapper[4754]: E0126 17:25:25.855441 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3b67190-caf6-444e-8c6f-bf2878f37421" containerName="extract-utilities" Jan 26 17:25:25 crc kubenswrapper[4754]: I0126 17:25:25.855459 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3b67190-caf6-444e-8c6f-bf2878f37421" containerName="extract-utilities" Jan 26 17:25:25 crc kubenswrapper[4754]: E0126 17:25:25.855488 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9057d2fb-d9b9-4ce9-b219-9f15c5b7f051" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Jan 26 17:25:25 crc kubenswrapper[4754]: I0126 17:25:25.855495 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="9057d2fb-d9b9-4ce9-b219-9f15c5b7f051" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Jan 26 17:25:25 crc kubenswrapper[4754]: E0126 17:25:25.855510 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3b67190-caf6-444e-8c6f-bf2878f37421" containerName="registry-server" Jan 26 17:25:25 crc kubenswrapper[4754]: I0126 17:25:25.855515 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3b67190-caf6-444e-8c6f-bf2878f37421" containerName="registry-server" Jan 26 17:25:25 crc kubenswrapper[4754]: E0126 17:25:25.855523 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3b67190-caf6-444e-8c6f-bf2878f37421" containerName="extract-content" Jan 26 17:25:25 crc kubenswrapper[4754]: I0126 17:25:25.855529 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3b67190-caf6-444e-8c6f-bf2878f37421" containerName="extract-content" Jan 26 17:25:25 crc kubenswrapper[4754]: I0126 17:25:25.855710 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="b3b67190-caf6-444e-8c6f-bf2878f37421" containerName="registry-server" Jan 26 17:25:25 crc kubenswrapper[4754]: I0126 17:25:25.855728 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="9057d2fb-d9b9-4ce9-b219-9f15c5b7f051" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Jan 26 17:25:25 crc kubenswrapper[4754]: I0126 17:25:25.856317 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wljzd" Jan 26 17:25:25 crc kubenswrapper[4754]: I0126 17:25:25.858447 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Jan 26 17:25:25 crc kubenswrapper[4754]: I0126 17:25:25.858545 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Jan 26 17:25:25 crc kubenswrapper[4754]: I0126 17:25:25.859413 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Jan 26 17:25:25 crc kubenswrapper[4754]: I0126 17:25:25.861964 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Jan 26 17:25:25 crc kubenswrapper[4754]: I0126 17:25:25.861987 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Jan 26 17:25:25 crc kubenswrapper[4754]: I0126 17:25:25.861988 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-fz7z8" Jan 26 17:25:25 crc kubenswrapper[4754]: I0126 17:25:25.872462 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wljzd"] Jan 26 17:25:25 crc kubenswrapper[4754]: I0126 17:25:25.965146 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7qcjl\" (UniqueName: \"kubernetes.io/projected/18417e65-9c20-4b5e-96df-b7b79c67433a-kube-api-access-7qcjl\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-wljzd\" (UID: \"18417e65-9c20-4b5e-96df-b7b79c67433a\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wljzd" Jan 26 17:25:25 crc kubenswrapper[4754]: I0126 17:25:25.965221 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18417e65-9c20-4b5e-96df-b7b79c67433a-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-wljzd\" (UID: \"18417e65-9c20-4b5e-96df-b7b79c67433a\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wljzd" Jan 26 17:25:25 crc kubenswrapper[4754]: I0126 17:25:25.965274 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/18417e65-9c20-4b5e-96df-b7b79c67433a-ssh-key-openstack-edpm-ipam\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-wljzd\" (UID: \"18417e65-9c20-4b5e-96df-b7b79c67433a\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wljzd" Jan 26 17:25:25 crc kubenswrapper[4754]: I0126 17:25:25.965310 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/18417e65-9c20-4b5e-96df-b7b79c67433a-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-wljzd\" (UID: \"18417e65-9c20-4b5e-96df-b7b79c67433a\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wljzd" Jan 26 17:25:25 crc kubenswrapper[4754]: I0126 17:25:25.965333 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/18417e65-9c20-4b5e-96df-b7b79c67433a-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-wljzd\" (UID: \"18417e65-9c20-4b5e-96df-b7b79c67433a\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wljzd" Jan 26 17:25:25 crc kubenswrapper[4754]: I0126 17:25:25.965355 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/18417e65-9c20-4b5e-96df-b7b79c67433a-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-wljzd\" (UID: \"18417e65-9c20-4b5e-96df-b7b79c67433a\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wljzd" Jan 26 17:25:26 crc kubenswrapper[4754]: I0126 17:25:26.066334 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/18417e65-9c20-4b5e-96df-b7b79c67433a-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-wljzd\" (UID: \"18417e65-9c20-4b5e-96df-b7b79c67433a\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wljzd" Jan 26 17:25:26 crc kubenswrapper[4754]: I0126 17:25:26.066798 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/18417e65-9c20-4b5e-96df-b7b79c67433a-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-wljzd\" (UID: \"18417e65-9c20-4b5e-96df-b7b79c67433a\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wljzd" Jan 26 17:25:26 crc kubenswrapper[4754]: I0126 17:25:26.067514 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/18417e65-9c20-4b5e-96df-b7b79c67433a-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-wljzd\" (UID: \"18417e65-9c20-4b5e-96df-b7b79c67433a\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wljzd" Jan 26 17:25:26 crc kubenswrapper[4754]: I0126 17:25:26.067856 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7qcjl\" (UniqueName: \"kubernetes.io/projected/18417e65-9c20-4b5e-96df-b7b79c67433a-kube-api-access-7qcjl\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-wljzd\" (UID: \"18417e65-9c20-4b5e-96df-b7b79c67433a\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wljzd" Jan 26 17:25:26 crc kubenswrapper[4754]: I0126 17:25:26.068013 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18417e65-9c20-4b5e-96df-b7b79c67433a-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-wljzd\" (UID: \"18417e65-9c20-4b5e-96df-b7b79c67433a\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wljzd" Jan 26 17:25:26 crc kubenswrapper[4754]: I0126 17:25:26.068202 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/18417e65-9c20-4b5e-96df-b7b79c67433a-ssh-key-openstack-edpm-ipam\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-wljzd\" (UID: \"18417e65-9c20-4b5e-96df-b7b79c67433a\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wljzd" Jan 26 17:25:26 crc kubenswrapper[4754]: I0126 17:25:26.072102 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/18417e65-9c20-4b5e-96df-b7b79c67433a-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-wljzd\" (UID: \"18417e65-9c20-4b5e-96df-b7b79c67433a\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wljzd" Jan 26 17:25:26 crc kubenswrapper[4754]: I0126 17:25:26.072510 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/18417e65-9c20-4b5e-96df-b7b79c67433a-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-wljzd\" (UID: \"18417e65-9c20-4b5e-96df-b7b79c67433a\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wljzd" Jan 26 17:25:26 crc kubenswrapper[4754]: I0126 17:25:26.073424 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18417e65-9c20-4b5e-96df-b7b79c67433a-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-wljzd\" (UID: \"18417e65-9c20-4b5e-96df-b7b79c67433a\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wljzd" Jan 26 17:25:26 crc kubenswrapper[4754]: I0126 17:25:26.073526 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/18417e65-9c20-4b5e-96df-b7b79c67433a-ssh-key-openstack-edpm-ipam\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-wljzd\" (UID: \"18417e65-9c20-4b5e-96df-b7b79c67433a\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wljzd" Jan 26 17:25:26 crc kubenswrapper[4754]: I0126 17:25:26.074911 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/18417e65-9c20-4b5e-96df-b7b79c67433a-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-wljzd\" (UID: \"18417e65-9c20-4b5e-96df-b7b79c67433a\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wljzd" Jan 26 17:25:26 crc kubenswrapper[4754]: I0126 17:25:26.086377 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7qcjl\" (UniqueName: \"kubernetes.io/projected/18417e65-9c20-4b5e-96df-b7b79c67433a-kube-api-access-7qcjl\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-wljzd\" (UID: \"18417e65-9c20-4b5e-96df-b7b79c67433a\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wljzd" Jan 26 17:25:26 crc kubenswrapper[4754]: I0126 17:25:26.174543 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wljzd" Jan 26 17:25:26 crc kubenswrapper[4754]: I0126 17:25:26.680272 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wljzd"] Jan 26 17:25:26 crc kubenswrapper[4754]: I0126 17:25:26.760733 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wljzd" event={"ID":"18417e65-9c20-4b5e-96df-b7b79c67433a","Type":"ContainerStarted","Data":"d5244bb9f79b8359dd2a48258ce8b656fcb6dc1492586967285422c49e355547"} Jan 26 17:25:27 crc kubenswrapper[4754]: I0126 17:25:27.777355 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wljzd" event={"ID":"18417e65-9c20-4b5e-96df-b7b79c67433a","Type":"ContainerStarted","Data":"1f2efeefc9cd8710cd38763d7776404db47adebb63fd3f67d56ee5227863ec8a"} Jan 26 17:25:27 crc kubenswrapper[4754]: I0126 17:25:27.792276 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wljzd" podStartSLOduration=2.2617917260000002 podStartE2EDuration="2.792258621s" podCreationTimestamp="2026-01-26 17:25:25 +0000 UTC" firstStartedPulling="2026-01-26 17:25:26.689343336 +0000 UTC m=+2293.213523770" lastFinishedPulling="2026-01-26 17:25:27.219810231 +0000 UTC m=+2293.743990665" observedRunningTime="2026-01-26 17:25:27.785144303 +0000 UTC m=+2294.309324737" watchObservedRunningTime="2026-01-26 17:25:27.792258621 +0000 UTC m=+2294.316439055" Jan 26 17:25:35 crc kubenswrapper[4754]: I0126 17:25:35.766938 4754 scope.go:117] "RemoveContainer" containerID="00e15c82b01f7eb3c11968e396c0748604b39b71891170fde13a876ab0a851b3" Jan 26 17:25:35 crc kubenswrapper[4754]: E0126 17:25:35.767810 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x65wv_openshift-machine-config-operator(8c3718a4-f354-4284-92e0-fdfb45a692bd)\"" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" Jan 26 17:25:48 crc kubenswrapper[4754]: I0126 17:25:48.768271 4754 scope.go:117] "RemoveContainer" containerID="00e15c82b01f7eb3c11968e396c0748604b39b71891170fde13a876ab0a851b3" Jan 26 17:25:48 crc kubenswrapper[4754]: E0126 17:25:48.769095 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x65wv_openshift-machine-config-operator(8c3718a4-f354-4284-92e0-fdfb45a692bd)\"" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" Jan 26 17:26:00 crc kubenswrapper[4754]: I0126 17:26:00.768349 4754 scope.go:117] "RemoveContainer" containerID="00e15c82b01f7eb3c11968e396c0748604b39b71891170fde13a876ab0a851b3" Jan 26 17:26:00 crc kubenswrapper[4754]: E0126 17:26:00.769112 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x65wv_openshift-machine-config-operator(8c3718a4-f354-4284-92e0-fdfb45a692bd)\"" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" Jan 26 17:26:14 crc kubenswrapper[4754]: I0126 17:26:14.768039 4754 scope.go:117] "RemoveContainer" containerID="00e15c82b01f7eb3c11968e396c0748604b39b71891170fde13a876ab0a851b3" Jan 26 17:26:14 crc kubenswrapper[4754]: E0126 17:26:14.768899 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x65wv_openshift-machine-config-operator(8c3718a4-f354-4284-92e0-fdfb45a692bd)\"" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" Jan 26 17:26:18 crc kubenswrapper[4754]: I0126 17:26:18.241518 4754 generic.go:334] "Generic (PLEG): container finished" podID="18417e65-9c20-4b5e-96df-b7b79c67433a" containerID="1f2efeefc9cd8710cd38763d7776404db47adebb63fd3f67d56ee5227863ec8a" exitCode=0 Jan 26 17:26:18 crc kubenswrapper[4754]: I0126 17:26:18.241727 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wljzd" event={"ID":"18417e65-9c20-4b5e-96df-b7b79c67433a","Type":"ContainerDied","Data":"1f2efeefc9cd8710cd38763d7776404db47adebb63fd3f67d56ee5227863ec8a"} Jan 26 17:26:19 crc kubenswrapper[4754]: I0126 17:26:19.713911 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wljzd" Jan 26 17:26:19 crc kubenswrapper[4754]: I0126 17:26:19.885646 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/18417e65-9c20-4b5e-96df-b7b79c67433a-neutron-ovn-metadata-agent-neutron-config-0\") pod \"18417e65-9c20-4b5e-96df-b7b79c67433a\" (UID: \"18417e65-9c20-4b5e-96df-b7b79c67433a\") " Jan 26 17:26:19 crc kubenswrapper[4754]: I0126 17:26:19.885750 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/18417e65-9c20-4b5e-96df-b7b79c67433a-nova-metadata-neutron-config-0\") pod \"18417e65-9c20-4b5e-96df-b7b79c67433a\" (UID: \"18417e65-9c20-4b5e-96df-b7b79c67433a\") " Jan 26 17:26:19 crc kubenswrapper[4754]: I0126 17:26:19.885822 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7qcjl\" (UniqueName: \"kubernetes.io/projected/18417e65-9c20-4b5e-96df-b7b79c67433a-kube-api-access-7qcjl\") pod \"18417e65-9c20-4b5e-96df-b7b79c67433a\" (UID: \"18417e65-9c20-4b5e-96df-b7b79c67433a\") " Jan 26 17:26:19 crc kubenswrapper[4754]: I0126 17:26:19.885911 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/18417e65-9c20-4b5e-96df-b7b79c67433a-ssh-key-openstack-edpm-ipam\") pod \"18417e65-9c20-4b5e-96df-b7b79c67433a\" (UID: \"18417e65-9c20-4b5e-96df-b7b79c67433a\") " Jan 26 17:26:19 crc kubenswrapper[4754]: I0126 17:26:19.886000 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/18417e65-9c20-4b5e-96df-b7b79c67433a-inventory\") pod \"18417e65-9c20-4b5e-96df-b7b79c67433a\" (UID: \"18417e65-9c20-4b5e-96df-b7b79c67433a\") " Jan 26 17:26:19 crc kubenswrapper[4754]: I0126 17:26:19.886059 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18417e65-9c20-4b5e-96df-b7b79c67433a-neutron-metadata-combined-ca-bundle\") pod \"18417e65-9c20-4b5e-96df-b7b79c67433a\" (UID: \"18417e65-9c20-4b5e-96df-b7b79c67433a\") " Jan 26 17:26:19 crc kubenswrapper[4754]: I0126 17:26:19.893127 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/18417e65-9c20-4b5e-96df-b7b79c67433a-kube-api-access-7qcjl" (OuterVolumeSpecName: "kube-api-access-7qcjl") pod "18417e65-9c20-4b5e-96df-b7b79c67433a" (UID: "18417e65-9c20-4b5e-96df-b7b79c67433a"). InnerVolumeSpecName "kube-api-access-7qcjl". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:26:19 crc kubenswrapper[4754]: I0126 17:26:19.894147 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/18417e65-9c20-4b5e-96df-b7b79c67433a-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "18417e65-9c20-4b5e-96df-b7b79c67433a" (UID: "18417e65-9c20-4b5e-96df-b7b79c67433a"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:26:19 crc kubenswrapper[4754]: I0126 17:26:19.915204 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/18417e65-9c20-4b5e-96df-b7b79c67433a-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "18417e65-9c20-4b5e-96df-b7b79c67433a" (UID: "18417e65-9c20-4b5e-96df-b7b79c67433a"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:26:19 crc kubenswrapper[4754]: I0126 17:26:19.916635 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/18417e65-9c20-4b5e-96df-b7b79c67433a-inventory" (OuterVolumeSpecName: "inventory") pod "18417e65-9c20-4b5e-96df-b7b79c67433a" (UID: "18417e65-9c20-4b5e-96df-b7b79c67433a"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:26:19 crc kubenswrapper[4754]: I0126 17:26:19.924982 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/18417e65-9c20-4b5e-96df-b7b79c67433a-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "18417e65-9c20-4b5e-96df-b7b79c67433a" (UID: "18417e65-9c20-4b5e-96df-b7b79c67433a"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:26:19 crc kubenswrapper[4754]: I0126 17:26:19.925067 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/18417e65-9c20-4b5e-96df-b7b79c67433a-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "18417e65-9c20-4b5e-96df-b7b79c67433a" (UID: "18417e65-9c20-4b5e-96df-b7b79c67433a"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:26:19 crc kubenswrapper[4754]: I0126 17:26:19.988286 4754 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/18417e65-9c20-4b5e-96df-b7b79c67433a-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Jan 26 17:26:19 crc kubenswrapper[4754]: I0126 17:26:19.988324 4754 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/18417e65-9c20-4b5e-96df-b7b79c67433a-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Jan 26 17:26:19 crc kubenswrapper[4754]: I0126 17:26:19.988334 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7qcjl\" (UniqueName: \"kubernetes.io/projected/18417e65-9c20-4b5e-96df-b7b79c67433a-kube-api-access-7qcjl\") on node \"crc\" DevicePath \"\"" Jan 26 17:26:19 crc kubenswrapper[4754]: I0126 17:26:19.988344 4754 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/18417e65-9c20-4b5e-96df-b7b79c67433a-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Jan 26 17:26:19 crc kubenswrapper[4754]: I0126 17:26:19.988353 4754 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/18417e65-9c20-4b5e-96df-b7b79c67433a-inventory\") on node \"crc\" DevicePath \"\"" Jan 26 17:26:19 crc kubenswrapper[4754]: I0126 17:26:19.988361 4754 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18417e65-9c20-4b5e-96df-b7b79c67433a-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 26 17:26:20 crc kubenswrapper[4754]: I0126 17:26:20.262959 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wljzd" event={"ID":"18417e65-9c20-4b5e-96df-b7b79c67433a","Type":"ContainerDied","Data":"d5244bb9f79b8359dd2a48258ce8b656fcb6dc1492586967285422c49e355547"} Jan 26 17:26:20 crc kubenswrapper[4754]: I0126 17:26:20.263281 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d5244bb9f79b8359dd2a48258ce8b656fcb6dc1492586967285422c49e355547" Jan 26 17:26:20 crc kubenswrapper[4754]: I0126 17:26:20.263003 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wljzd" Jan 26 17:26:20 crc kubenswrapper[4754]: I0126 17:26:20.371851 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lpbsc"] Jan 26 17:26:20 crc kubenswrapper[4754]: E0126 17:26:20.372302 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="18417e65-9c20-4b5e-96df-b7b79c67433a" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Jan 26 17:26:20 crc kubenswrapper[4754]: I0126 17:26:20.372333 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="18417e65-9c20-4b5e-96df-b7b79c67433a" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Jan 26 17:26:20 crc kubenswrapper[4754]: I0126 17:26:20.372591 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="18417e65-9c20-4b5e-96df-b7b79c67433a" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Jan 26 17:26:20 crc kubenswrapper[4754]: I0126 17:26:20.373394 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lpbsc" Jan 26 17:26:20 crc kubenswrapper[4754]: I0126 17:26:20.378022 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Jan 26 17:26:20 crc kubenswrapper[4754]: I0126 17:26:20.378191 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Jan 26 17:26:20 crc kubenswrapper[4754]: I0126 17:26:20.378477 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Jan 26 17:26:20 crc kubenswrapper[4754]: I0126 17:26:20.378538 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Jan 26 17:26:20 crc kubenswrapper[4754]: I0126 17:26:20.378093 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-fz7z8" Jan 26 17:26:20 crc kubenswrapper[4754]: I0126 17:26:20.391459 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lpbsc"] Jan 26 17:26:20 crc kubenswrapper[4754]: I0126 17:26:20.501475 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/a8d7c557-f25d-4aaa-94be-cf9e8e26bc19-ssh-key-openstack-edpm-ipam\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-lpbsc\" (UID: \"a8d7c557-f25d-4aaa-94be-cf9e8e26bc19\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lpbsc" Jan 26 17:26:20 crc kubenswrapper[4754]: I0126 17:26:20.501537 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dkwhs\" (UniqueName: \"kubernetes.io/projected/a8d7c557-f25d-4aaa-94be-cf9e8e26bc19-kube-api-access-dkwhs\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-lpbsc\" (UID: \"a8d7c557-f25d-4aaa-94be-cf9e8e26bc19\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lpbsc" Jan 26 17:26:20 crc kubenswrapper[4754]: I0126 17:26:20.501601 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a8d7c557-f25d-4aaa-94be-cf9e8e26bc19-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-lpbsc\" (UID: \"a8d7c557-f25d-4aaa-94be-cf9e8e26bc19\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lpbsc" Jan 26 17:26:20 crc kubenswrapper[4754]: I0126 17:26:20.501736 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/a8d7c557-f25d-4aaa-94be-cf9e8e26bc19-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-lpbsc\" (UID: \"a8d7c557-f25d-4aaa-94be-cf9e8e26bc19\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lpbsc" Jan 26 17:26:20 crc kubenswrapper[4754]: I0126 17:26:20.501782 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8d7c557-f25d-4aaa-94be-cf9e8e26bc19-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-lpbsc\" (UID: \"a8d7c557-f25d-4aaa-94be-cf9e8e26bc19\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lpbsc" Jan 26 17:26:20 crc kubenswrapper[4754]: I0126 17:26:20.604323 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8d7c557-f25d-4aaa-94be-cf9e8e26bc19-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-lpbsc\" (UID: \"a8d7c557-f25d-4aaa-94be-cf9e8e26bc19\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lpbsc" Jan 26 17:26:20 crc kubenswrapper[4754]: I0126 17:26:20.604418 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/a8d7c557-f25d-4aaa-94be-cf9e8e26bc19-ssh-key-openstack-edpm-ipam\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-lpbsc\" (UID: \"a8d7c557-f25d-4aaa-94be-cf9e8e26bc19\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lpbsc" Jan 26 17:26:20 crc kubenswrapper[4754]: I0126 17:26:20.604465 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dkwhs\" (UniqueName: \"kubernetes.io/projected/a8d7c557-f25d-4aaa-94be-cf9e8e26bc19-kube-api-access-dkwhs\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-lpbsc\" (UID: \"a8d7c557-f25d-4aaa-94be-cf9e8e26bc19\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lpbsc" Jan 26 17:26:20 crc kubenswrapper[4754]: I0126 17:26:20.604561 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a8d7c557-f25d-4aaa-94be-cf9e8e26bc19-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-lpbsc\" (UID: \"a8d7c557-f25d-4aaa-94be-cf9e8e26bc19\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lpbsc" Jan 26 17:26:20 crc kubenswrapper[4754]: I0126 17:26:20.604649 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/a8d7c557-f25d-4aaa-94be-cf9e8e26bc19-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-lpbsc\" (UID: \"a8d7c557-f25d-4aaa-94be-cf9e8e26bc19\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lpbsc" Jan 26 17:26:20 crc kubenswrapper[4754]: I0126 17:26:20.609273 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/a8d7c557-f25d-4aaa-94be-cf9e8e26bc19-ssh-key-openstack-edpm-ipam\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-lpbsc\" (UID: \"a8d7c557-f25d-4aaa-94be-cf9e8e26bc19\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lpbsc" Jan 26 17:26:20 crc kubenswrapper[4754]: I0126 17:26:20.609566 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/a8d7c557-f25d-4aaa-94be-cf9e8e26bc19-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-lpbsc\" (UID: \"a8d7c557-f25d-4aaa-94be-cf9e8e26bc19\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lpbsc" Jan 26 17:26:20 crc kubenswrapper[4754]: I0126 17:26:20.609902 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8d7c557-f25d-4aaa-94be-cf9e8e26bc19-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-lpbsc\" (UID: \"a8d7c557-f25d-4aaa-94be-cf9e8e26bc19\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lpbsc" Jan 26 17:26:20 crc kubenswrapper[4754]: I0126 17:26:20.612541 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a8d7c557-f25d-4aaa-94be-cf9e8e26bc19-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-lpbsc\" (UID: \"a8d7c557-f25d-4aaa-94be-cf9e8e26bc19\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lpbsc" Jan 26 17:26:20 crc kubenswrapper[4754]: I0126 17:26:20.624329 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dkwhs\" (UniqueName: \"kubernetes.io/projected/a8d7c557-f25d-4aaa-94be-cf9e8e26bc19-kube-api-access-dkwhs\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-lpbsc\" (UID: \"a8d7c557-f25d-4aaa-94be-cf9e8e26bc19\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lpbsc" Jan 26 17:26:20 crc kubenswrapper[4754]: I0126 17:26:20.699901 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lpbsc" Jan 26 17:26:21 crc kubenswrapper[4754]: I0126 17:26:21.259460 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lpbsc"] Jan 26 17:26:21 crc kubenswrapper[4754]: I0126 17:26:21.275638 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lpbsc" event={"ID":"a8d7c557-f25d-4aaa-94be-cf9e8e26bc19","Type":"ContainerStarted","Data":"eba98b958806894d5f78bd4e32ec566fa959345e672dd4ff2027908a33fdcc88"} Jan 26 17:26:22 crc kubenswrapper[4754]: I0126 17:26:22.286594 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lpbsc" event={"ID":"a8d7c557-f25d-4aaa-94be-cf9e8e26bc19","Type":"ContainerStarted","Data":"509f7cb363a2df5dd5a50fcb7fe8105677cac729172c881a60f9656e1067d6b5"} Jan 26 17:26:22 crc kubenswrapper[4754]: I0126 17:26:22.316305 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lpbsc" podStartSLOduration=1.8670311000000002 podStartE2EDuration="2.316269214s" podCreationTimestamp="2026-01-26 17:26:20 +0000 UTC" firstStartedPulling="2026-01-26 17:26:21.266282607 +0000 UTC m=+2347.790463041" lastFinishedPulling="2026-01-26 17:26:21.715520721 +0000 UTC m=+2348.239701155" observedRunningTime="2026-01-26 17:26:22.308592271 +0000 UTC m=+2348.832772725" watchObservedRunningTime="2026-01-26 17:26:22.316269214 +0000 UTC m=+2348.840449648" Jan 26 17:26:26 crc kubenswrapper[4754]: I0126 17:26:26.768325 4754 scope.go:117] "RemoveContainer" containerID="00e15c82b01f7eb3c11968e396c0748604b39b71891170fde13a876ab0a851b3" Jan 26 17:26:26 crc kubenswrapper[4754]: E0126 17:26:26.769159 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x65wv_openshift-machine-config-operator(8c3718a4-f354-4284-92e0-fdfb45a692bd)\"" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" Jan 26 17:26:40 crc kubenswrapper[4754]: I0126 17:26:40.768356 4754 scope.go:117] "RemoveContainer" containerID="00e15c82b01f7eb3c11968e396c0748604b39b71891170fde13a876ab0a851b3" Jan 26 17:26:40 crc kubenswrapper[4754]: E0126 17:26:40.769289 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x65wv_openshift-machine-config-operator(8c3718a4-f354-4284-92e0-fdfb45a692bd)\"" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" Jan 26 17:26:53 crc kubenswrapper[4754]: I0126 17:26:53.775503 4754 scope.go:117] "RemoveContainer" containerID="00e15c82b01f7eb3c11968e396c0748604b39b71891170fde13a876ab0a851b3" Jan 26 17:26:53 crc kubenswrapper[4754]: E0126 17:26:53.776929 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x65wv_openshift-machine-config-operator(8c3718a4-f354-4284-92e0-fdfb45a692bd)\"" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" Jan 26 17:27:06 crc kubenswrapper[4754]: I0126 17:27:06.767489 4754 scope.go:117] "RemoveContainer" containerID="00e15c82b01f7eb3c11968e396c0748604b39b71891170fde13a876ab0a851b3" Jan 26 17:27:06 crc kubenswrapper[4754]: E0126 17:27:06.768149 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x65wv_openshift-machine-config-operator(8c3718a4-f354-4284-92e0-fdfb45a692bd)\"" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" Jan 26 17:27:19 crc kubenswrapper[4754]: I0126 17:27:19.767496 4754 scope.go:117] "RemoveContainer" containerID="00e15c82b01f7eb3c11968e396c0748604b39b71891170fde13a876ab0a851b3" Jan 26 17:27:19 crc kubenswrapper[4754]: E0126 17:27:19.768261 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x65wv_openshift-machine-config-operator(8c3718a4-f354-4284-92e0-fdfb45a692bd)\"" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" Jan 26 17:27:31 crc kubenswrapper[4754]: I0126 17:27:31.767695 4754 scope.go:117] "RemoveContainer" containerID="00e15c82b01f7eb3c11968e396c0748604b39b71891170fde13a876ab0a851b3" Jan 26 17:27:31 crc kubenswrapper[4754]: E0126 17:27:31.768888 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x65wv_openshift-machine-config-operator(8c3718a4-f354-4284-92e0-fdfb45a692bd)\"" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" Jan 26 17:27:46 crc kubenswrapper[4754]: I0126 17:27:46.767522 4754 scope.go:117] "RemoveContainer" containerID="00e15c82b01f7eb3c11968e396c0748604b39b71891170fde13a876ab0a851b3" Jan 26 17:27:46 crc kubenswrapper[4754]: E0126 17:27:46.768369 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x65wv_openshift-machine-config-operator(8c3718a4-f354-4284-92e0-fdfb45a692bd)\"" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" Jan 26 17:27:58 crc kubenswrapper[4754]: I0126 17:27:58.768552 4754 scope.go:117] "RemoveContainer" containerID="00e15c82b01f7eb3c11968e396c0748604b39b71891170fde13a876ab0a851b3" Jan 26 17:27:58 crc kubenswrapper[4754]: E0126 17:27:58.769341 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x65wv_openshift-machine-config-operator(8c3718a4-f354-4284-92e0-fdfb45a692bd)\"" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" Jan 26 17:28:10 crc kubenswrapper[4754]: I0126 17:28:10.767402 4754 scope.go:117] "RemoveContainer" containerID="00e15c82b01f7eb3c11968e396c0748604b39b71891170fde13a876ab0a851b3" Jan 26 17:28:10 crc kubenswrapper[4754]: E0126 17:28:10.768154 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x65wv_openshift-machine-config-operator(8c3718a4-f354-4284-92e0-fdfb45a692bd)\"" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" Jan 26 17:28:25 crc kubenswrapper[4754]: I0126 17:28:25.767842 4754 scope.go:117] "RemoveContainer" containerID="00e15c82b01f7eb3c11968e396c0748604b39b71891170fde13a876ab0a851b3" Jan 26 17:28:25 crc kubenswrapper[4754]: E0126 17:28:25.768568 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x65wv_openshift-machine-config-operator(8c3718a4-f354-4284-92e0-fdfb45a692bd)\"" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" Jan 26 17:28:40 crc kubenswrapper[4754]: I0126 17:28:40.767851 4754 scope.go:117] "RemoveContainer" containerID="00e15c82b01f7eb3c11968e396c0748604b39b71891170fde13a876ab0a851b3" Jan 26 17:28:40 crc kubenswrapper[4754]: E0126 17:28:40.768615 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x65wv_openshift-machine-config-operator(8c3718a4-f354-4284-92e0-fdfb45a692bd)\"" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" Jan 26 17:28:54 crc kubenswrapper[4754]: I0126 17:28:54.768549 4754 scope.go:117] "RemoveContainer" containerID="00e15c82b01f7eb3c11968e396c0748604b39b71891170fde13a876ab0a851b3" Jan 26 17:28:54 crc kubenswrapper[4754]: E0126 17:28:54.769305 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x65wv_openshift-machine-config-operator(8c3718a4-f354-4284-92e0-fdfb45a692bd)\"" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" Jan 26 17:29:06 crc kubenswrapper[4754]: I0126 17:29:06.768188 4754 scope.go:117] "RemoveContainer" containerID="00e15c82b01f7eb3c11968e396c0748604b39b71891170fde13a876ab0a851b3" Jan 26 17:29:06 crc kubenswrapper[4754]: E0126 17:29:06.769043 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x65wv_openshift-machine-config-operator(8c3718a4-f354-4284-92e0-fdfb45a692bd)\"" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" Jan 26 17:29:19 crc kubenswrapper[4754]: I0126 17:29:19.769934 4754 scope.go:117] "RemoveContainer" containerID="00e15c82b01f7eb3c11968e396c0748604b39b71891170fde13a876ab0a851b3" Jan 26 17:29:19 crc kubenswrapper[4754]: E0126 17:29:19.771339 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x65wv_openshift-machine-config-operator(8c3718a4-f354-4284-92e0-fdfb45a692bd)\"" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" Jan 26 17:29:34 crc kubenswrapper[4754]: I0126 17:29:34.767880 4754 scope.go:117] "RemoveContainer" containerID="00e15c82b01f7eb3c11968e396c0748604b39b71891170fde13a876ab0a851b3" Jan 26 17:29:34 crc kubenswrapper[4754]: E0126 17:29:34.768635 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x65wv_openshift-machine-config-operator(8c3718a4-f354-4284-92e0-fdfb45a692bd)\"" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" Jan 26 17:29:48 crc kubenswrapper[4754]: I0126 17:29:48.767375 4754 scope.go:117] "RemoveContainer" containerID="00e15c82b01f7eb3c11968e396c0748604b39b71891170fde13a876ab0a851b3" Jan 26 17:29:49 crc kubenswrapper[4754]: I0126 17:29:49.118599 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" event={"ID":"8c3718a4-f354-4284-92e0-fdfb45a692bd","Type":"ContainerStarted","Data":"efeb34d7b9f84cbdd86409079c972f84a47710b40532d5bb54abdaa0f23b0f5d"} Jan 26 17:30:00 crc kubenswrapper[4754]: I0126 17:30:00.146192 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29490810-dwtwj"] Jan 26 17:30:00 crc kubenswrapper[4754]: I0126 17:30:00.149130 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29490810-dwtwj" Jan 26 17:30:00 crc kubenswrapper[4754]: I0126 17:30:00.152537 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Jan 26 17:30:00 crc kubenswrapper[4754]: I0126 17:30:00.153518 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Jan 26 17:30:00 crc kubenswrapper[4754]: I0126 17:30:00.155967 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29490810-dwtwj"] Jan 26 17:30:00 crc kubenswrapper[4754]: I0126 17:30:00.169123 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vwv9w\" (UniqueName: \"kubernetes.io/projected/50db957e-d22d-48b3-8f21-9d6d4868df5b-kube-api-access-vwv9w\") pod \"collect-profiles-29490810-dwtwj\" (UID: \"50db957e-d22d-48b3-8f21-9d6d4868df5b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29490810-dwtwj" Jan 26 17:30:00 crc kubenswrapper[4754]: I0126 17:30:00.169184 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/50db957e-d22d-48b3-8f21-9d6d4868df5b-config-volume\") pod \"collect-profiles-29490810-dwtwj\" (UID: \"50db957e-d22d-48b3-8f21-9d6d4868df5b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29490810-dwtwj" Jan 26 17:30:00 crc kubenswrapper[4754]: I0126 17:30:00.171872 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/50db957e-d22d-48b3-8f21-9d6d4868df5b-secret-volume\") pod \"collect-profiles-29490810-dwtwj\" (UID: \"50db957e-d22d-48b3-8f21-9d6d4868df5b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29490810-dwtwj" Jan 26 17:30:00 crc kubenswrapper[4754]: I0126 17:30:00.274217 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/50db957e-d22d-48b3-8f21-9d6d4868df5b-secret-volume\") pod \"collect-profiles-29490810-dwtwj\" (UID: \"50db957e-d22d-48b3-8f21-9d6d4868df5b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29490810-dwtwj" Jan 26 17:30:00 crc kubenswrapper[4754]: I0126 17:30:00.274282 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vwv9w\" (UniqueName: \"kubernetes.io/projected/50db957e-d22d-48b3-8f21-9d6d4868df5b-kube-api-access-vwv9w\") pod \"collect-profiles-29490810-dwtwj\" (UID: \"50db957e-d22d-48b3-8f21-9d6d4868df5b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29490810-dwtwj" Jan 26 17:30:00 crc kubenswrapper[4754]: I0126 17:30:00.274336 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/50db957e-d22d-48b3-8f21-9d6d4868df5b-config-volume\") pod \"collect-profiles-29490810-dwtwj\" (UID: \"50db957e-d22d-48b3-8f21-9d6d4868df5b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29490810-dwtwj" Jan 26 17:30:00 crc kubenswrapper[4754]: I0126 17:30:00.275551 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/50db957e-d22d-48b3-8f21-9d6d4868df5b-config-volume\") pod \"collect-profiles-29490810-dwtwj\" (UID: \"50db957e-d22d-48b3-8f21-9d6d4868df5b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29490810-dwtwj" Jan 26 17:30:00 crc kubenswrapper[4754]: I0126 17:30:00.282166 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/50db957e-d22d-48b3-8f21-9d6d4868df5b-secret-volume\") pod \"collect-profiles-29490810-dwtwj\" (UID: \"50db957e-d22d-48b3-8f21-9d6d4868df5b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29490810-dwtwj" Jan 26 17:30:00 crc kubenswrapper[4754]: I0126 17:30:00.293147 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vwv9w\" (UniqueName: \"kubernetes.io/projected/50db957e-d22d-48b3-8f21-9d6d4868df5b-kube-api-access-vwv9w\") pod \"collect-profiles-29490810-dwtwj\" (UID: \"50db957e-d22d-48b3-8f21-9d6d4868df5b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29490810-dwtwj" Jan 26 17:30:00 crc kubenswrapper[4754]: I0126 17:30:00.476009 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29490810-dwtwj" Jan 26 17:30:00 crc kubenswrapper[4754]: I0126 17:30:00.918203 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29490810-dwtwj"] Jan 26 17:30:01 crc kubenswrapper[4754]: I0126 17:30:01.219988 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29490810-dwtwj" event={"ID":"50db957e-d22d-48b3-8f21-9d6d4868df5b","Type":"ContainerStarted","Data":"ec67176c3876d5599edb2dbb9e9e7429a2765710c035d779519f8ce1a87fda94"} Jan 26 17:30:01 crc kubenswrapper[4754]: I0126 17:30:01.220378 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29490810-dwtwj" event={"ID":"50db957e-d22d-48b3-8f21-9d6d4868df5b","Type":"ContainerStarted","Data":"384b10d2eac72826605f51b1901d4a3873869fae80646d414734f6c785b5ded9"} Jan 26 17:30:01 crc kubenswrapper[4754]: I0126 17:30:01.237260 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29490810-dwtwj" podStartSLOduration=1.237242987 podStartE2EDuration="1.237242987s" podCreationTimestamp="2026-01-26 17:30:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 17:30:01.233734816 +0000 UTC m=+2567.757915270" watchObservedRunningTime="2026-01-26 17:30:01.237242987 +0000 UTC m=+2567.761423421" Jan 26 17:30:02 crc kubenswrapper[4754]: I0126 17:30:02.230951 4754 generic.go:334] "Generic (PLEG): container finished" podID="50db957e-d22d-48b3-8f21-9d6d4868df5b" containerID="ec67176c3876d5599edb2dbb9e9e7429a2765710c035d779519f8ce1a87fda94" exitCode=0 Jan 26 17:30:02 crc kubenswrapper[4754]: I0126 17:30:02.231026 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29490810-dwtwj" event={"ID":"50db957e-d22d-48b3-8f21-9d6d4868df5b","Type":"ContainerDied","Data":"ec67176c3876d5599edb2dbb9e9e7429a2765710c035d779519f8ce1a87fda94"} Jan 26 17:30:03 crc kubenswrapper[4754]: I0126 17:30:03.547174 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29490810-dwtwj" Jan 26 17:30:03 crc kubenswrapper[4754]: I0126 17:30:03.629065 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/50db957e-d22d-48b3-8f21-9d6d4868df5b-config-volume\") pod \"50db957e-d22d-48b3-8f21-9d6d4868df5b\" (UID: \"50db957e-d22d-48b3-8f21-9d6d4868df5b\") " Jan 26 17:30:03 crc kubenswrapper[4754]: I0126 17:30:03.629876 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/50db957e-d22d-48b3-8f21-9d6d4868df5b-config-volume" (OuterVolumeSpecName: "config-volume") pod "50db957e-d22d-48b3-8f21-9d6d4868df5b" (UID: "50db957e-d22d-48b3-8f21-9d6d4868df5b"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:30:03 crc kubenswrapper[4754]: I0126 17:30:03.629982 4754 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/50db957e-d22d-48b3-8f21-9d6d4868df5b-config-volume\") on node \"crc\" DevicePath \"\"" Jan 26 17:30:03 crc kubenswrapper[4754]: I0126 17:30:03.730761 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vwv9w\" (UniqueName: \"kubernetes.io/projected/50db957e-d22d-48b3-8f21-9d6d4868df5b-kube-api-access-vwv9w\") pod \"50db957e-d22d-48b3-8f21-9d6d4868df5b\" (UID: \"50db957e-d22d-48b3-8f21-9d6d4868df5b\") " Jan 26 17:30:03 crc kubenswrapper[4754]: I0126 17:30:03.730853 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/50db957e-d22d-48b3-8f21-9d6d4868df5b-secret-volume\") pod \"50db957e-d22d-48b3-8f21-9d6d4868df5b\" (UID: \"50db957e-d22d-48b3-8f21-9d6d4868df5b\") " Jan 26 17:30:03 crc kubenswrapper[4754]: I0126 17:30:03.739035 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/50db957e-d22d-48b3-8f21-9d6d4868df5b-kube-api-access-vwv9w" (OuterVolumeSpecName: "kube-api-access-vwv9w") pod "50db957e-d22d-48b3-8f21-9d6d4868df5b" (UID: "50db957e-d22d-48b3-8f21-9d6d4868df5b"). InnerVolumeSpecName "kube-api-access-vwv9w". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:30:03 crc kubenswrapper[4754]: I0126 17:30:03.742830 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/50db957e-d22d-48b3-8f21-9d6d4868df5b-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "50db957e-d22d-48b3-8f21-9d6d4868df5b" (UID: "50db957e-d22d-48b3-8f21-9d6d4868df5b"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:30:03 crc kubenswrapper[4754]: I0126 17:30:03.832387 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vwv9w\" (UniqueName: \"kubernetes.io/projected/50db957e-d22d-48b3-8f21-9d6d4868df5b-kube-api-access-vwv9w\") on node \"crc\" DevicePath \"\"" Jan 26 17:30:03 crc kubenswrapper[4754]: I0126 17:30:03.832431 4754 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/50db957e-d22d-48b3-8f21-9d6d4868df5b-secret-volume\") on node \"crc\" DevicePath \"\"" Jan 26 17:30:04 crc kubenswrapper[4754]: I0126 17:30:04.247179 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29490810-dwtwj" event={"ID":"50db957e-d22d-48b3-8f21-9d6d4868df5b","Type":"ContainerDied","Data":"384b10d2eac72826605f51b1901d4a3873869fae80646d414734f6c785b5ded9"} Jan 26 17:30:04 crc kubenswrapper[4754]: I0126 17:30:04.247222 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="384b10d2eac72826605f51b1901d4a3873869fae80646d414734f6c785b5ded9" Jan 26 17:30:04 crc kubenswrapper[4754]: I0126 17:30:04.247226 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29490810-dwtwj" Jan 26 17:30:04 crc kubenswrapper[4754]: I0126 17:30:04.308803 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29490765-w8zqq"] Jan 26 17:30:04 crc kubenswrapper[4754]: I0126 17:30:04.317768 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29490765-w8zqq"] Jan 26 17:30:05 crc kubenswrapper[4754]: I0126 17:30:05.777472 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f5fabb97-e818-4c00-8023-db5e8477878f" path="/var/lib/kubelet/pods/f5fabb97-e818-4c00-8023-db5e8477878f/volumes" Jan 26 17:30:27 crc kubenswrapper[4754]: I0126 17:30:27.427119 4754 scope.go:117] "RemoveContainer" containerID="e4fb9df6906463bd5ee7e3667dc19c4ad04a1cf49b018f0036e4fbd87047d607" Jan 26 17:30:30 crc kubenswrapper[4754]: I0126 17:30:30.472902 4754 generic.go:334] "Generic (PLEG): container finished" podID="a8d7c557-f25d-4aaa-94be-cf9e8e26bc19" containerID="509f7cb363a2df5dd5a50fcb7fe8105677cac729172c881a60f9656e1067d6b5" exitCode=0 Jan 26 17:30:30 crc kubenswrapper[4754]: I0126 17:30:30.472946 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lpbsc" event={"ID":"a8d7c557-f25d-4aaa-94be-cf9e8e26bc19","Type":"ContainerDied","Data":"509f7cb363a2df5dd5a50fcb7fe8105677cac729172c881a60f9656e1067d6b5"} Jan 26 17:30:31 crc kubenswrapper[4754]: I0126 17:30:31.842125 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lpbsc" Jan 26 17:30:32 crc kubenswrapper[4754]: I0126 17:30:32.035859 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/a8d7c557-f25d-4aaa-94be-cf9e8e26bc19-libvirt-secret-0\") pod \"a8d7c557-f25d-4aaa-94be-cf9e8e26bc19\" (UID: \"a8d7c557-f25d-4aaa-94be-cf9e8e26bc19\") " Jan 26 17:30:32 crc kubenswrapper[4754]: I0126 17:30:32.035933 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8d7c557-f25d-4aaa-94be-cf9e8e26bc19-libvirt-combined-ca-bundle\") pod \"a8d7c557-f25d-4aaa-94be-cf9e8e26bc19\" (UID: \"a8d7c557-f25d-4aaa-94be-cf9e8e26bc19\") " Jan 26 17:30:32 crc kubenswrapper[4754]: I0126 17:30:32.036142 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/a8d7c557-f25d-4aaa-94be-cf9e8e26bc19-ssh-key-openstack-edpm-ipam\") pod \"a8d7c557-f25d-4aaa-94be-cf9e8e26bc19\" (UID: \"a8d7c557-f25d-4aaa-94be-cf9e8e26bc19\") " Jan 26 17:30:32 crc kubenswrapper[4754]: I0126 17:30:32.036243 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a8d7c557-f25d-4aaa-94be-cf9e8e26bc19-inventory\") pod \"a8d7c557-f25d-4aaa-94be-cf9e8e26bc19\" (UID: \"a8d7c557-f25d-4aaa-94be-cf9e8e26bc19\") " Jan 26 17:30:32 crc kubenswrapper[4754]: I0126 17:30:32.036273 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dkwhs\" (UniqueName: \"kubernetes.io/projected/a8d7c557-f25d-4aaa-94be-cf9e8e26bc19-kube-api-access-dkwhs\") pod \"a8d7c557-f25d-4aaa-94be-cf9e8e26bc19\" (UID: \"a8d7c557-f25d-4aaa-94be-cf9e8e26bc19\") " Jan 26 17:30:32 crc kubenswrapper[4754]: I0126 17:30:32.043145 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a8d7c557-f25d-4aaa-94be-cf9e8e26bc19-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "a8d7c557-f25d-4aaa-94be-cf9e8e26bc19" (UID: "a8d7c557-f25d-4aaa-94be-cf9e8e26bc19"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:30:32 crc kubenswrapper[4754]: I0126 17:30:32.043246 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a8d7c557-f25d-4aaa-94be-cf9e8e26bc19-kube-api-access-dkwhs" (OuterVolumeSpecName: "kube-api-access-dkwhs") pod "a8d7c557-f25d-4aaa-94be-cf9e8e26bc19" (UID: "a8d7c557-f25d-4aaa-94be-cf9e8e26bc19"). InnerVolumeSpecName "kube-api-access-dkwhs". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:30:32 crc kubenswrapper[4754]: I0126 17:30:32.064312 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a8d7c557-f25d-4aaa-94be-cf9e8e26bc19-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "a8d7c557-f25d-4aaa-94be-cf9e8e26bc19" (UID: "a8d7c557-f25d-4aaa-94be-cf9e8e26bc19"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:30:32 crc kubenswrapper[4754]: I0126 17:30:32.064436 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a8d7c557-f25d-4aaa-94be-cf9e8e26bc19-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "a8d7c557-f25d-4aaa-94be-cf9e8e26bc19" (UID: "a8d7c557-f25d-4aaa-94be-cf9e8e26bc19"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:30:32 crc kubenswrapper[4754]: I0126 17:30:32.067117 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a8d7c557-f25d-4aaa-94be-cf9e8e26bc19-inventory" (OuterVolumeSpecName: "inventory") pod "a8d7c557-f25d-4aaa-94be-cf9e8e26bc19" (UID: "a8d7c557-f25d-4aaa-94be-cf9e8e26bc19"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:30:32 crc kubenswrapper[4754]: I0126 17:30:32.137911 4754 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/a8d7c557-f25d-4aaa-94be-cf9e8e26bc19-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Jan 26 17:30:32 crc kubenswrapper[4754]: I0126 17:30:32.138132 4754 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8d7c557-f25d-4aaa-94be-cf9e8e26bc19-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 26 17:30:32 crc kubenswrapper[4754]: I0126 17:30:32.138195 4754 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/a8d7c557-f25d-4aaa-94be-cf9e8e26bc19-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Jan 26 17:30:32 crc kubenswrapper[4754]: I0126 17:30:32.138257 4754 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a8d7c557-f25d-4aaa-94be-cf9e8e26bc19-inventory\") on node \"crc\" DevicePath \"\"" Jan 26 17:30:32 crc kubenswrapper[4754]: I0126 17:30:32.138310 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dkwhs\" (UniqueName: \"kubernetes.io/projected/a8d7c557-f25d-4aaa-94be-cf9e8e26bc19-kube-api-access-dkwhs\") on node \"crc\" DevicePath \"\"" Jan 26 17:30:32 crc kubenswrapper[4754]: I0126 17:30:32.492156 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lpbsc" event={"ID":"a8d7c557-f25d-4aaa-94be-cf9e8e26bc19","Type":"ContainerDied","Data":"eba98b958806894d5f78bd4e32ec566fa959345e672dd4ff2027908a33fdcc88"} Jan 26 17:30:32 crc kubenswrapper[4754]: I0126 17:30:32.492238 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="eba98b958806894d5f78bd4e32ec566fa959345e672dd4ff2027908a33fdcc88" Jan 26 17:30:32 crc kubenswrapper[4754]: I0126 17:30:32.492274 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lpbsc" Jan 26 17:30:32 crc kubenswrapper[4754]: I0126 17:30:32.585723 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-jswg2"] Jan 26 17:30:32 crc kubenswrapper[4754]: E0126 17:30:32.586479 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a8d7c557-f25d-4aaa-94be-cf9e8e26bc19" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Jan 26 17:30:32 crc kubenswrapper[4754]: I0126 17:30:32.586497 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="a8d7c557-f25d-4aaa-94be-cf9e8e26bc19" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Jan 26 17:30:32 crc kubenswrapper[4754]: E0126 17:30:32.586515 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="50db957e-d22d-48b3-8f21-9d6d4868df5b" containerName="collect-profiles" Jan 26 17:30:32 crc kubenswrapper[4754]: I0126 17:30:32.586523 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="50db957e-d22d-48b3-8f21-9d6d4868df5b" containerName="collect-profiles" Jan 26 17:30:32 crc kubenswrapper[4754]: I0126 17:30:32.586778 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="50db957e-d22d-48b3-8f21-9d6d4868df5b" containerName="collect-profiles" Jan 26 17:30:32 crc kubenswrapper[4754]: I0126 17:30:32.586799 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="a8d7c557-f25d-4aaa-94be-cf9e8e26bc19" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Jan 26 17:30:32 crc kubenswrapper[4754]: I0126 17:30:32.587535 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-jswg2" Jan 26 17:30:32 crc kubenswrapper[4754]: I0126 17:30:32.589375 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Jan 26 17:30:32 crc kubenswrapper[4754]: I0126 17:30:32.590567 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-extra-config" Jan 26 17:30:32 crc kubenswrapper[4754]: I0126 17:30:32.590823 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Jan 26 17:30:32 crc kubenswrapper[4754]: I0126 17:30:32.592873 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Jan 26 17:30:32 crc kubenswrapper[4754]: I0126 17:30:32.592972 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-fz7z8" Jan 26 17:30:32 crc kubenswrapper[4754]: I0126 17:30:32.593017 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Jan 26 17:30:32 crc kubenswrapper[4754]: I0126 17:30:32.593024 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Jan 26 17:30:32 crc kubenswrapper[4754]: I0126 17:30:32.597507 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-jswg2"] Jan 26 17:30:32 crc kubenswrapper[4754]: I0126 17:30:32.747515 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/b5408b33-3b71-46ac-87a4-413ecb9614b0-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-jswg2\" (UID: \"b5408b33-3b71-46ac-87a4-413ecb9614b0\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-jswg2" Jan 26 17:30:32 crc kubenswrapper[4754]: I0126 17:30:32.747553 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/b5408b33-3b71-46ac-87a4-413ecb9614b0-ssh-key-openstack-edpm-ipam\") pod \"nova-edpm-deployment-openstack-edpm-ipam-jswg2\" (UID: \"b5408b33-3b71-46ac-87a4-413ecb9614b0\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-jswg2" Jan 26 17:30:32 crc kubenswrapper[4754]: I0126 17:30:32.747577 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/b5408b33-3b71-46ac-87a4-413ecb9614b0-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-jswg2\" (UID: \"b5408b33-3b71-46ac-87a4-413ecb9614b0\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-jswg2" Jan 26 17:30:32 crc kubenswrapper[4754]: I0126 17:30:32.748424 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5408b33-3b71-46ac-87a4-413ecb9614b0-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-jswg2\" (UID: \"b5408b33-3b71-46ac-87a4-413ecb9614b0\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-jswg2" Jan 26 17:30:32 crc kubenswrapper[4754]: I0126 17:30:32.748489 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/b5408b33-3b71-46ac-87a4-413ecb9614b0-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-jswg2\" (UID: \"b5408b33-3b71-46ac-87a4-413ecb9614b0\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-jswg2" Jan 26 17:30:32 crc kubenswrapper[4754]: I0126 17:30:32.748546 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b5408b33-3b71-46ac-87a4-413ecb9614b0-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-jswg2\" (UID: \"b5408b33-3b71-46ac-87a4-413ecb9614b0\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-jswg2" Jan 26 17:30:32 crc kubenswrapper[4754]: I0126 17:30:32.748648 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/b5408b33-3b71-46ac-87a4-413ecb9614b0-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-jswg2\" (UID: \"b5408b33-3b71-46ac-87a4-413ecb9614b0\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-jswg2" Jan 26 17:30:32 crc kubenswrapper[4754]: I0126 17:30:32.748999 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/b5408b33-3b71-46ac-87a4-413ecb9614b0-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-jswg2\" (UID: \"b5408b33-3b71-46ac-87a4-413ecb9614b0\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-jswg2" Jan 26 17:30:32 crc kubenswrapper[4754]: I0126 17:30:32.749089 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p5phm\" (UniqueName: \"kubernetes.io/projected/b5408b33-3b71-46ac-87a4-413ecb9614b0-kube-api-access-p5phm\") pod \"nova-edpm-deployment-openstack-edpm-ipam-jswg2\" (UID: \"b5408b33-3b71-46ac-87a4-413ecb9614b0\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-jswg2" Jan 26 17:30:32 crc kubenswrapper[4754]: I0126 17:30:32.850604 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p5phm\" (UniqueName: \"kubernetes.io/projected/b5408b33-3b71-46ac-87a4-413ecb9614b0-kube-api-access-p5phm\") pod \"nova-edpm-deployment-openstack-edpm-ipam-jswg2\" (UID: \"b5408b33-3b71-46ac-87a4-413ecb9614b0\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-jswg2" Jan 26 17:30:32 crc kubenswrapper[4754]: I0126 17:30:32.850734 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/b5408b33-3b71-46ac-87a4-413ecb9614b0-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-jswg2\" (UID: \"b5408b33-3b71-46ac-87a4-413ecb9614b0\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-jswg2" Jan 26 17:30:32 crc kubenswrapper[4754]: I0126 17:30:32.850757 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/b5408b33-3b71-46ac-87a4-413ecb9614b0-ssh-key-openstack-edpm-ipam\") pod \"nova-edpm-deployment-openstack-edpm-ipam-jswg2\" (UID: \"b5408b33-3b71-46ac-87a4-413ecb9614b0\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-jswg2" Jan 26 17:30:32 crc kubenswrapper[4754]: I0126 17:30:32.850778 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/b5408b33-3b71-46ac-87a4-413ecb9614b0-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-jswg2\" (UID: \"b5408b33-3b71-46ac-87a4-413ecb9614b0\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-jswg2" Jan 26 17:30:32 crc kubenswrapper[4754]: I0126 17:30:32.850843 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5408b33-3b71-46ac-87a4-413ecb9614b0-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-jswg2\" (UID: \"b5408b33-3b71-46ac-87a4-413ecb9614b0\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-jswg2" Jan 26 17:30:32 crc kubenswrapper[4754]: I0126 17:30:32.850873 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/b5408b33-3b71-46ac-87a4-413ecb9614b0-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-jswg2\" (UID: \"b5408b33-3b71-46ac-87a4-413ecb9614b0\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-jswg2" Jan 26 17:30:32 crc kubenswrapper[4754]: I0126 17:30:32.850905 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b5408b33-3b71-46ac-87a4-413ecb9614b0-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-jswg2\" (UID: \"b5408b33-3b71-46ac-87a4-413ecb9614b0\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-jswg2" Jan 26 17:30:32 crc kubenswrapper[4754]: I0126 17:30:32.850946 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/b5408b33-3b71-46ac-87a4-413ecb9614b0-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-jswg2\" (UID: \"b5408b33-3b71-46ac-87a4-413ecb9614b0\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-jswg2" Jan 26 17:30:32 crc kubenswrapper[4754]: I0126 17:30:32.850971 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/b5408b33-3b71-46ac-87a4-413ecb9614b0-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-jswg2\" (UID: \"b5408b33-3b71-46ac-87a4-413ecb9614b0\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-jswg2" Jan 26 17:30:32 crc kubenswrapper[4754]: I0126 17:30:32.852414 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/b5408b33-3b71-46ac-87a4-413ecb9614b0-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-jswg2\" (UID: \"b5408b33-3b71-46ac-87a4-413ecb9614b0\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-jswg2" Jan 26 17:30:32 crc kubenswrapper[4754]: I0126 17:30:32.856587 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/b5408b33-3b71-46ac-87a4-413ecb9614b0-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-jswg2\" (UID: \"b5408b33-3b71-46ac-87a4-413ecb9614b0\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-jswg2" Jan 26 17:30:32 crc kubenswrapper[4754]: I0126 17:30:32.856587 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/b5408b33-3b71-46ac-87a4-413ecb9614b0-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-jswg2\" (UID: \"b5408b33-3b71-46ac-87a4-413ecb9614b0\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-jswg2" Jan 26 17:30:32 crc kubenswrapper[4754]: I0126 17:30:32.857017 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/b5408b33-3b71-46ac-87a4-413ecb9614b0-ssh-key-openstack-edpm-ipam\") pod \"nova-edpm-deployment-openstack-edpm-ipam-jswg2\" (UID: \"b5408b33-3b71-46ac-87a4-413ecb9614b0\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-jswg2" Jan 26 17:30:32 crc kubenswrapper[4754]: I0126 17:30:32.857034 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b5408b33-3b71-46ac-87a4-413ecb9614b0-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-jswg2\" (UID: \"b5408b33-3b71-46ac-87a4-413ecb9614b0\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-jswg2" Jan 26 17:30:32 crc kubenswrapper[4754]: I0126 17:30:32.857266 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5408b33-3b71-46ac-87a4-413ecb9614b0-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-jswg2\" (UID: \"b5408b33-3b71-46ac-87a4-413ecb9614b0\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-jswg2" Jan 26 17:30:32 crc kubenswrapper[4754]: I0126 17:30:32.858190 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/b5408b33-3b71-46ac-87a4-413ecb9614b0-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-jswg2\" (UID: \"b5408b33-3b71-46ac-87a4-413ecb9614b0\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-jswg2" Jan 26 17:30:32 crc kubenswrapper[4754]: I0126 17:30:32.858903 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/b5408b33-3b71-46ac-87a4-413ecb9614b0-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-jswg2\" (UID: \"b5408b33-3b71-46ac-87a4-413ecb9614b0\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-jswg2" Jan 26 17:30:32 crc kubenswrapper[4754]: I0126 17:30:32.868723 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p5phm\" (UniqueName: \"kubernetes.io/projected/b5408b33-3b71-46ac-87a4-413ecb9614b0-kube-api-access-p5phm\") pod \"nova-edpm-deployment-openstack-edpm-ipam-jswg2\" (UID: \"b5408b33-3b71-46ac-87a4-413ecb9614b0\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-jswg2" Jan 26 17:30:32 crc kubenswrapper[4754]: I0126 17:30:32.909822 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-jswg2" Jan 26 17:30:33 crc kubenswrapper[4754]: I0126 17:30:33.389700 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-jswg2"] Jan 26 17:30:33 crc kubenswrapper[4754]: I0126 17:30:33.392407 4754 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 26 17:30:33 crc kubenswrapper[4754]: I0126 17:30:33.501611 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-jswg2" event={"ID":"b5408b33-3b71-46ac-87a4-413ecb9614b0","Type":"ContainerStarted","Data":"2216d89e7a7b851415c7080c6d481507a0bdd04990cdf0408576e6e7469f6652"} Jan 26 17:30:34 crc kubenswrapper[4754]: I0126 17:30:34.510784 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-jswg2" event={"ID":"b5408b33-3b71-46ac-87a4-413ecb9614b0","Type":"ContainerStarted","Data":"ba97fdca943084140236a71b63afdaa65a1ed00ca09bc5f0dd10aeb80ef90c6d"} Jan 26 17:30:34 crc kubenswrapper[4754]: I0126 17:30:34.532911 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-jswg2" podStartSLOduration=1.72894769 podStartE2EDuration="2.532893082s" podCreationTimestamp="2026-01-26 17:30:32 +0000 UTC" firstStartedPulling="2026-01-26 17:30:33.392135559 +0000 UTC m=+2599.916315993" lastFinishedPulling="2026-01-26 17:30:34.196080951 +0000 UTC m=+2600.720261385" observedRunningTime="2026-01-26 17:30:34.525755607 +0000 UTC m=+2601.049936051" watchObservedRunningTime="2026-01-26 17:30:34.532893082 +0000 UTC m=+2601.057073516" Jan 26 17:31:17 crc kubenswrapper[4754]: I0126 17:31:17.127110 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-jvfdk"] Jan 26 17:31:17 crc kubenswrapper[4754]: I0126 17:31:17.134256 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jvfdk" Jan 26 17:31:17 crc kubenswrapper[4754]: I0126 17:31:17.152143 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-jvfdk"] Jan 26 17:31:17 crc kubenswrapper[4754]: I0126 17:31:17.292505 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ceb4005f-7a24-4c06-8afd-5cad6c195c57-catalog-content\") pod \"certified-operators-jvfdk\" (UID: \"ceb4005f-7a24-4c06-8afd-5cad6c195c57\") " pod="openshift-marketplace/certified-operators-jvfdk" Jan 26 17:31:17 crc kubenswrapper[4754]: I0126 17:31:17.292635 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4w2g9\" (UniqueName: \"kubernetes.io/projected/ceb4005f-7a24-4c06-8afd-5cad6c195c57-kube-api-access-4w2g9\") pod \"certified-operators-jvfdk\" (UID: \"ceb4005f-7a24-4c06-8afd-5cad6c195c57\") " pod="openshift-marketplace/certified-operators-jvfdk" Jan 26 17:31:17 crc kubenswrapper[4754]: I0126 17:31:17.292819 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ceb4005f-7a24-4c06-8afd-5cad6c195c57-utilities\") pod \"certified-operators-jvfdk\" (UID: \"ceb4005f-7a24-4c06-8afd-5cad6c195c57\") " pod="openshift-marketplace/certified-operators-jvfdk" Jan 26 17:31:17 crc kubenswrapper[4754]: I0126 17:31:17.394409 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ceb4005f-7a24-4c06-8afd-5cad6c195c57-utilities\") pod \"certified-operators-jvfdk\" (UID: \"ceb4005f-7a24-4c06-8afd-5cad6c195c57\") " pod="openshift-marketplace/certified-operators-jvfdk" Jan 26 17:31:17 crc kubenswrapper[4754]: I0126 17:31:17.394516 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ceb4005f-7a24-4c06-8afd-5cad6c195c57-catalog-content\") pod \"certified-operators-jvfdk\" (UID: \"ceb4005f-7a24-4c06-8afd-5cad6c195c57\") " pod="openshift-marketplace/certified-operators-jvfdk" Jan 26 17:31:17 crc kubenswrapper[4754]: I0126 17:31:17.394603 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4w2g9\" (UniqueName: \"kubernetes.io/projected/ceb4005f-7a24-4c06-8afd-5cad6c195c57-kube-api-access-4w2g9\") pod \"certified-operators-jvfdk\" (UID: \"ceb4005f-7a24-4c06-8afd-5cad6c195c57\") " pod="openshift-marketplace/certified-operators-jvfdk" Jan 26 17:31:17 crc kubenswrapper[4754]: I0126 17:31:17.395093 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ceb4005f-7a24-4c06-8afd-5cad6c195c57-catalog-content\") pod \"certified-operators-jvfdk\" (UID: \"ceb4005f-7a24-4c06-8afd-5cad6c195c57\") " pod="openshift-marketplace/certified-operators-jvfdk" Jan 26 17:31:17 crc kubenswrapper[4754]: I0126 17:31:17.395328 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ceb4005f-7a24-4c06-8afd-5cad6c195c57-utilities\") pod \"certified-operators-jvfdk\" (UID: \"ceb4005f-7a24-4c06-8afd-5cad6c195c57\") " pod="openshift-marketplace/certified-operators-jvfdk" Jan 26 17:31:17 crc kubenswrapper[4754]: I0126 17:31:17.417092 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4w2g9\" (UniqueName: \"kubernetes.io/projected/ceb4005f-7a24-4c06-8afd-5cad6c195c57-kube-api-access-4w2g9\") pod \"certified-operators-jvfdk\" (UID: \"ceb4005f-7a24-4c06-8afd-5cad6c195c57\") " pod="openshift-marketplace/certified-operators-jvfdk" Jan 26 17:31:17 crc kubenswrapper[4754]: I0126 17:31:17.457153 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jvfdk" Jan 26 17:31:17 crc kubenswrapper[4754]: I0126 17:31:17.973613 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-jvfdk"] Jan 26 17:31:18 crc kubenswrapper[4754]: I0126 17:31:18.865937 4754 generic.go:334] "Generic (PLEG): container finished" podID="ceb4005f-7a24-4c06-8afd-5cad6c195c57" containerID="d24ae875af77bd9ee7990ac43b0af766e8b9083229869f673f03d7640f059d0a" exitCode=0 Jan 26 17:31:18 crc kubenswrapper[4754]: I0126 17:31:18.866048 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jvfdk" event={"ID":"ceb4005f-7a24-4c06-8afd-5cad6c195c57","Type":"ContainerDied","Data":"d24ae875af77bd9ee7990ac43b0af766e8b9083229869f673f03d7640f059d0a"} Jan 26 17:31:18 crc kubenswrapper[4754]: I0126 17:31:18.866137 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jvfdk" event={"ID":"ceb4005f-7a24-4c06-8afd-5cad6c195c57","Type":"ContainerStarted","Data":"05a9c13184a75a678522cf5830b9d4702ca86a56dc92e61c4077897f8f2c0bda"} Jan 26 17:31:22 crc kubenswrapper[4754]: I0126 17:31:22.908550 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jvfdk" event={"ID":"ceb4005f-7a24-4c06-8afd-5cad6c195c57","Type":"ContainerStarted","Data":"6bec897356445c42c63125de67159d6e6befa55e179a895705b42fbae6cf06f4"} Jan 26 17:31:23 crc kubenswrapper[4754]: I0126 17:31:23.918268 4754 generic.go:334] "Generic (PLEG): container finished" podID="ceb4005f-7a24-4c06-8afd-5cad6c195c57" containerID="6bec897356445c42c63125de67159d6e6befa55e179a895705b42fbae6cf06f4" exitCode=0 Jan 26 17:31:23 crc kubenswrapper[4754]: I0126 17:31:23.918319 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jvfdk" event={"ID":"ceb4005f-7a24-4c06-8afd-5cad6c195c57","Type":"ContainerDied","Data":"6bec897356445c42c63125de67159d6e6befa55e179a895705b42fbae6cf06f4"} Jan 26 17:31:27 crc kubenswrapper[4754]: I0126 17:31:27.955347 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jvfdk" event={"ID":"ceb4005f-7a24-4c06-8afd-5cad6c195c57","Type":"ContainerStarted","Data":"4262dd831cb45a47435f460d1353b0a16361f5f677395f3f474207738be2d687"} Jan 26 17:31:27 crc kubenswrapper[4754]: I0126 17:31:27.982352 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-jvfdk" podStartSLOduration=2.392424438 podStartE2EDuration="10.982328331s" podCreationTimestamp="2026-01-26 17:31:17 +0000 UTC" firstStartedPulling="2026-01-26 17:31:18.869408941 +0000 UTC m=+2645.393589375" lastFinishedPulling="2026-01-26 17:31:27.459312834 +0000 UTC m=+2653.983493268" observedRunningTime="2026-01-26 17:31:27.975556946 +0000 UTC m=+2654.499737400" watchObservedRunningTime="2026-01-26 17:31:27.982328331 +0000 UTC m=+2654.506508765" Jan 26 17:31:37 crc kubenswrapper[4754]: I0126 17:31:37.457754 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-jvfdk" Jan 26 17:31:37 crc kubenswrapper[4754]: I0126 17:31:37.458326 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-jvfdk" Jan 26 17:31:37 crc kubenswrapper[4754]: I0126 17:31:37.499395 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-jvfdk" Jan 26 17:31:38 crc kubenswrapper[4754]: I0126 17:31:38.127707 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-jvfdk" Jan 26 17:31:38 crc kubenswrapper[4754]: I0126 17:31:38.189972 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-jvfdk"] Jan 26 17:31:38 crc kubenswrapper[4754]: I0126 17:31:38.227866 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-l482d"] Jan 26 17:31:38 crc kubenswrapper[4754]: I0126 17:31:38.228100 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-l482d" podUID="833110b5-684c-4097-9e52-d3d8d1838e73" containerName="registry-server" containerID="cri-o://9082e167aba36fe6b7a850694a1fb9c6d2a872460309547b98fd483c3424871f" gracePeriod=2 Jan 26 17:31:38 crc kubenswrapper[4754]: I0126 17:31:38.786321 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-l482d" Jan 26 17:31:38 crc kubenswrapper[4754]: I0126 17:31:38.911016 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6s9wg\" (UniqueName: \"kubernetes.io/projected/833110b5-684c-4097-9e52-d3d8d1838e73-kube-api-access-6s9wg\") pod \"833110b5-684c-4097-9e52-d3d8d1838e73\" (UID: \"833110b5-684c-4097-9e52-d3d8d1838e73\") " Jan 26 17:31:38 crc kubenswrapper[4754]: I0126 17:31:38.911141 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/833110b5-684c-4097-9e52-d3d8d1838e73-utilities\") pod \"833110b5-684c-4097-9e52-d3d8d1838e73\" (UID: \"833110b5-684c-4097-9e52-d3d8d1838e73\") " Jan 26 17:31:38 crc kubenswrapper[4754]: I0126 17:31:38.911237 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/833110b5-684c-4097-9e52-d3d8d1838e73-catalog-content\") pod \"833110b5-684c-4097-9e52-d3d8d1838e73\" (UID: \"833110b5-684c-4097-9e52-d3d8d1838e73\") " Jan 26 17:31:38 crc kubenswrapper[4754]: I0126 17:31:38.914374 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/833110b5-684c-4097-9e52-d3d8d1838e73-utilities" (OuterVolumeSpecName: "utilities") pod "833110b5-684c-4097-9e52-d3d8d1838e73" (UID: "833110b5-684c-4097-9e52-d3d8d1838e73"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:31:38 crc kubenswrapper[4754]: I0126 17:31:38.918347 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/833110b5-684c-4097-9e52-d3d8d1838e73-kube-api-access-6s9wg" (OuterVolumeSpecName: "kube-api-access-6s9wg") pod "833110b5-684c-4097-9e52-d3d8d1838e73" (UID: "833110b5-684c-4097-9e52-d3d8d1838e73"). InnerVolumeSpecName "kube-api-access-6s9wg". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:31:38 crc kubenswrapper[4754]: I0126 17:31:38.964149 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/833110b5-684c-4097-9e52-d3d8d1838e73-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "833110b5-684c-4097-9e52-d3d8d1838e73" (UID: "833110b5-684c-4097-9e52-d3d8d1838e73"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:31:39 crc kubenswrapper[4754]: I0126 17:31:39.013929 4754 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/833110b5-684c-4097-9e52-d3d8d1838e73-utilities\") on node \"crc\" DevicePath \"\"" Jan 26 17:31:39 crc kubenswrapper[4754]: I0126 17:31:39.013968 4754 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/833110b5-684c-4097-9e52-d3d8d1838e73-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 26 17:31:39 crc kubenswrapper[4754]: I0126 17:31:39.013980 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6s9wg\" (UniqueName: \"kubernetes.io/projected/833110b5-684c-4097-9e52-d3d8d1838e73-kube-api-access-6s9wg\") on node \"crc\" DevicePath \"\"" Jan 26 17:31:39 crc kubenswrapper[4754]: I0126 17:31:39.069325 4754 generic.go:334] "Generic (PLEG): container finished" podID="833110b5-684c-4097-9e52-d3d8d1838e73" containerID="9082e167aba36fe6b7a850694a1fb9c6d2a872460309547b98fd483c3424871f" exitCode=0 Jan 26 17:31:39 crc kubenswrapper[4754]: I0126 17:31:39.069388 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-l482d" Jan 26 17:31:39 crc kubenswrapper[4754]: I0126 17:31:39.069386 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-l482d" event={"ID":"833110b5-684c-4097-9e52-d3d8d1838e73","Type":"ContainerDied","Data":"9082e167aba36fe6b7a850694a1fb9c6d2a872460309547b98fd483c3424871f"} Jan 26 17:31:39 crc kubenswrapper[4754]: I0126 17:31:39.069552 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-l482d" event={"ID":"833110b5-684c-4097-9e52-d3d8d1838e73","Type":"ContainerDied","Data":"a9596fc361a04720af3a7646e6e2f9902add150a9947aeda0c6cbdf57d75b591"} Jan 26 17:31:39 crc kubenswrapper[4754]: I0126 17:31:39.069576 4754 scope.go:117] "RemoveContainer" containerID="9082e167aba36fe6b7a850694a1fb9c6d2a872460309547b98fd483c3424871f" Jan 26 17:31:39 crc kubenswrapper[4754]: I0126 17:31:39.096616 4754 scope.go:117] "RemoveContainer" containerID="567bc05d44e76fe18b6a1e23914ff4f2bfb8eb85a95eecdfe15f9a5d80eb1ef2" Jan 26 17:31:39 crc kubenswrapper[4754]: I0126 17:31:39.110972 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-l482d"] Jan 26 17:31:39 crc kubenswrapper[4754]: I0126 17:31:39.120340 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-l482d"] Jan 26 17:31:39 crc kubenswrapper[4754]: I0126 17:31:39.129838 4754 scope.go:117] "RemoveContainer" containerID="cdafb1c82434540940dc94751916e14ab3f96474a9b0b50c0dfcee61e2b193fd" Jan 26 17:31:39 crc kubenswrapper[4754]: I0126 17:31:39.183376 4754 scope.go:117] "RemoveContainer" containerID="9082e167aba36fe6b7a850694a1fb9c6d2a872460309547b98fd483c3424871f" Jan 26 17:31:39 crc kubenswrapper[4754]: E0126 17:31:39.183914 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9082e167aba36fe6b7a850694a1fb9c6d2a872460309547b98fd483c3424871f\": container with ID starting with 9082e167aba36fe6b7a850694a1fb9c6d2a872460309547b98fd483c3424871f not found: ID does not exist" containerID="9082e167aba36fe6b7a850694a1fb9c6d2a872460309547b98fd483c3424871f" Jan 26 17:31:39 crc kubenswrapper[4754]: I0126 17:31:39.183947 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9082e167aba36fe6b7a850694a1fb9c6d2a872460309547b98fd483c3424871f"} err="failed to get container status \"9082e167aba36fe6b7a850694a1fb9c6d2a872460309547b98fd483c3424871f\": rpc error: code = NotFound desc = could not find container \"9082e167aba36fe6b7a850694a1fb9c6d2a872460309547b98fd483c3424871f\": container with ID starting with 9082e167aba36fe6b7a850694a1fb9c6d2a872460309547b98fd483c3424871f not found: ID does not exist" Jan 26 17:31:39 crc kubenswrapper[4754]: I0126 17:31:39.183970 4754 scope.go:117] "RemoveContainer" containerID="567bc05d44e76fe18b6a1e23914ff4f2bfb8eb85a95eecdfe15f9a5d80eb1ef2" Jan 26 17:31:39 crc kubenswrapper[4754]: E0126 17:31:39.184611 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"567bc05d44e76fe18b6a1e23914ff4f2bfb8eb85a95eecdfe15f9a5d80eb1ef2\": container with ID starting with 567bc05d44e76fe18b6a1e23914ff4f2bfb8eb85a95eecdfe15f9a5d80eb1ef2 not found: ID does not exist" containerID="567bc05d44e76fe18b6a1e23914ff4f2bfb8eb85a95eecdfe15f9a5d80eb1ef2" Jan 26 17:31:39 crc kubenswrapper[4754]: I0126 17:31:39.184711 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"567bc05d44e76fe18b6a1e23914ff4f2bfb8eb85a95eecdfe15f9a5d80eb1ef2"} err="failed to get container status \"567bc05d44e76fe18b6a1e23914ff4f2bfb8eb85a95eecdfe15f9a5d80eb1ef2\": rpc error: code = NotFound desc = could not find container \"567bc05d44e76fe18b6a1e23914ff4f2bfb8eb85a95eecdfe15f9a5d80eb1ef2\": container with ID starting with 567bc05d44e76fe18b6a1e23914ff4f2bfb8eb85a95eecdfe15f9a5d80eb1ef2 not found: ID does not exist" Jan 26 17:31:39 crc kubenswrapper[4754]: I0126 17:31:39.184744 4754 scope.go:117] "RemoveContainer" containerID="cdafb1c82434540940dc94751916e14ab3f96474a9b0b50c0dfcee61e2b193fd" Jan 26 17:31:39 crc kubenswrapper[4754]: E0126 17:31:39.185109 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cdafb1c82434540940dc94751916e14ab3f96474a9b0b50c0dfcee61e2b193fd\": container with ID starting with cdafb1c82434540940dc94751916e14ab3f96474a9b0b50c0dfcee61e2b193fd not found: ID does not exist" containerID="cdafb1c82434540940dc94751916e14ab3f96474a9b0b50c0dfcee61e2b193fd" Jan 26 17:31:39 crc kubenswrapper[4754]: I0126 17:31:39.185164 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cdafb1c82434540940dc94751916e14ab3f96474a9b0b50c0dfcee61e2b193fd"} err="failed to get container status \"cdafb1c82434540940dc94751916e14ab3f96474a9b0b50c0dfcee61e2b193fd\": rpc error: code = NotFound desc = could not find container \"cdafb1c82434540940dc94751916e14ab3f96474a9b0b50c0dfcee61e2b193fd\": container with ID starting with cdafb1c82434540940dc94751916e14ab3f96474a9b0b50c0dfcee61e2b193fd not found: ID does not exist" Jan 26 17:31:39 crc kubenswrapper[4754]: I0126 17:31:39.778781 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="833110b5-684c-4097-9e52-d3d8d1838e73" path="/var/lib/kubelet/pods/833110b5-684c-4097-9e52-d3d8d1838e73/volumes" Jan 26 17:31:40 crc kubenswrapper[4754]: I0126 17:31:40.746412 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-n69nh"] Jan 26 17:31:40 crc kubenswrapper[4754]: E0126 17:31:40.747141 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="833110b5-684c-4097-9e52-d3d8d1838e73" containerName="registry-server" Jan 26 17:31:40 crc kubenswrapper[4754]: I0126 17:31:40.747159 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="833110b5-684c-4097-9e52-d3d8d1838e73" containerName="registry-server" Jan 26 17:31:40 crc kubenswrapper[4754]: E0126 17:31:40.747184 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="833110b5-684c-4097-9e52-d3d8d1838e73" containerName="extract-content" Jan 26 17:31:40 crc kubenswrapper[4754]: I0126 17:31:40.747191 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="833110b5-684c-4097-9e52-d3d8d1838e73" containerName="extract-content" Jan 26 17:31:40 crc kubenswrapper[4754]: E0126 17:31:40.747220 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="833110b5-684c-4097-9e52-d3d8d1838e73" containerName="extract-utilities" Jan 26 17:31:40 crc kubenswrapper[4754]: I0126 17:31:40.747229 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="833110b5-684c-4097-9e52-d3d8d1838e73" containerName="extract-utilities" Jan 26 17:31:40 crc kubenswrapper[4754]: I0126 17:31:40.747461 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="833110b5-684c-4097-9e52-d3d8d1838e73" containerName="registry-server" Jan 26 17:31:40 crc kubenswrapper[4754]: I0126 17:31:40.748920 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-n69nh" Jan 26 17:31:40 crc kubenswrapper[4754]: I0126 17:31:40.760227 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-n69nh"] Jan 26 17:31:40 crc kubenswrapper[4754]: I0126 17:31:40.848795 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e1e1571f-18fd-402c-b212-70635bbb1fb2-utilities\") pod \"redhat-operators-n69nh\" (UID: \"e1e1571f-18fd-402c-b212-70635bbb1fb2\") " pod="openshift-marketplace/redhat-operators-n69nh" Jan 26 17:31:40 crc kubenswrapper[4754]: I0126 17:31:40.848878 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-km4pj\" (UniqueName: \"kubernetes.io/projected/e1e1571f-18fd-402c-b212-70635bbb1fb2-kube-api-access-km4pj\") pod \"redhat-operators-n69nh\" (UID: \"e1e1571f-18fd-402c-b212-70635bbb1fb2\") " pod="openshift-marketplace/redhat-operators-n69nh" Jan 26 17:31:40 crc kubenswrapper[4754]: I0126 17:31:40.848912 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e1e1571f-18fd-402c-b212-70635bbb1fb2-catalog-content\") pod \"redhat-operators-n69nh\" (UID: \"e1e1571f-18fd-402c-b212-70635bbb1fb2\") " pod="openshift-marketplace/redhat-operators-n69nh" Jan 26 17:31:40 crc kubenswrapper[4754]: I0126 17:31:40.951115 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e1e1571f-18fd-402c-b212-70635bbb1fb2-utilities\") pod \"redhat-operators-n69nh\" (UID: \"e1e1571f-18fd-402c-b212-70635bbb1fb2\") " pod="openshift-marketplace/redhat-operators-n69nh" Jan 26 17:31:40 crc kubenswrapper[4754]: I0126 17:31:40.951435 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-km4pj\" (UniqueName: \"kubernetes.io/projected/e1e1571f-18fd-402c-b212-70635bbb1fb2-kube-api-access-km4pj\") pod \"redhat-operators-n69nh\" (UID: \"e1e1571f-18fd-402c-b212-70635bbb1fb2\") " pod="openshift-marketplace/redhat-operators-n69nh" Jan 26 17:31:40 crc kubenswrapper[4754]: I0126 17:31:40.951592 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e1e1571f-18fd-402c-b212-70635bbb1fb2-catalog-content\") pod \"redhat-operators-n69nh\" (UID: \"e1e1571f-18fd-402c-b212-70635bbb1fb2\") " pod="openshift-marketplace/redhat-operators-n69nh" Jan 26 17:31:40 crc kubenswrapper[4754]: I0126 17:31:40.951794 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e1e1571f-18fd-402c-b212-70635bbb1fb2-utilities\") pod \"redhat-operators-n69nh\" (UID: \"e1e1571f-18fd-402c-b212-70635bbb1fb2\") " pod="openshift-marketplace/redhat-operators-n69nh" Jan 26 17:31:40 crc kubenswrapper[4754]: I0126 17:31:40.952022 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e1e1571f-18fd-402c-b212-70635bbb1fb2-catalog-content\") pod \"redhat-operators-n69nh\" (UID: \"e1e1571f-18fd-402c-b212-70635bbb1fb2\") " pod="openshift-marketplace/redhat-operators-n69nh" Jan 26 17:31:40 crc kubenswrapper[4754]: I0126 17:31:40.970185 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-km4pj\" (UniqueName: \"kubernetes.io/projected/e1e1571f-18fd-402c-b212-70635bbb1fb2-kube-api-access-km4pj\") pod \"redhat-operators-n69nh\" (UID: \"e1e1571f-18fd-402c-b212-70635bbb1fb2\") " pod="openshift-marketplace/redhat-operators-n69nh" Jan 26 17:31:41 crc kubenswrapper[4754]: I0126 17:31:41.067984 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-n69nh" Jan 26 17:31:41 crc kubenswrapper[4754]: I0126 17:31:41.618362 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-n69nh"] Jan 26 17:31:42 crc kubenswrapper[4754]: I0126 17:31:42.109718 4754 generic.go:334] "Generic (PLEG): container finished" podID="e1e1571f-18fd-402c-b212-70635bbb1fb2" containerID="506cf3d5a90041c2f08e2700b4371d48ea643c6c7343afa62486a23177167427" exitCode=0 Jan 26 17:31:42 crc kubenswrapper[4754]: I0126 17:31:42.109768 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n69nh" event={"ID":"e1e1571f-18fd-402c-b212-70635bbb1fb2","Type":"ContainerDied","Data":"506cf3d5a90041c2f08e2700b4371d48ea643c6c7343afa62486a23177167427"} Jan 26 17:31:42 crc kubenswrapper[4754]: I0126 17:31:42.109804 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n69nh" event={"ID":"e1e1571f-18fd-402c-b212-70635bbb1fb2","Type":"ContainerStarted","Data":"a2ac043784bc5caea1d75aef38caeb182f4cd45bc5f607b1852b68edd5154223"} Jan 26 17:31:43 crc kubenswrapper[4754]: I0126 17:31:43.120991 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n69nh" event={"ID":"e1e1571f-18fd-402c-b212-70635bbb1fb2","Type":"ContainerStarted","Data":"9c4e4f002f6f00c6d5b1ff9b4558808dfb64b93e1cd488f57bff582355f00b8e"} Jan 26 17:31:46 crc kubenswrapper[4754]: I0126 17:31:46.146957 4754 generic.go:334] "Generic (PLEG): container finished" podID="e1e1571f-18fd-402c-b212-70635bbb1fb2" containerID="9c4e4f002f6f00c6d5b1ff9b4558808dfb64b93e1cd488f57bff582355f00b8e" exitCode=0 Jan 26 17:31:46 crc kubenswrapper[4754]: I0126 17:31:46.147015 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n69nh" event={"ID":"e1e1571f-18fd-402c-b212-70635bbb1fb2","Type":"ContainerDied","Data":"9c4e4f002f6f00c6d5b1ff9b4558808dfb64b93e1cd488f57bff582355f00b8e"} Jan 26 17:31:47 crc kubenswrapper[4754]: I0126 17:31:47.157339 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n69nh" event={"ID":"e1e1571f-18fd-402c-b212-70635bbb1fb2","Type":"ContainerStarted","Data":"63b879fde0367ba80cafee073294ca5fdc0931ef0861a3c3cb437c494bd0d3b5"} Jan 26 17:31:47 crc kubenswrapper[4754]: I0126 17:31:47.180983 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-n69nh" podStartSLOduration=2.631527338 podStartE2EDuration="7.180960731s" podCreationTimestamp="2026-01-26 17:31:40 +0000 UTC" firstStartedPulling="2026-01-26 17:31:42.111372384 +0000 UTC m=+2668.635552818" lastFinishedPulling="2026-01-26 17:31:46.660805777 +0000 UTC m=+2673.184986211" observedRunningTime="2026-01-26 17:31:47.174989886 +0000 UTC m=+2673.699170340" watchObservedRunningTime="2026-01-26 17:31:47.180960731 +0000 UTC m=+2673.705141165" Jan 26 17:31:51 crc kubenswrapper[4754]: I0126 17:31:51.068644 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-n69nh" Jan 26 17:31:51 crc kubenswrapper[4754]: I0126 17:31:51.069222 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-n69nh" Jan 26 17:31:52 crc kubenswrapper[4754]: I0126 17:31:52.115726 4754 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-n69nh" podUID="e1e1571f-18fd-402c-b212-70635bbb1fb2" containerName="registry-server" probeResult="failure" output=< Jan 26 17:31:52 crc kubenswrapper[4754]: timeout: failed to connect service ":50051" within 1s Jan 26 17:31:52 crc kubenswrapper[4754]: > Jan 26 17:32:01 crc kubenswrapper[4754]: I0126 17:32:01.124808 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-n69nh" Jan 26 17:32:01 crc kubenswrapper[4754]: I0126 17:32:01.179533 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-n69nh" Jan 26 17:32:01 crc kubenswrapper[4754]: I0126 17:32:01.364354 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-n69nh"] Jan 26 17:32:02 crc kubenswrapper[4754]: I0126 17:32:02.282222 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-n69nh" podUID="e1e1571f-18fd-402c-b212-70635bbb1fb2" containerName="registry-server" containerID="cri-o://63b879fde0367ba80cafee073294ca5fdc0931ef0861a3c3cb437c494bd0d3b5" gracePeriod=2 Jan 26 17:32:03 crc kubenswrapper[4754]: I0126 17:32:03.206005 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-n69nh" Jan 26 17:32:03 crc kubenswrapper[4754]: I0126 17:32:03.282045 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e1e1571f-18fd-402c-b212-70635bbb1fb2-utilities\") pod \"e1e1571f-18fd-402c-b212-70635bbb1fb2\" (UID: \"e1e1571f-18fd-402c-b212-70635bbb1fb2\") " Jan 26 17:32:03 crc kubenswrapper[4754]: I0126 17:32:03.282131 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e1e1571f-18fd-402c-b212-70635bbb1fb2-catalog-content\") pod \"e1e1571f-18fd-402c-b212-70635bbb1fb2\" (UID: \"e1e1571f-18fd-402c-b212-70635bbb1fb2\") " Jan 26 17:32:03 crc kubenswrapper[4754]: I0126 17:32:03.282226 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-km4pj\" (UniqueName: \"kubernetes.io/projected/e1e1571f-18fd-402c-b212-70635bbb1fb2-kube-api-access-km4pj\") pod \"e1e1571f-18fd-402c-b212-70635bbb1fb2\" (UID: \"e1e1571f-18fd-402c-b212-70635bbb1fb2\") " Jan 26 17:32:03 crc kubenswrapper[4754]: I0126 17:32:03.283056 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e1e1571f-18fd-402c-b212-70635bbb1fb2-utilities" (OuterVolumeSpecName: "utilities") pod "e1e1571f-18fd-402c-b212-70635bbb1fb2" (UID: "e1e1571f-18fd-402c-b212-70635bbb1fb2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:32:03 crc kubenswrapper[4754]: I0126 17:32:03.283657 4754 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e1e1571f-18fd-402c-b212-70635bbb1fb2-utilities\") on node \"crc\" DevicePath \"\"" Jan 26 17:32:03 crc kubenswrapper[4754]: I0126 17:32:03.288108 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e1e1571f-18fd-402c-b212-70635bbb1fb2-kube-api-access-km4pj" (OuterVolumeSpecName: "kube-api-access-km4pj") pod "e1e1571f-18fd-402c-b212-70635bbb1fb2" (UID: "e1e1571f-18fd-402c-b212-70635bbb1fb2"). InnerVolumeSpecName "kube-api-access-km4pj". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:32:03 crc kubenswrapper[4754]: I0126 17:32:03.293726 4754 generic.go:334] "Generic (PLEG): container finished" podID="e1e1571f-18fd-402c-b212-70635bbb1fb2" containerID="63b879fde0367ba80cafee073294ca5fdc0931ef0861a3c3cb437c494bd0d3b5" exitCode=0 Jan 26 17:32:03 crc kubenswrapper[4754]: I0126 17:32:03.293773 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n69nh" event={"ID":"e1e1571f-18fd-402c-b212-70635bbb1fb2","Type":"ContainerDied","Data":"63b879fde0367ba80cafee073294ca5fdc0931ef0861a3c3cb437c494bd0d3b5"} Jan 26 17:32:03 crc kubenswrapper[4754]: I0126 17:32:03.293800 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n69nh" event={"ID":"e1e1571f-18fd-402c-b212-70635bbb1fb2","Type":"ContainerDied","Data":"a2ac043784bc5caea1d75aef38caeb182f4cd45bc5f607b1852b68edd5154223"} Jan 26 17:32:03 crc kubenswrapper[4754]: I0126 17:32:03.293818 4754 scope.go:117] "RemoveContainer" containerID="63b879fde0367ba80cafee073294ca5fdc0931ef0861a3c3cb437c494bd0d3b5" Jan 26 17:32:03 crc kubenswrapper[4754]: I0126 17:32:03.293994 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-n69nh" Jan 26 17:32:03 crc kubenswrapper[4754]: I0126 17:32:03.345365 4754 scope.go:117] "RemoveContainer" containerID="9c4e4f002f6f00c6d5b1ff9b4558808dfb64b93e1cd488f57bff582355f00b8e" Jan 26 17:32:03 crc kubenswrapper[4754]: I0126 17:32:03.369939 4754 scope.go:117] "RemoveContainer" containerID="506cf3d5a90041c2f08e2700b4371d48ea643c6c7343afa62486a23177167427" Jan 26 17:32:03 crc kubenswrapper[4754]: I0126 17:32:03.386011 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-km4pj\" (UniqueName: \"kubernetes.io/projected/e1e1571f-18fd-402c-b212-70635bbb1fb2-kube-api-access-km4pj\") on node \"crc\" DevicePath \"\"" Jan 26 17:32:03 crc kubenswrapper[4754]: I0126 17:32:03.408627 4754 scope.go:117] "RemoveContainer" containerID="63b879fde0367ba80cafee073294ca5fdc0931ef0861a3c3cb437c494bd0d3b5" Jan 26 17:32:03 crc kubenswrapper[4754]: E0126 17:32:03.409439 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"63b879fde0367ba80cafee073294ca5fdc0931ef0861a3c3cb437c494bd0d3b5\": container with ID starting with 63b879fde0367ba80cafee073294ca5fdc0931ef0861a3c3cb437c494bd0d3b5 not found: ID does not exist" containerID="63b879fde0367ba80cafee073294ca5fdc0931ef0861a3c3cb437c494bd0d3b5" Jan 26 17:32:03 crc kubenswrapper[4754]: I0126 17:32:03.409474 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"63b879fde0367ba80cafee073294ca5fdc0931ef0861a3c3cb437c494bd0d3b5"} err="failed to get container status \"63b879fde0367ba80cafee073294ca5fdc0931ef0861a3c3cb437c494bd0d3b5\": rpc error: code = NotFound desc = could not find container \"63b879fde0367ba80cafee073294ca5fdc0931ef0861a3c3cb437c494bd0d3b5\": container with ID starting with 63b879fde0367ba80cafee073294ca5fdc0931ef0861a3c3cb437c494bd0d3b5 not found: ID does not exist" Jan 26 17:32:03 crc kubenswrapper[4754]: I0126 17:32:03.409505 4754 scope.go:117] "RemoveContainer" containerID="9c4e4f002f6f00c6d5b1ff9b4558808dfb64b93e1cd488f57bff582355f00b8e" Jan 26 17:32:03 crc kubenswrapper[4754]: E0126 17:32:03.409912 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9c4e4f002f6f00c6d5b1ff9b4558808dfb64b93e1cd488f57bff582355f00b8e\": container with ID starting with 9c4e4f002f6f00c6d5b1ff9b4558808dfb64b93e1cd488f57bff582355f00b8e not found: ID does not exist" containerID="9c4e4f002f6f00c6d5b1ff9b4558808dfb64b93e1cd488f57bff582355f00b8e" Jan 26 17:32:03 crc kubenswrapper[4754]: I0126 17:32:03.409934 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9c4e4f002f6f00c6d5b1ff9b4558808dfb64b93e1cd488f57bff582355f00b8e"} err="failed to get container status \"9c4e4f002f6f00c6d5b1ff9b4558808dfb64b93e1cd488f57bff582355f00b8e\": rpc error: code = NotFound desc = could not find container \"9c4e4f002f6f00c6d5b1ff9b4558808dfb64b93e1cd488f57bff582355f00b8e\": container with ID starting with 9c4e4f002f6f00c6d5b1ff9b4558808dfb64b93e1cd488f57bff582355f00b8e not found: ID does not exist" Jan 26 17:32:03 crc kubenswrapper[4754]: I0126 17:32:03.409947 4754 scope.go:117] "RemoveContainer" containerID="506cf3d5a90041c2f08e2700b4371d48ea643c6c7343afa62486a23177167427" Jan 26 17:32:03 crc kubenswrapper[4754]: E0126 17:32:03.410241 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"506cf3d5a90041c2f08e2700b4371d48ea643c6c7343afa62486a23177167427\": container with ID starting with 506cf3d5a90041c2f08e2700b4371d48ea643c6c7343afa62486a23177167427 not found: ID does not exist" containerID="506cf3d5a90041c2f08e2700b4371d48ea643c6c7343afa62486a23177167427" Jan 26 17:32:03 crc kubenswrapper[4754]: I0126 17:32:03.410286 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"506cf3d5a90041c2f08e2700b4371d48ea643c6c7343afa62486a23177167427"} err="failed to get container status \"506cf3d5a90041c2f08e2700b4371d48ea643c6c7343afa62486a23177167427\": rpc error: code = NotFound desc = could not find container \"506cf3d5a90041c2f08e2700b4371d48ea643c6c7343afa62486a23177167427\": container with ID starting with 506cf3d5a90041c2f08e2700b4371d48ea643c6c7343afa62486a23177167427 not found: ID does not exist" Jan 26 17:32:03 crc kubenswrapper[4754]: I0126 17:32:03.415125 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e1e1571f-18fd-402c-b212-70635bbb1fb2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e1e1571f-18fd-402c-b212-70635bbb1fb2" (UID: "e1e1571f-18fd-402c-b212-70635bbb1fb2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:32:03 crc kubenswrapper[4754]: I0126 17:32:03.487594 4754 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e1e1571f-18fd-402c-b212-70635bbb1fb2-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 26 17:32:03 crc kubenswrapper[4754]: I0126 17:32:03.628347 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-n69nh"] Jan 26 17:32:03 crc kubenswrapper[4754]: I0126 17:32:03.637227 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-n69nh"] Jan 26 17:32:03 crc kubenswrapper[4754]: I0126 17:32:03.801716 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e1e1571f-18fd-402c-b212-70635bbb1fb2" path="/var/lib/kubelet/pods/e1e1571f-18fd-402c-b212-70635bbb1fb2/volumes" Jan 26 17:32:07 crc kubenswrapper[4754]: I0126 17:32:07.128988 4754 patch_prober.go:28] interesting pod/machine-config-daemon-x65wv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 26 17:32:07 crc kubenswrapper[4754]: I0126 17:32:07.129481 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 26 17:32:37 crc kubenswrapper[4754]: I0126 17:32:37.129103 4754 patch_prober.go:28] interesting pod/machine-config-daemon-x65wv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 26 17:32:37 crc kubenswrapper[4754]: I0126 17:32:37.130959 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 26 17:32:57 crc kubenswrapper[4754]: I0126 17:32:57.966230 4754 generic.go:334] "Generic (PLEG): container finished" podID="b5408b33-3b71-46ac-87a4-413ecb9614b0" containerID="ba97fdca943084140236a71b63afdaa65a1ed00ca09bc5f0dd10aeb80ef90c6d" exitCode=0 Jan 26 17:32:57 crc kubenswrapper[4754]: I0126 17:32:57.966327 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-jswg2" event={"ID":"b5408b33-3b71-46ac-87a4-413ecb9614b0","Type":"ContainerDied","Data":"ba97fdca943084140236a71b63afdaa65a1ed00ca09bc5f0dd10aeb80ef90c6d"} Jan 26 17:32:59 crc kubenswrapper[4754]: I0126 17:32:59.404778 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-jswg2" Jan 26 17:32:59 crc kubenswrapper[4754]: I0126 17:32:59.535731 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/b5408b33-3b71-46ac-87a4-413ecb9614b0-nova-migration-ssh-key-0\") pod \"b5408b33-3b71-46ac-87a4-413ecb9614b0\" (UID: \"b5408b33-3b71-46ac-87a4-413ecb9614b0\") " Jan 26 17:32:59 crc kubenswrapper[4754]: I0126 17:32:59.535797 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/b5408b33-3b71-46ac-87a4-413ecb9614b0-nova-extra-config-0\") pod \"b5408b33-3b71-46ac-87a4-413ecb9614b0\" (UID: \"b5408b33-3b71-46ac-87a4-413ecb9614b0\") " Jan 26 17:32:59 crc kubenswrapper[4754]: I0126 17:32:59.535829 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p5phm\" (UniqueName: \"kubernetes.io/projected/b5408b33-3b71-46ac-87a4-413ecb9614b0-kube-api-access-p5phm\") pod \"b5408b33-3b71-46ac-87a4-413ecb9614b0\" (UID: \"b5408b33-3b71-46ac-87a4-413ecb9614b0\") " Jan 26 17:32:59 crc kubenswrapper[4754]: I0126 17:32:59.535867 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/b5408b33-3b71-46ac-87a4-413ecb9614b0-nova-cell1-compute-config-0\") pod \"b5408b33-3b71-46ac-87a4-413ecb9614b0\" (UID: \"b5408b33-3b71-46ac-87a4-413ecb9614b0\") " Jan 26 17:32:59 crc kubenswrapper[4754]: I0126 17:32:59.535921 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/b5408b33-3b71-46ac-87a4-413ecb9614b0-ssh-key-openstack-edpm-ipam\") pod \"b5408b33-3b71-46ac-87a4-413ecb9614b0\" (UID: \"b5408b33-3b71-46ac-87a4-413ecb9614b0\") " Jan 26 17:32:59 crc kubenswrapper[4754]: I0126 17:32:59.535988 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5408b33-3b71-46ac-87a4-413ecb9614b0-nova-combined-ca-bundle\") pod \"b5408b33-3b71-46ac-87a4-413ecb9614b0\" (UID: \"b5408b33-3b71-46ac-87a4-413ecb9614b0\") " Jan 26 17:32:59 crc kubenswrapper[4754]: I0126 17:32:59.536057 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/b5408b33-3b71-46ac-87a4-413ecb9614b0-nova-migration-ssh-key-1\") pod \"b5408b33-3b71-46ac-87a4-413ecb9614b0\" (UID: \"b5408b33-3b71-46ac-87a4-413ecb9614b0\") " Jan 26 17:32:59 crc kubenswrapper[4754]: I0126 17:32:59.536122 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b5408b33-3b71-46ac-87a4-413ecb9614b0-inventory\") pod \"b5408b33-3b71-46ac-87a4-413ecb9614b0\" (UID: \"b5408b33-3b71-46ac-87a4-413ecb9614b0\") " Jan 26 17:32:59 crc kubenswrapper[4754]: I0126 17:32:59.536164 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/b5408b33-3b71-46ac-87a4-413ecb9614b0-nova-cell1-compute-config-1\") pod \"b5408b33-3b71-46ac-87a4-413ecb9614b0\" (UID: \"b5408b33-3b71-46ac-87a4-413ecb9614b0\") " Jan 26 17:32:59 crc kubenswrapper[4754]: I0126 17:32:59.545063 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b5408b33-3b71-46ac-87a4-413ecb9614b0-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "b5408b33-3b71-46ac-87a4-413ecb9614b0" (UID: "b5408b33-3b71-46ac-87a4-413ecb9614b0"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:32:59 crc kubenswrapper[4754]: I0126 17:32:59.545588 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b5408b33-3b71-46ac-87a4-413ecb9614b0-kube-api-access-p5phm" (OuterVolumeSpecName: "kube-api-access-p5phm") pod "b5408b33-3b71-46ac-87a4-413ecb9614b0" (UID: "b5408b33-3b71-46ac-87a4-413ecb9614b0"). InnerVolumeSpecName "kube-api-access-p5phm". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:32:59 crc kubenswrapper[4754]: I0126 17:32:59.567435 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b5408b33-3b71-46ac-87a4-413ecb9614b0-nova-extra-config-0" (OuterVolumeSpecName: "nova-extra-config-0") pod "b5408b33-3b71-46ac-87a4-413ecb9614b0" (UID: "b5408b33-3b71-46ac-87a4-413ecb9614b0"). InnerVolumeSpecName "nova-extra-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:32:59 crc kubenswrapper[4754]: I0126 17:32:59.581830 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b5408b33-3b71-46ac-87a4-413ecb9614b0-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "b5408b33-3b71-46ac-87a4-413ecb9614b0" (UID: "b5408b33-3b71-46ac-87a4-413ecb9614b0"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:32:59 crc kubenswrapper[4754]: I0126 17:32:59.582880 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b5408b33-3b71-46ac-87a4-413ecb9614b0-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "b5408b33-3b71-46ac-87a4-413ecb9614b0" (UID: "b5408b33-3b71-46ac-87a4-413ecb9614b0"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:32:59 crc kubenswrapper[4754]: I0126 17:32:59.583368 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b5408b33-3b71-46ac-87a4-413ecb9614b0-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "b5408b33-3b71-46ac-87a4-413ecb9614b0" (UID: "b5408b33-3b71-46ac-87a4-413ecb9614b0"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:32:59 crc kubenswrapper[4754]: I0126 17:32:59.584410 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b5408b33-3b71-46ac-87a4-413ecb9614b0-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "b5408b33-3b71-46ac-87a4-413ecb9614b0" (UID: "b5408b33-3b71-46ac-87a4-413ecb9614b0"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:32:59 crc kubenswrapper[4754]: I0126 17:32:59.588236 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b5408b33-3b71-46ac-87a4-413ecb9614b0-inventory" (OuterVolumeSpecName: "inventory") pod "b5408b33-3b71-46ac-87a4-413ecb9614b0" (UID: "b5408b33-3b71-46ac-87a4-413ecb9614b0"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:32:59 crc kubenswrapper[4754]: I0126 17:32:59.595143 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b5408b33-3b71-46ac-87a4-413ecb9614b0-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "b5408b33-3b71-46ac-87a4-413ecb9614b0" (UID: "b5408b33-3b71-46ac-87a4-413ecb9614b0"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:32:59 crc kubenswrapper[4754]: I0126 17:32:59.637781 4754 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/b5408b33-3b71-46ac-87a4-413ecb9614b0-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Jan 26 17:32:59 crc kubenswrapper[4754]: I0126 17:32:59.637820 4754 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/b5408b33-3b71-46ac-87a4-413ecb9614b0-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Jan 26 17:32:59 crc kubenswrapper[4754]: I0126 17:32:59.637831 4754 reconciler_common.go:293] "Volume detached for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/b5408b33-3b71-46ac-87a4-413ecb9614b0-nova-extra-config-0\") on node \"crc\" DevicePath \"\"" Jan 26 17:32:59 crc kubenswrapper[4754]: I0126 17:32:59.637840 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p5phm\" (UniqueName: \"kubernetes.io/projected/b5408b33-3b71-46ac-87a4-413ecb9614b0-kube-api-access-p5phm\") on node \"crc\" DevicePath \"\"" Jan 26 17:32:59 crc kubenswrapper[4754]: I0126 17:32:59.637849 4754 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/b5408b33-3b71-46ac-87a4-413ecb9614b0-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Jan 26 17:32:59 crc kubenswrapper[4754]: I0126 17:32:59.637856 4754 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/b5408b33-3b71-46ac-87a4-413ecb9614b0-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Jan 26 17:32:59 crc kubenswrapper[4754]: I0126 17:32:59.637865 4754 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5408b33-3b71-46ac-87a4-413ecb9614b0-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 26 17:32:59 crc kubenswrapper[4754]: I0126 17:32:59.637877 4754 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/b5408b33-3b71-46ac-87a4-413ecb9614b0-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Jan 26 17:32:59 crc kubenswrapper[4754]: I0126 17:32:59.637887 4754 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b5408b33-3b71-46ac-87a4-413ecb9614b0-inventory\") on node \"crc\" DevicePath \"\"" Jan 26 17:32:59 crc kubenswrapper[4754]: I0126 17:32:59.990037 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-jswg2" event={"ID":"b5408b33-3b71-46ac-87a4-413ecb9614b0","Type":"ContainerDied","Data":"2216d89e7a7b851415c7080c6d481507a0bdd04990cdf0408576e6e7469f6652"} Jan 26 17:32:59 crc kubenswrapper[4754]: I0126 17:32:59.990081 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2216d89e7a7b851415c7080c6d481507a0bdd04990cdf0408576e6e7469f6652" Jan 26 17:32:59 crc kubenswrapper[4754]: I0126 17:32:59.990096 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-jswg2" Jan 26 17:33:00 crc kubenswrapper[4754]: I0126 17:33:00.082904 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-59vz4"] Jan 26 17:33:00 crc kubenswrapper[4754]: E0126 17:33:00.083445 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1e1571f-18fd-402c-b212-70635bbb1fb2" containerName="extract-content" Jan 26 17:33:00 crc kubenswrapper[4754]: I0126 17:33:00.083472 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1e1571f-18fd-402c-b212-70635bbb1fb2" containerName="extract-content" Jan 26 17:33:00 crc kubenswrapper[4754]: E0126 17:33:00.083485 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1e1571f-18fd-402c-b212-70635bbb1fb2" containerName="registry-server" Jan 26 17:33:00 crc kubenswrapper[4754]: I0126 17:33:00.083493 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1e1571f-18fd-402c-b212-70635bbb1fb2" containerName="registry-server" Jan 26 17:33:00 crc kubenswrapper[4754]: E0126 17:33:00.083510 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1e1571f-18fd-402c-b212-70635bbb1fb2" containerName="extract-utilities" Jan 26 17:33:00 crc kubenswrapper[4754]: I0126 17:33:00.083518 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1e1571f-18fd-402c-b212-70635bbb1fb2" containerName="extract-utilities" Jan 26 17:33:00 crc kubenswrapper[4754]: E0126 17:33:00.083547 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5408b33-3b71-46ac-87a4-413ecb9614b0" containerName="nova-edpm-deployment-openstack-edpm-ipam" Jan 26 17:33:00 crc kubenswrapper[4754]: I0126 17:33:00.083555 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5408b33-3b71-46ac-87a4-413ecb9614b0" containerName="nova-edpm-deployment-openstack-edpm-ipam" Jan 26 17:33:00 crc kubenswrapper[4754]: I0126 17:33:00.083790 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="e1e1571f-18fd-402c-b212-70635bbb1fb2" containerName="registry-server" Jan 26 17:33:00 crc kubenswrapper[4754]: I0126 17:33:00.083817 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="b5408b33-3b71-46ac-87a4-413ecb9614b0" containerName="nova-edpm-deployment-openstack-edpm-ipam" Jan 26 17:33:00 crc kubenswrapper[4754]: I0126 17:33:00.084573 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-59vz4" Jan 26 17:33:00 crc kubenswrapper[4754]: I0126 17:33:00.091961 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-compute-config-data" Jan 26 17:33:00 crc kubenswrapper[4754]: I0126 17:33:00.091984 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Jan 26 17:33:00 crc kubenswrapper[4754]: I0126 17:33:00.092046 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-fz7z8" Jan 26 17:33:00 crc kubenswrapper[4754]: I0126 17:33:00.092174 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Jan 26 17:33:00 crc kubenswrapper[4754]: I0126 17:33:00.092202 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Jan 26 17:33:00 crc kubenswrapper[4754]: I0126 17:33:00.114467 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-59vz4"] Jan 26 17:33:00 crc kubenswrapper[4754]: I0126 17:33:00.146494 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tsfmz\" (UniqueName: \"kubernetes.io/projected/5fade260-b289-4c35-b77e-ca9f8ec754c3-kube-api-access-tsfmz\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-59vz4\" (UID: \"5fade260-b289-4c35-b77e-ca9f8ec754c3\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-59vz4" Jan 26 17:33:00 crc kubenswrapper[4754]: I0126 17:33:00.147833 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/5fade260-b289-4c35-b77e-ca9f8ec754c3-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-59vz4\" (UID: \"5fade260-b289-4c35-b77e-ca9f8ec754c3\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-59vz4" Jan 26 17:33:00 crc kubenswrapper[4754]: I0126 17:33:00.148084 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/5fade260-b289-4c35-b77e-ca9f8ec754c3-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-59vz4\" (UID: \"5fade260-b289-4c35-b77e-ca9f8ec754c3\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-59vz4" Jan 26 17:33:00 crc kubenswrapper[4754]: I0126 17:33:00.148196 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5fade260-b289-4c35-b77e-ca9f8ec754c3-ssh-key-openstack-edpm-ipam\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-59vz4\" (UID: \"5fade260-b289-4c35-b77e-ca9f8ec754c3\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-59vz4" Jan 26 17:33:00 crc kubenswrapper[4754]: I0126 17:33:00.148284 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5fade260-b289-4c35-b77e-ca9f8ec754c3-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-59vz4\" (UID: \"5fade260-b289-4c35-b77e-ca9f8ec754c3\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-59vz4" Jan 26 17:33:00 crc kubenswrapper[4754]: I0126 17:33:00.148427 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/5fade260-b289-4c35-b77e-ca9f8ec754c3-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-59vz4\" (UID: \"5fade260-b289-4c35-b77e-ca9f8ec754c3\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-59vz4" Jan 26 17:33:00 crc kubenswrapper[4754]: I0126 17:33:00.148556 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5fade260-b289-4c35-b77e-ca9f8ec754c3-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-59vz4\" (UID: \"5fade260-b289-4c35-b77e-ca9f8ec754c3\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-59vz4" Jan 26 17:33:00 crc kubenswrapper[4754]: I0126 17:33:00.249905 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5fade260-b289-4c35-b77e-ca9f8ec754c3-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-59vz4\" (UID: \"5fade260-b289-4c35-b77e-ca9f8ec754c3\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-59vz4" Jan 26 17:33:00 crc kubenswrapper[4754]: I0126 17:33:00.249975 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tsfmz\" (UniqueName: \"kubernetes.io/projected/5fade260-b289-4c35-b77e-ca9f8ec754c3-kube-api-access-tsfmz\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-59vz4\" (UID: \"5fade260-b289-4c35-b77e-ca9f8ec754c3\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-59vz4" Jan 26 17:33:00 crc kubenswrapper[4754]: I0126 17:33:00.250008 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/5fade260-b289-4c35-b77e-ca9f8ec754c3-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-59vz4\" (UID: \"5fade260-b289-4c35-b77e-ca9f8ec754c3\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-59vz4" Jan 26 17:33:00 crc kubenswrapper[4754]: I0126 17:33:00.250075 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/5fade260-b289-4c35-b77e-ca9f8ec754c3-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-59vz4\" (UID: \"5fade260-b289-4c35-b77e-ca9f8ec754c3\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-59vz4" Jan 26 17:33:00 crc kubenswrapper[4754]: I0126 17:33:00.250101 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5fade260-b289-4c35-b77e-ca9f8ec754c3-ssh-key-openstack-edpm-ipam\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-59vz4\" (UID: \"5fade260-b289-4c35-b77e-ca9f8ec754c3\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-59vz4" Jan 26 17:33:00 crc kubenswrapper[4754]: I0126 17:33:00.250126 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5fade260-b289-4c35-b77e-ca9f8ec754c3-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-59vz4\" (UID: \"5fade260-b289-4c35-b77e-ca9f8ec754c3\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-59vz4" Jan 26 17:33:00 crc kubenswrapper[4754]: I0126 17:33:00.250175 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/5fade260-b289-4c35-b77e-ca9f8ec754c3-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-59vz4\" (UID: \"5fade260-b289-4c35-b77e-ca9f8ec754c3\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-59vz4" Jan 26 17:33:00 crc kubenswrapper[4754]: I0126 17:33:00.255101 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/5fade260-b289-4c35-b77e-ca9f8ec754c3-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-59vz4\" (UID: \"5fade260-b289-4c35-b77e-ca9f8ec754c3\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-59vz4" Jan 26 17:33:00 crc kubenswrapper[4754]: I0126 17:33:00.256050 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/5fade260-b289-4c35-b77e-ca9f8ec754c3-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-59vz4\" (UID: \"5fade260-b289-4c35-b77e-ca9f8ec754c3\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-59vz4" Jan 26 17:33:00 crc kubenswrapper[4754]: I0126 17:33:00.256339 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5fade260-b289-4c35-b77e-ca9f8ec754c3-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-59vz4\" (UID: \"5fade260-b289-4c35-b77e-ca9f8ec754c3\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-59vz4" Jan 26 17:33:00 crc kubenswrapper[4754]: I0126 17:33:00.256526 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5fade260-b289-4c35-b77e-ca9f8ec754c3-ssh-key-openstack-edpm-ipam\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-59vz4\" (UID: \"5fade260-b289-4c35-b77e-ca9f8ec754c3\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-59vz4" Jan 26 17:33:00 crc kubenswrapper[4754]: I0126 17:33:00.256631 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/5fade260-b289-4c35-b77e-ca9f8ec754c3-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-59vz4\" (UID: \"5fade260-b289-4c35-b77e-ca9f8ec754c3\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-59vz4" Jan 26 17:33:00 crc kubenswrapper[4754]: I0126 17:33:00.257450 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5fade260-b289-4c35-b77e-ca9f8ec754c3-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-59vz4\" (UID: \"5fade260-b289-4c35-b77e-ca9f8ec754c3\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-59vz4" Jan 26 17:33:00 crc kubenswrapper[4754]: I0126 17:33:00.265384 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tsfmz\" (UniqueName: \"kubernetes.io/projected/5fade260-b289-4c35-b77e-ca9f8ec754c3-kube-api-access-tsfmz\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-59vz4\" (UID: \"5fade260-b289-4c35-b77e-ca9f8ec754c3\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-59vz4" Jan 26 17:33:00 crc kubenswrapper[4754]: I0126 17:33:00.407836 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-59vz4" Jan 26 17:33:00 crc kubenswrapper[4754]: I0126 17:33:00.930976 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-59vz4"] Jan 26 17:33:00 crc kubenswrapper[4754]: I0126 17:33:00.999114 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-59vz4" event={"ID":"5fade260-b289-4c35-b77e-ca9f8ec754c3","Type":"ContainerStarted","Data":"58b810266bda80a7a3f9af826f449ae0be7384d7256e04bb2ffa71d4441625f7"} Jan 26 17:33:03 crc kubenswrapper[4754]: I0126 17:33:03.021801 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-59vz4" event={"ID":"5fade260-b289-4c35-b77e-ca9f8ec754c3","Type":"ContainerStarted","Data":"fca7458db5b576d8e8083ee721d60295d3124bc9455683e170a9c3b4845a3051"} Jan 26 17:33:03 crc kubenswrapper[4754]: I0126 17:33:03.045129 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-59vz4" podStartSLOduration=1.877275093 podStartE2EDuration="3.04510461s" podCreationTimestamp="2026-01-26 17:33:00 +0000 UTC" firstStartedPulling="2026-01-26 17:33:00.935766865 +0000 UTC m=+2747.459947299" lastFinishedPulling="2026-01-26 17:33:02.103596382 +0000 UTC m=+2748.627776816" observedRunningTime="2026-01-26 17:33:03.041548082 +0000 UTC m=+2749.565728526" watchObservedRunningTime="2026-01-26 17:33:03.04510461 +0000 UTC m=+2749.569285044" Jan 26 17:33:07 crc kubenswrapper[4754]: I0126 17:33:07.129055 4754 patch_prober.go:28] interesting pod/machine-config-daemon-x65wv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 26 17:33:07 crc kubenswrapper[4754]: I0126 17:33:07.129591 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 26 17:33:07 crc kubenswrapper[4754]: I0126 17:33:07.129634 4754 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" Jan 26 17:33:07 crc kubenswrapper[4754]: I0126 17:33:07.130352 4754 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"efeb34d7b9f84cbdd86409079c972f84a47710b40532d5bb54abdaa0f23b0f5d"} pod="openshift-machine-config-operator/machine-config-daemon-x65wv" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 26 17:33:07 crc kubenswrapper[4754]: I0126 17:33:07.130399 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" containerName="machine-config-daemon" containerID="cri-o://efeb34d7b9f84cbdd86409079c972f84a47710b40532d5bb54abdaa0f23b0f5d" gracePeriod=600 Jan 26 17:33:08 crc kubenswrapper[4754]: I0126 17:33:08.074846 4754 generic.go:334] "Generic (PLEG): container finished" podID="8c3718a4-f354-4284-92e0-fdfb45a692bd" containerID="efeb34d7b9f84cbdd86409079c972f84a47710b40532d5bb54abdaa0f23b0f5d" exitCode=0 Jan 26 17:33:08 crc kubenswrapper[4754]: I0126 17:33:08.074924 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" event={"ID":"8c3718a4-f354-4284-92e0-fdfb45a692bd","Type":"ContainerDied","Data":"efeb34d7b9f84cbdd86409079c972f84a47710b40532d5bb54abdaa0f23b0f5d"} Jan 26 17:33:08 crc kubenswrapper[4754]: I0126 17:33:08.075266 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" event={"ID":"8c3718a4-f354-4284-92e0-fdfb45a692bd","Type":"ContainerStarted","Data":"2a52617105aca80bd6aa782680de53fad558b6014f1d373ef1592509df2dabc0"} Jan 26 17:33:08 crc kubenswrapper[4754]: I0126 17:33:08.075295 4754 scope.go:117] "RemoveContainer" containerID="00e15c82b01f7eb3c11968e396c0748604b39b71891170fde13a876ab0a851b3" Jan 26 17:33:39 crc kubenswrapper[4754]: I0126 17:33:39.299021 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-6nwcr"] Jan 26 17:33:39 crc kubenswrapper[4754]: I0126 17:33:39.303132 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6nwcr" Jan 26 17:33:39 crc kubenswrapper[4754]: I0126 17:33:39.308601 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-6nwcr"] Jan 26 17:33:39 crc kubenswrapper[4754]: I0126 17:33:39.484837 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5b54a1ca-9c37-4258-863b-10141af1389d-catalog-content\") pod \"community-operators-6nwcr\" (UID: \"5b54a1ca-9c37-4258-863b-10141af1389d\") " pod="openshift-marketplace/community-operators-6nwcr" Jan 26 17:33:39 crc kubenswrapper[4754]: I0126 17:33:39.484952 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5b54a1ca-9c37-4258-863b-10141af1389d-utilities\") pod \"community-operators-6nwcr\" (UID: \"5b54a1ca-9c37-4258-863b-10141af1389d\") " pod="openshift-marketplace/community-operators-6nwcr" Jan 26 17:33:39 crc kubenswrapper[4754]: I0126 17:33:39.485000 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2g94s\" (UniqueName: \"kubernetes.io/projected/5b54a1ca-9c37-4258-863b-10141af1389d-kube-api-access-2g94s\") pod \"community-operators-6nwcr\" (UID: \"5b54a1ca-9c37-4258-863b-10141af1389d\") " pod="openshift-marketplace/community-operators-6nwcr" Jan 26 17:33:39 crc kubenswrapper[4754]: I0126 17:33:39.586237 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5b54a1ca-9c37-4258-863b-10141af1389d-catalog-content\") pod \"community-operators-6nwcr\" (UID: \"5b54a1ca-9c37-4258-863b-10141af1389d\") " pod="openshift-marketplace/community-operators-6nwcr" Jan 26 17:33:39 crc kubenswrapper[4754]: I0126 17:33:39.586406 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5b54a1ca-9c37-4258-863b-10141af1389d-utilities\") pod \"community-operators-6nwcr\" (UID: \"5b54a1ca-9c37-4258-863b-10141af1389d\") " pod="openshift-marketplace/community-operators-6nwcr" Jan 26 17:33:39 crc kubenswrapper[4754]: I0126 17:33:39.586479 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2g94s\" (UniqueName: \"kubernetes.io/projected/5b54a1ca-9c37-4258-863b-10141af1389d-kube-api-access-2g94s\") pod \"community-operators-6nwcr\" (UID: \"5b54a1ca-9c37-4258-863b-10141af1389d\") " pod="openshift-marketplace/community-operators-6nwcr" Jan 26 17:33:39 crc kubenswrapper[4754]: I0126 17:33:39.586765 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5b54a1ca-9c37-4258-863b-10141af1389d-catalog-content\") pod \"community-operators-6nwcr\" (UID: \"5b54a1ca-9c37-4258-863b-10141af1389d\") " pod="openshift-marketplace/community-operators-6nwcr" Jan 26 17:33:39 crc kubenswrapper[4754]: I0126 17:33:39.586947 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5b54a1ca-9c37-4258-863b-10141af1389d-utilities\") pod \"community-operators-6nwcr\" (UID: \"5b54a1ca-9c37-4258-863b-10141af1389d\") " pod="openshift-marketplace/community-operators-6nwcr" Jan 26 17:33:39 crc kubenswrapper[4754]: I0126 17:33:39.624761 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2g94s\" (UniqueName: \"kubernetes.io/projected/5b54a1ca-9c37-4258-863b-10141af1389d-kube-api-access-2g94s\") pod \"community-operators-6nwcr\" (UID: \"5b54a1ca-9c37-4258-863b-10141af1389d\") " pod="openshift-marketplace/community-operators-6nwcr" Jan 26 17:33:39 crc kubenswrapper[4754]: I0126 17:33:39.923923 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6nwcr" Jan 26 17:33:40 crc kubenswrapper[4754]: I0126 17:33:40.351734 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-6nwcr"] Jan 26 17:33:41 crc kubenswrapper[4754]: I0126 17:33:41.359715 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6nwcr" event={"ID":"5b54a1ca-9c37-4258-863b-10141af1389d","Type":"ContainerStarted","Data":"0f2383cede51079981535ec7e11ec77ab4219d88948c099a19c35521085f9f85"} Jan 26 17:33:41 crc kubenswrapper[4754]: I0126 17:33:41.360172 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6nwcr" event={"ID":"5b54a1ca-9c37-4258-863b-10141af1389d","Type":"ContainerStarted","Data":"35d335c1a400182a4c4acdfb373b98e26605b69fac748d6ccb3429ee53770dc0"} Jan 26 17:33:42 crc kubenswrapper[4754]: I0126 17:33:42.373265 4754 generic.go:334] "Generic (PLEG): container finished" podID="5b54a1ca-9c37-4258-863b-10141af1389d" containerID="0f2383cede51079981535ec7e11ec77ab4219d88948c099a19c35521085f9f85" exitCode=0 Jan 26 17:33:42 crc kubenswrapper[4754]: I0126 17:33:42.373321 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6nwcr" event={"ID":"5b54a1ca-9c37-4258-863b-10141af1389d","Type":"ContainerDied","Data":"0f2383cede51079981535ec7e11ec77ab4219d88948c099a19c35521085f9f85"} Jan 26 17:33:43 crc kubenswrapper[4754]: I0126 17:33:43.386962 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6nwcr" event={"ID":"5b54a1ca-9c37-4258-863b-10141af1389d","Type":"ContainerStarted","Data":"da1e482fb6e405a990f6e40a10434173cf7000c3a12e2fa3659affec563e39f5"} Jan 26 17:33:44 crc kubenswrapper[4754]: I0126 17:33:44.398441 4754 generic.go:334] "Generic (PLEG): container finished" podID="5b54a1ca-9c37-4258-863b-10141af1389d" containerID="da1e482fb6e405a990f6e40a10434173cf7000c3a12e2fa3659affec563e39f5" exitCode=0 Jan 26 17:33:44 crc kubenswrapper[4754]: I0126 17:33:44.398489 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6nwcr" event={"ID":"5b54a1ca-9c37-4258-863b-10141af1389d","Type":"ContainerDied","Data":"da1e482fb6e405a990f6e40a10434173cf7000c3a12e2fa3659affec563e39f5"} Jan 26 17:33:45 crc kubenswrapper[4754]: I0126 17:33:45.410519 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6nwcr" event={"ID":"5b54a1ca-9c37-4258-863b-10141af1389d","Type":"ContainerStarted","Data":"30c12f20b61a46798058b7be18e5bf22eb47041ab2f4a2606fa75d1fcd41e1f5"} Jan 26 17:33:45 crc kubenswrapper[4754]: I0126 17:33:45.435838 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-6nwcr" podStartSLOduration=4.041829023 podStartE2EDuration="6.435817472s" podCreationTimestamp="2026-01-26 17:33:39 +0000 UTC" firstStartedPulling="2026-01-26 17:33:42.37526558 +0000 UTC m=+2788.899446014" lastFinishedPulling="2026-01-26 17:33:44.769254019 +0000 UTC m=+2791.293434463" observedRunningTime="2026-01-26 17:33:45.427485801 +0000 UTC m=+2791.951666255" watchObservedRunningTime="2026-01-26 17:33:45.435817472 +0000 UTC m=+2791.959997906" Jan 26 17:33:49 crc kubenswrapper[4754]: I0126 17:33:49.925024 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-6nwcr" Jan 26 17:33:49 crc kubenswrapper[4754]: I0126 17:33:49.925485 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-6nwcr" Jan 26 17:33:49 crc kubenswrapper[4754]: I0126 17:33:49.973457 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-6nwcr" Jan 26 17:33:50 crc kubenswrapper[4754]: I0126 17:33:50.507164 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-6nwcr" Jan 26 17:33:50 crc kubenswrapper[4754]: I0126 17:33:50.549330 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-6nwcr"] Jan 26 17:33:52 crc kubenswrapper[4754]: I0126 17:33:52.468590 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-6nwcr" podUID="5b54a1ca-9c37-4258-863b-10141af1389d" containerName="registry-server" containerID="cri-o://30c12f20b61a46798058b7be18e5bf22eb47041ab2f4a2606fa75d1fcd41e1f5" gracePeriod=2 Jan 26 17:33:53 crc kubenswrapper[4754]: I0126 17:33:53.435956 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6nwcr" Jan 26 17:33:53 crc kubenswrapper[4754]: I0126 17:33:53.479313 4754 generic.go:334] "Generic (PLEG): container finished" podID="5b54a1ca-9c37-4258-863b-10141af1389d" containerID="30c12f20b61a46798058b7be18e5bf22eb47041ab2f4a2606fa75d1fcd41e1f5" exitCode=0 Jan 26 17:33:53 crc kubenswrapper[4754]: I0126 17:33:53.479360 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6nwcr" event={"ID":"5b54a1ca-9c37-4258-863b-10141af1389d","Type":"ContainerDied","Data":"30c12f20b61a46798058b7be18e5bf22eb47041ab2f4a2606fa75d1fcd41e1f5"} Jan 26 17:33:53 crc kubenswrapper[4754]: I0126 17:33:53.479397 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6nwcr" event={"ID":"5b54a1ca-9c37-4258-863b-10141af1389d","Type":"ContainerDied","Data":"35d335c1a400182a4c4acdfb373b98e26605b69fac748d6ccb3429ee53770dc0"} Jan 26 17:33:53 crc kubenswrapper[4754]: I0126 17:33:53.479414 4754 scope.go:117] "RemoveContainer" containerID="30c12f20b61a46798058b7be18e5bf22eb47041ab2f4a2606fa75d1fcd41e1f5" Jan 26 17:33:53 crc kubenswrapper[4754]: I0126 17:33:53.479426 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6nwcr" Jan 26 17:33:53 crc kubenswrapper[4754]: I0126 17:33:53.505931 4754 scope.go:117] "RemoveContainer" containerID="da1e482fb6e405a990f6e40a10434173cf7000c3a12e2fa3659affec563e39f5" Jan 26 17:33:53 crc kubenswrapper[4754]: I0126 17:33:53.527604 4754 scope.go:117] "RemoveContainer" containerID="0f2383cede51079981535ec7e11ec77ab4219d88948c099a19c35521085f9f85" Jan 26 17:33:53 crc kubenswrapper[4754]: I0126 17:33:53.535559 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5b54a1ca-9c37-4258-863b-10141af1389d-utilities\") pod \"5b54a1ca-9c37-4258-863b-10141af1389d\" (UID: \"5b54a1ca-9c37-4258-863b-10141af1389d\") " Jan 26 17:33:53 crc kubenswrapper[4754]: I0126 17:33:53.535636 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5b54a1ca-9c37-4258-863b-10141af1389d-catalog-content\") pod \"5b54a1ca-9c37-4258-863b-10141af1389d\" (UID: \"5b54a1ca-9c37-4258-863b-10141af1389d\") " Jan 26 17:33:53 crc kubenswrapper[4754]: I0126 17:33:53.535804 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2g94s\" (UniqueName: \"kubernetes.io/projected/5b54a1ca-9c37-4258-863b-10141af1389d-kube-api-access-2g94s\") pod \"5b54a1ca-9c37-4258-863b-10141af1389d\" (UID: \"5b54a1ca-9c37-4258-863b-10141af1389d\") " Jan 26 17:33:53 crc kubenswrapper[4754]: I0126 17:33:53.536325 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5b54a1ca-9c37-4258-863b-10141af1389d-utilities" (OuterVolumeSpecName: "utilities") pod "5b54a1ca-9c37-4258-863b-10141af1389d" (UID: "5b54a1ca-9c37-4258-863b-10141af1389d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:33:53 crc kubenswrapper[4754]: I0126 17:33:53.543093 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b54a1ca-9c37-4258-863b-10141af1389d-kube-api-access-2g94s" (OuterVolumeSpecName: "kube-api-access-2g94s") pod "5b54a1ca-9c37-4258-863b-10141af1389d" (UID: "5b54a1ca-9c37-4258-863b-10141af1389d"). InnerVolumeSpecName "kube-api-access-2g94s". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:33:53 crc kubenswrapper[4754]: I0126 17:33:53.629254 4754 scope.go:117] "RemoveContainer" containerID="30c12f20b61a46798058b7be18e5bf22eb47041ab2f4a2606fa75d1fcd41e1f5" Jan 26 17:33:53 crc kubenswrapper[4754]: E0126 17:33:53.629752 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"30c12f20b61a46798058b7be18e5bf22eb47041ab2f4a2606fa75d1fcd41e1f5\": container with ID starting with 30c12f20b61a46798058b7be18e5bf22eb47041ab2f4a2606fa75d1fcd41e1f5 not found: ID does not exist" containerID="30c12f20b61a46798058b7be18e5bf22eb47041ab2f4a2606fa75d1fcd41e1f5" Jan 26 17:33:53 crc kubenswrapper[4754]: I0126 17:33:53.629785 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"30c12f20b61a46798058b7be18e5bf22eb47041ab2f4a2606fa75d1fcd41e1f5"} err="failed to get container status \"30c12f20b61a46798058b7be18e5bf22eb47041ab2f4a2606fa75d1fcd41e1f5\": rpc error: code = NotFound desc = could not find container \"30c12f20b61a46798058b7be18e5bf22eb47041ab2f4a2606fa75d1fcd41e1f5\": container with ID starting with 30c12f20b61a46798058b7be18e5bf22eb47041ab2f4a2606fa75d1fcd41e1f5 not found: ID does not exist" Jan 26 17:33:53 crc kubenswrapper[4754]: I0126 17:33:53.629806 4754 scope.go:117] "RemoveContainer" containerID="da1e482fb6e405a990f6e40a10434173cf7000c3a12e2fa3659affec563e39f5" Jan 26 17:33:53 crc kubenswrapper[4754]: E0126 17:33:53.630091 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"da1e482fb6e405a990f6e40a10434173cf7000c3a12e2fa3659affec563e39f5\": container with ID starting with da1e482fb6e405a990f6e40a10434173cf7000c3a12e2fa3659affec563e39f5 not found: ID does not exist" containerID="da1e482fb6e405a990f6e40a10434173cf7000c3a12e2fa3659affec563e39f5" Jan 26 17:33:53 crc kubenswrapper[4754]: I0126 17:33:53.630149 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"da1e482fb6e405a990f6e40a10434173cf7000c3a12e2fa3659affec563e39f5"} err="failed to get container status \"da1e482fb6e405a990f6e40a10434173cf7000c3a12e2fa3659affec563e39f5\": rpc error: code = NotFound desc = could not find container \"da1e482fb6e405a990f6e40a10434173cf7000c3a12e2fa3659affec563e39f5\": container with ID starting with da1e482fb6e405a990f6e40a10434173cf7000c3a12e2fa3659affec563e39f5 not found: ID does not exist" Jan 26 17:33:53 crc kubenswrapper[4754]: I0126 17:33:53.630165 4754 scope.go:117] "RemoveContainer" containerID="0f2383cede51079981535ec7e11ec77ab4219d88948c099a19c35521085f9f85" Jan 26 17:33:53 crc kubenswrapper[4754]: E0126 17:33:53.630398 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0f2383cede51079981535ec7e11ec77ab4219d88948c099a19c35521085f9f85\": container with ID starting with 0f2383cede51079981535ec7e11ec77ab4219d88948c099a19c35521085f9f85 not found: ID does not exist" containerID="0f2383cede51079981535ec7e11ec77ab4219d88948c099a19c35521085f9f85" Jan 26 17:33:53 crc kubenswrapper[4754]: I0126 17:33:53.630414 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0f2383cede51079981535ec7e11ec77ab4219d88948c099a19c35521085f9f85"} err="failed to get container status \"0f2383cede51079981535ec7e11ec77ab4219d88948c099a19c35521085f9f85\": rpc error: code = NotFound desc = could not find container \"0f2383cede51079981535ec7e11ec77ab4219d88948c099a19c35521085f9f85\": container with ID starting with 0f2383cede51079981535ec7e11ec77ab4219d88948c099a19c35521085f9f85 not found: ID does not exist" Jan 26 17:33:53 crc kubenswrapper[4754]: I0126 17:33:53.637579 4754 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5b54a1ca-9c37-4258-863b-10141af1389d-utilities\") on node \"crc\" DevicePath \"\"" Jan 26 17:33:53 crc kubenswrapper[4754]: I0126 17:33:53.637607 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2g94s\" (UniqueName: \"kubernetes.io/projected/5b54a1ca-9c37-4258-863b-10141af1389d-kube-api-access-2g94s\") on node \"crc\" DevicePath \"\"" Jan 26 17:33:53 crc kubenswrapper[4754]: I0126 17:33:53.834836 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5b54a1ca-9c37-4258-863b-10141af1389d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5b54a1ca-9c37-4258-863b-10141af1389d" (UID: "5b54a1ca-9c37-4258-863b-10141af1389d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:33:53 crc kubenswrapper[4754]: I0126 17:33:53.846688 4754 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5b54a1ca-9c37-4258-863b-10141af1389d-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 26 17:33:54 crc kubenswrapper[4754]: I0126 17:33:54.111451 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-6nwcr"] Jan 26 17:33:54 crc kubenswrapper[4754]: I0126 17:33:54.121743 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-6nwcr"] Jan 26 17:33:55 crc kubenswrapper[4754]: I0126 17:33:55.778819 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b54a1ca-9c37-4258-863b-10141af1389d" path="/var/lib/kubelet/pods/5b54a1ca-9c37-4258-863b-10141af1389d/volumes" Jan 26 17:35:29 crc kubenswrapper[4754]: I0126 17:35:29.344061 4754 generic.go:334] "Generic (PLEG): container finished" podID="5fade260-b289-4c35-b77e-ca9f8ec754c3" containerID="fca7458db5b576d8e8083ee721d60295d3124bc9455683e170a9c3b4845a3051" exitCode=0 Jan 26 17:35:29 crc kubenswrapper[4754]: I0126 17:35:29.344567 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-59vz4" event={"ID":"5fade260-b289-4c35-b77e-ca9f8ec754c3","Type":"ContainerDied","Data":"fca7458db5b576d8e8083ee721d60295d3124bc9455683e170a9c3b4845a3051"} Jan 26 17:35:30 crc kubenswrapper[4754]: I0126 17:35:30.857855 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-59vz4" Jan 26 17:35:30 crc kubenswrapper[4754]: I0126 17:35:30.904155 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/5fade260-b289-4c35-b77e-ca9f8ec754c3-ceilometer-compute-config-data-0\") pod \"5fade260-b289-4c35-b77e-ca9f8ec754c3\" (UID: \"5fade260-b289-4c35-b77e-ca9f8ec754c3\") " Jan 26 17:35:30 crc kubenswrapper[4754]: I0126 17:35:30.904227 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/5fade260-b289-4c35-b77e-ca9f8ec754c3-ceilometer-compute-config-data-2\") pod \"5fade260-b289-4c35-b77e-ca9f8ec754c3\" (UID: \"5fade260-b289-4c35-b77e-ca9f8ec754c3\") " Jan 26 17:35:30 crc kubenswrapper[4754]: I0126 17:35:30.904263 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5fade260-b289-4c35-b77e-ca9f8ec754c3-inventory\") pod \"5fade260-b289-4c35-b77e-ca9f8ec754c3\" (UID: \"5fade260-b289-4c35-b77e-ca9f8ec754c3\") " Jan 26 17:35:30 crc kubenswrapper[4754]: I0126 17:35:30.904320 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tsfmz\" (UniqueName: \"kubernetes.io/projected/5fade260-b289-4c35-b77e-ca9f8ec754c3-kube-api-access-tsfmz\") pod \"5fade260-b289-4c35-b77e-ca9f8ec754c3\" (UID: \"5fade260-b289-4c35-b77e-ca9f8ec754c3\") " Jan 26 17:35:30 crc kubenswrapper[4754]: I0126 17:35:30.904487 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/5fade260-b289-4c35-b77e-ca9f8ec754c3-ceilometer-compute-config-data-1\") pod \"5fade260-b289-4c35-b77e-ca9f8ec754c3\" (UID: \"5fade260-b289-4c35-b77e-ca9f8ec754c3\") " Jan 26 17:35:30 crc kubenswrapper[4754]: I0126 17:35:30.904516 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5fade260-b289-4c35-b77e-ca9f8ec754c3-ssh-key-openstack-edpm-ipam\") pod \"5fade260-b289-4c35-b77e-ca9f8ec754c3\" (UID: \"5fade260-b289-4c35-b77e-ca9f8ec754c3\") " Jan 26 17:35:30 crc kubenswrapper[4754]: I0126 17:35:30.904603 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5fade260-b289-4c35-b77e-ca9f8ec754c3-telemetry-combined-ca-bundle\") pod \"5fade260-b289-4c35-b77e-ca9f8ec754c3\" (UID: \"5fade260-b289-4c35-b77e-ca9f8ec754c3\") " Jan 26 17:35:30 crc kubenswrapper[4754]: I0126 17:35:30.910461 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fade260-b289-4c35-b77e-ca9f8ec754c3-kube-api-access-tsfmz" (OuterVolumeSpecName: "kube-api-access-tsfmz") pod "5fade260-b289-4c35-b77e-ca9f8ec754c3" (UID: "5fade260-b289-4c35-b77e-ca9f8ec754c3"). InnerVolumeSpecName "kube-api-access-tsfmz". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:35:30 crc kubenswrapper[4754]: I0126 17:35:30.913848 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fade260-b289-4c35-b77e-ca9f8ec754c3-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "5fade260-b289-4c35-b77e-ca9f8ec754c3" (UID: "5fade260-b289-4c35-b77e-ca9f8ec754c3"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:35:30 crc kubenswrapper[4754]: I0126 17:35:30.940179 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fade260-b289-4c35-b77e-ca9f8ec754c3-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "5fade260-b289-4c35-b77e-ca9f8ec754c3" (UID: "5fade260-b289-4c35-b77e-ca9f8ec754c3"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:35:30 crc kubenswrapper[4754]: I0126 17:35:30.943134 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fade260-b289-4c35-b77e-ca9f8ec754c3-ceilometer-compute-config-data-1" (OuterVolumeSpecName: "ceilometer-compute-config-data-1") pod "5fade260-b289-4c35-b77e-ca9f8ec754c3" (UID: "5fade260-b289-4c35-b77e-ca9f8ec754c3"). InnerVolumeSpecName "ceilometer-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:35:30 crc kubenswrapper[4754]: I0126 17:35:30.947752 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fade260-b289-4c35-b77e-ca9f8ec754c3-ceilometer-compute-config-data-2" (OuterVolumeSpecName: "ceilometer-compute-config-data-2") pod "5fade260-b289-4c35-b77e-ca9f8ec754c3" (UID: "5fade260-b289-4c35-b77e-ca9f8ec754c3"). InnerVolumeSpecName "ceilometer-compute-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:35:30 crc kubenswrapper[4754]: I0126 17:35:30.947918 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fade260-b289-4c35-b77e-ca9f8ec754c3-ceilometer-compute-config-data-0" (OuterVolumeSpecName: "ceilometer-compute-config-data-0") pod "5fade260-b289-4c35-b77e-ca9f8ec754c3" (UID: "5fade260-b289-4c35-b77e-ca9f8ec754c3"). InnerVolumeSpecName "ceilometer-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:35:30 crc kubenswrapper[4754]: I0126 17:35:30.955746 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fade260-b289-4c35-b77e-ca9f8ec754c3-inventory" (OuterVolumeSpecName: "inventory") pod "5fade260-b289-4c35-b77e-ca9f8ec754c3" (UID: "5fade260-b289-4c35-b77e-ca9f8ec754c3"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:35:31 crc kubenswrapper[4754]: I0126 17:35:31.007276 4754 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/5fade260-b289-4c35-b77e-ca9f8ec754c3-ceilometer-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Jan 26 17:35:31 crc kubenswrapper[4754]: I0126 17:35:31.007308 4754 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5fade260-b289-4c35-b77e-ca9f8ec754c3-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Jan 26 17:35:31 crc kubenswrapper[4754]: I0126 17:35:31.007319 4754 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5fade260-b289-4c35-b77e-ca9f8ec754c3-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 26 17:35:31 crc kubenswrapper[4754]: I0126 17:35:31.007328 4754 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/5fade260-b289-4c35-b77e-ca9f8ec754c3-ceilometer-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Jan 26 17:35:31 crc kubenswrapper[4754]: I0126 17:35:31.007337 4754 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/5fade260-b289-4c35-b77e-ca9f8ec754c3-ceilometer-compute-config-data-2\") on node \"crc\" DevicePath \"\"" Jan 26 17:35:31 crc kubenswrapper[4754]: I0126 17:35:31.007347 4754 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5fade260-b289-4c35-b77e-ca9f8ec754c3-inventory\") on node \"crc\" DevicePath \"\"" Jan 26 17:35:31 crc kubenswrapper[4754]: I0126 17:35:31.007358 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tsfmz\" (UniqueName: \"kubernetes.io/projected/5fade260-b289-4c35-b77e-ca9f8ec754c3-kube-api-access-tsfmz\") on node \"crc\" DevicePath \"\"" Jan 26 17:35:31 crc kubenswrapper[4754]: I0126 17:35:31.360805 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-59vz4" event={"ID":"5fade260-b289-4c35-b77e-ca9f8ec754c3","Type":"ContainerDied","Data":"58b810266bda80a7a3f9af826f449ae0be7384d7256e04bb2ffa71d4441625f7"} Jan 26 17:35:31 crc kubenswrapper[4754]: I0126 17:35:31.361146 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="58b810266bda80a7a3f9af826f449ae0be7384d7256e04bb2ffa71d4441625f7" Jan 26 17:35:31 crc kubenswrapper[4754]: I0126 17:35:31.361119 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-59vz4" Jan 26 17:35:37 crc kubenswrapper[4754]: I0126 17:35:37.129344 4754 patch_prober.go:28] interesting pod/machine-config-daemon-x65wv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 26 17:35:37 crc kubenswrapper[4754]: I0126 17:35:37.129815 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 26 17:36:07 crc kubenswrapper[4754]: I0126 17:36:07.129526 4754 patch_prober.go:28] interesting pod/machine-config-daemon-x65wv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 26 17:36:07 crc kubenswrapper[4754]: I0126 17:36:07.131119 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 26 17:36:27 crc kubenswrapper[4754]: I0126 17:36:27.563033 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tempest-tests-tempest"] Jan 26 17:36:27 crc kubenswrapper[4754]: E0126 17:36:27.565952 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b54a1ca-9c37-4258-863b-10141af1389d" containerName="extract-content" Jan 26 17:36:27 crc kubenswrapper[4754]: I0126 17:36:27.566080 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b54a1ca-9c37-4258-863b-10141af1389d" containerName="extract-content" Jan 26 17:36:27 crc kubenswrapper[4754]: E0126 17:36:27.566169 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b54a1ca-9c37-4258-863b-10141af1389d" containerName="registry-server" Jan 26 17:36:27 crc kubenswrapper[4754]: I0126 17:36:27.566240 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b54a1ca-9c37-4258-863b-10141af1389d" containerName="registry-server" Jan 26 17:36:27 crc kubenswrapper[4754]: E0126 17:36:27.566352 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5fade260-b289-4c35-b77e-ca9f8ec754c3" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Jan 26 17:36:27 crc kubenswrapper[4754]: I0126 17:36:27.566427 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="5fade260-b289-4c35-b77e-ca9f8ec754c3" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Jan 26 17:36:27 crc kubenswrapper[4754]: E0126 17:36:27.566513 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b54a1ca-9c37-4258-863b-10141af1389d" containerName="extract-utilities" Jan 26 17:36:27 crc kubenswrapper[4754]: I0126 17:36:27.566589 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b54a1ca-9c37-4258-863b-10141af1389d" containerName="extract-utilities" Jan 26 17:36:27 crc kubenswrapper[4754]: I0126 17:36:27.566917 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="5b54a1ca-9c37-4258-863b-10141af1389d" containerName="registry-server" Jan 26 17:36:27 crc kubenswrapper[4754]: I0126 17:36:27.567013 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="5fade260-b289-4c35-b77e-ca9f8ec754c3" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Jan 26 17:36:27 crc kubenswrapper[4754]: I0126 17:36:27.568230 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Jan 26 17:36:27 crc kubenswrapper[4754]: I0126 17:36:27.571470 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-hm66p" Jan 26 17:36:27 crc kubenswrapper[4754]: I0126 17:36:27.571506 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"test-operator-controller-priv-key" Jan 26 17:36:27 crc kubenswrapper[4754]: I0126 17:36:27.571470 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-custom-data-s0" Jan 26 17:36:27 crc kubenswrapper[4754]: I0126 17:36:27.571497 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Jan 26 17:36:27 crc kubenswrapper[4754]: I0126 17:36:27.579132 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Jan 26 17:36:27 crc kubenswrapper[4754]: I0126 17:36:27.630961 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/6f413c96-a0cb-4d11-9def-df8c3218db98-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"6f413c96-a0cb-4d11-9def-df8c3218db98\") " pod="openstack/tempest-tests-tempest" Jan 26 17:36:27 crc kubenswrapper[4754]: I0126 17:36:27.631199 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/6f413c96-a0cb-4d11-9def-df8c3218db98-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"6f413c96-a0cb-4d11-9def-df8c3218db98\") " pod="openstack/tempest-tests-tempest" Jan 26 17:36:27 crc kubenswrapper[4754]: I0126 17:36:27.631535 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6f413c96-a0cb-4d11-9def-df8c3218db98-config-data\") pod \"tempest-tests-tempest\" (UID: \"6f413c96-a0cb-4d11-9def-df8c3218db98\") " pod="openstack/tempest-tests-tempest" Jan 26 17:36:27 crc kubenswrapper[4754]: I0126 17:36:27.733627 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/6f413c96-a0cb-4d11-9def-df8c3218db98-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"6f413c96-a0cb-4d11-9def-df8c3218db98\") " pod="openstack/tempest-tests-tempest" Jan 26 17:36:27 crc kubenswrapper[4754]: I0126 17:36:27.733834 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/6f413c96-a0cb-4d11-9def-df8c3218db98-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"6f413c96-a0cb-4d11-9def-df8c3218db98\") " pod="openstack/tempest-tests-tempest" Jan 26 17:36:27 crc kubenswrapper[4754]: I0126 17:36:27.733899 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6f413c96-a0cb-4d11-9def-df8c3218db98-config-data\") pod \"tempest-tests-tempest\" (UID: \"6f413c96-a0cb-4d11-9def-df8c3218db98\") " pod="openstack/tempest-tests-tempest" Jan 26 17:36:27 crc kubenswrapper[4754]: I0126 17:36:27.733971 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/6f413c96-a0cb-4d11-9def-df8c3218db98-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"6f413c96-a0cb-4d11-9def-df8c3218db98\") " pod="openstack/tempest-tests-tempest" Jan 26 17:36:27 crc kubenswrapper[4754]: I0126 17:36:27.734037 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"tempest-tests-tempest\" (UID: \"6f413c96-a0cb-4d11-9def-df8c3218db98\") " pod="openstack/tempest-tests-tempest" Jan 26 17:36:27 crc kubenswrapper[4754]: I0126 17:36:27.734138 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6f413c96-a0cb-4d11-9def-df8c3218db98-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"6f413c96-a0cb-4d11-9def-df8c3218db98\") " pod="openstack/tempest-tests-tempest" Jan 26 17:36:27 crc kubenswrapper[4754]: I0126 17:36:27.734167 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p72sf\" (UniqueName: \"kubernetes.io/projected/6f413c96-a0cb-4d11-9def-df8c3218db98-kube-api-access-p72sf\") pod \"tempest-tests-tempest\" (UID: \"6f413c96-a0cb-4d11-9def-df8c3218db98\") " pod="openstack/tempest-tests-tempest" Jan 26 17:36:27 crc kubenswrapper[4754]: I0126 17:36:27.734220 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/6f413c96-a0cb-4d11-9def-df8c3218db98-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"6f413c96-a0cb-4d11-9def-df8c3218db98\") " pod="openstack/tempest-tests-tempest" Jan 26 17:36:27 crc kubenswrapper[4754]: I0126 17:36:27.734261 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/6f413c96-a0cb-4d11-9def-df8c3218db98-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"6f413c96-a0cb-4d11-9def-df8c3218db98\") " pod="openstack/tempest-tests-tempest" Jan 26 17:36:27 crc kubenswrapper[4754]: I0126 17:36:27.736897 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/6f413c96-a0cb-4d11-9def-df8c3218db98-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"6f413c96-a0cb-4d11-9def-df8c3218db98\") " pod="openstack/tempest-tests-tempest" Jan 26 17:36:27 crc kubenswrapper[4754]: I0126 17:36:27.737361 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6f413c96-a0cb-4d11-9def-df8c3218db98-config-data\") pod \"tempest-tests-tempest\" (UID: \"6f413c96-a0cb-4d11-9def-df8c3218db98\") " pod="openstack/tempest-tests-tempest" Jan 26 17:36:27 crc kubenswrapper[4754]: I0126 17:36:27.751571 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/6f413c96-a0cb-4d11-9def-df8c3218db98-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"6f413c96-a0cb-4d11-9def-df8c3218db98\") " pod="openstack/tempest-tests-tempest" Jan 26 17:36:27 crc kubenswrapper[4754]: I0126 17:36:27.836202 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6f413c96-a0cb-4d11-9def-df8c3218db98-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"6f413c96-a0cb-4d11-9def-df8c3218db98\") " pod="openstack/tempest-tests-tempest" Jan 26 17:36:27 crc kubenswrapper[4754]: I0126 17:36:27.836296 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p72sf\" (UniqueName: \"kubernetes.io/projected/6f413c96-a0cb-4d11-9def-df8c3218db98-kube-api-access-p72sf\") pod \"tempest-tests-tempest\" (UID: \"6f413c96-a0cb-4d11-9def-df8c3218db98\") " pod="openstack/tempest-tests-tempest" Jan 26 17:36:27 crc kubenswrapper[4754]: I0126 17:36:27.836377 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/6f413c96-a0cb-4d11-9def-df8c3218db98-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"6f413c96-a0cb-4d11-9def-df8c3218db98\") " pod="openstack/tempest-tests-tempest" Jan 26 17:36:27 crc kubenswrapper[4754]: I0126 17:36:27.836543 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/6f413c96-a0cb-4d11-9def-df8c3218db98-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"6f413c96-a0cb-4d11-9def-df8c3218db98\") " pod="openstack/tempest-tests-tempest" Jan 26 17:36:27 crc kubenswrapper[4754]: I0126 17:36:27.836638 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/6f413c96-a0cb-4d11-9def-df8c3218db98-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"6f413c96-a0cb-4d11-9def-df8c3218db98\") " pod="openstack/tempest-tests-tempest" Jan 26 17:36:27 crc kubenswrapper[4754]: I0126 17:36:27.836811 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"tempest-tests-tempest\" (UID: \"6f413c96-a0cb-4d11-9def-df8c3218db98\") " pod="openstack/tempest-tests-tempest" Jan 26 17:36:27 crc kubenswrapper[4754]: I0126 17:36:27.837250 4754 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"tempest-tests-tempest\" (UID: \"6f413c96-a0cb-4d11-9def-df8c3218db98\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/tempest-tests-tempest" Jan 26 17:36:27 crc kubenswrapper[4754]: I0126 17:36:27.837290 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/6f413c96-a0cb-4d11-9def-df8c3218db98-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"6f413c96-a0cb-4d11-9def-df8c3218db98\") " pod="openstack/tempest-tests-tempest" Jan 26 17:36:27 crc kubenswrapper[4754]: I0126 17:36:27.837705 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/6f413c96-a0cb-4d11-9def-df8c3218db98-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"6f413c96-a0cb-4d11-9def-df8c3218db98\") " pod="openstack/tempest-tests-tempest" Jan 26 17:36:27 crc kubenswrapper[4754]: I0126 17:36:27.841697 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6f413c96-a0cb-4d11-9def-df8c3218db98-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"6f413c96-a0cb-4d11-9def-df8c3218db98\") " pod="openstack/tempest-tests-tempest" Jan 26 17:36:27 crc kubenswrapper[4754]: I0126 17:36:27.844175 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/6f413c96-a0cb-4d11-9def-df8c3218db98-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"6f413c96-a0cb-4d11-9def-df8c3218db98\") " pod="openstack/tempest-tests-tempest" Jan 26 17:36:27 crc kubenswrapper[4754]: I0126 17:36:27.863193 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p72sf\" (UniqueName: \"kubernetes.io/projected/6f413c96-a0cb-4d11-9def-df8c3218db98-kube-api-access-p72sf\") pod \"tempest-tests-tempest\" (UID: \"6f413c96-a0cb-4d11-9def-df8c3218db98\") " pod="openstack/tempest-tests-tempest" Jan 26 17:36:27 crc kubenswrapper[4754]: I0126 17:36:27.885074 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"tempest-tests-tempest\" (UID: \"6f413c96-a0cb-4d11-9def-df8c3218db98\") " pod="openstack/tempest-tests-tempest" Jan 26 17:36:27 crc kubenswrapper[4754]: I0126 17:36:27.892739 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Jan 26 17:36:28 crc kubenswrapper[4754]: I0126 17:36:28.348760 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Jan 26 17:36:28 crc kubenswrapper[4754]: I0126 17:36:28.354916 4754 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 26 17:36:28 crc kubenswrapper[4754]: I0126 17:36:28.911075 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"6f413c96-a0cb-4d11-9def-df8c3218db98","Type":"ContainerStarted","Data":"d60a26d5912f2c0f9acdb318d8eec23b93eea368db920c2db06859102ea6d69a"} Jan 26 17:36:37 crc kubenswrapper[4754]: I0126 17:36:37.129081 4754 patch_prober.go:28] interesting pod/machine-config-daemon-x65wv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 26 17:36:37 crc kubenswrapper[4754]: I0126 17:36:37.129602 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 26 17:36:37 crc kubenswrapper[4754]: I0126 17:36:37.130408 4754 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" Jan 26 17:36:37 crc kubenswrapper[4754]: I0126 17:36:37.131449 4754 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"2a52617105aca80bd6aa782680de53fad558b6014f1d373ef1592509df2dabc0"} pod="openshift-machine-config-operator/machine-config-daemon-x65wv" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 26 17:36:37 crc kubenswrapper[4754]: I0126 17:36:37.131537 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" containerName="machine-config-daemon" containerID="cri-o://2a52617105aca80bd6aa782680de53fad558b6014f1d373ef1592509df2dabc0" gracePeriod=600 Jan 26 17:36:37 crc kubenswrapper[4754]: E0126 17:36:37.255988 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x65wv_openshift-machine-config-operator(8c3718a4-f354-4284-92e0-fdfb45a692bd)\"" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" Jan 26 17:36:37 crc kubenswrapper[4754]: I0126 17:36:37.993321 4754 generic.go:334] "Generic (PLEG): container finished" podID="8c3718a4-f354-4284-92e0-fdfb45a692bd" containerID="2a52617105aca80bd6aa782680de53fad558b6014f1d373ef1592509df2dabc0" exitCode=0 Jan 26 17:36:37 crc kubenswrapper[4754]: I0126 17:36:37.993378 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" event={"ID":"8c3718a4-f354-4284-92e0-fdfb45a692bd","Type":"ContainerDied","Data":"2a52617105aca80bd6aa782680de53fad558b6014f1d373ef1592509df2dabc0"} Jan 26 17:36:37 crc kubenswrapper[4754]: I0126 17:36:37.993426 4754 scope.go:117] "RemoveContainer" containerID="efeb34d7b9f84cbdd86409079c972f84a47710b40532d5bb54abdaa0f23b0f5d" Jan 26 17:36:37 crc kubenswrapper[4754]: I0126 17:36:37.994066 4754 scope.go:117] "RemoveContainer" containerID="2a52617105aca80bd6aa782680de53fad558b6014f1d373ef1592509df2dabc0" Jan 26 17:36:37 crc kubenswrapper[4754]: E0126 17:36:37.994406 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x65wv_openshift-machine-config-operator(8c3718a4-f354-4284-92e0-fdfb45a692bd)\"" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" Jan 26 17:36:48 crc kubenswrapper[4754]: I0126 17:36:48.767744 4754 scope.go:117] "RemoveContainer" containerID="2a52617105aca80bd6aa782680de53fad558b6014f1d373ef1592509df2dabc0" Jan 26 17:36:48 crc kubenswrapper[4754]: E0126 17:36:48.768715 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x65wv_openshift-machine-config-operator(8c3718a4-f354-4284-92e0-fdfb45a692bd)\"" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" Jan 26 17:37:01 crc kubenswrapper[4754]: I0126 17:37:01.770006 4754 scope.go:117] "RemoveContainer" containerID="2a52617105aca80bd6aa782680de53fad558b6014f1d373ef1592509df2dabc0" Jan 26 17:37:01 crc kubenswrapper[4754]: E0126 17:37:01.770811 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x65wv_openshift-machine-config-operator(8c3718a4-f354-4284-92e0-fdfb45a692bd)\"" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" Jan 26 17:37:03 crc kubenswrapper[4754]: E0126 17:37:03.342457 4754 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified" Jan 26 17:37:03 crc kubenswrapper[4754]: E0126 17:37:03.342735 4754 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:tempest-tests-tempest-tests-runner,Image:quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/test_operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-workdir,ReadOnly:false,MountPath:/var/lib/tempest,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-temporary,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-logs,ReadOnly:false,MountPath:/var/lib/tempest/external_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/etc/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/var/lib/tempest/.config/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/etc/openstack/secure.yaml,SubPath:secure.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ca-certs,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ssh-key,ReadOnly:false,MountPath:/var/lib/tempest/id_ecdsa,SubPath:ssh_key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-p72sf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42480,RunAsNonRoot:*false,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:*true,RunAsGroup:*42480,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-custom-data-s0,},Optional:nil,},SecretRef:nil,},EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-env-vars-s0,},Optional:nil,},SecretRef:nil,},},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod tempest-tests-tempest_openstack(6f413c96-a0cb-4d11-9def-df8c3218db98): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Jan 26 17:37:03 crc kubenswrapper[4754]: E0126 17:37:03.344048 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/tempest-tests-tempest" podUID="6f413c96-a0cb-4d11-9def-df8c3218db98" Jan 26 17:37:04 crc kubenswrapper[4754]: E0126 17:37:04.280141 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified\\\"\"" pod="openstack/tempest-tests-tempest" podUID="6f413c96-a0cb-4d11-9def-df8c3218db98" Jan 26 17:37:16 crc kubenswrapper[4754]: I0126 17:37:16.506957 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Jan 26 17:37:16 crc kubenswrapper[4754]: I0126 17:37:16.768966 4754 scope.go:117] "RemoveContainer" containerID="2a52617105aca80bd6aa782680de53fad558b6014f1d373ef1592509df2dabc0" Jan 26 17:37:16 crc kubenswrapper[4754]: E0126 17:37:16.769292 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x65wv_openshift-machine-config-operator(8c3718a4-f354-4284-92e0-fdfb45a692bd)\"" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" Jan 26 17:37:18 crc kubenswrapper[4754]: I0126 17:37:18.417735 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"6f413c96-a0cb-4d11-9def-df8c3218db98","Type":"ContainerStarted","Data":"3426d5689db28b717decc25263f3ffdc456630c56b24555f60c60ddab335bf44"} Jan 26 17:37:18 crc kubenswrapper[4754]: I0126 17:37:18.440250 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tempest-tests-tempest" podStartSLOduration=4.289573701 podStartE2EDuration="52.440229516s" podCreationTimestamp="2026-01-26 17:36:26 +0000 UTC" firstStartedPulling="2026-01-26 17:36:28.354332601 +0000 UTC m=+2954.878513045" lastFinishedPulling="2026-01-26 17:37:16.504988426 +0000 UTC m=+3003.029168860" observedRunningTime="2026-01-26 17:37:18.432800652 +0000 UTC m=+3004.956981116" watchObservedRunningTime="2026-01-26 17:37:18.440229516 +0000 UTC m=+3004.964409950" Jan 26 17:37:29 crc kubenswrapper[4754]: I0126 17:37:29.766947 4754 scope.go:117] "RemoveContainer" containerID="2a52617105aca80bd6aa782680de53fad558b6014f1d373ef1592509df2dabc0" Jan 26 17:37:29 crc kubenswrapper[4754]: E0126 17:37:29.769212 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x65wv_openshift-machine-config-operator(8c3718a4-f354-4284-92e0-fdfb45a692bd)\"" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" Jan 26 17:37:44 crc kubenswrapper[4754]: I0126 17:37:44.769810 4754 scope.go:117] "RemoveContainer" containerID="2a52617105aca80bd6aa782680de53fad558b6014f1d373ef1592509df2dabc0" Jan 26 17:37:44 crc kubenswrapper[4754]: E0126 17:37:44.770946 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x65wv_openshift-machine-config-operator(8c3718a4-f354-4284-92e0-fdfb45a692bd)\"" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" Jan 26 17:37:57 crc kubenswrapper[4754]: I0126 17:37:57.768380 4754 scope.go:117] "RemoveContainer" containerID="2a52617105aca80bd6aa782680de53fad558b6014f1d373ef1592509df2dabc0" Jan 26 17:37:57 crc kubenswrapper[4754]: E0126 17:37:57.769332 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x65wv_openshift-machine-config-operator(8c3718a4-f354-4284-92e0-fdfb45a692bd)\"" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" Jan 26 17:38:08 crc kubenswrapper[4754]: I0126 17:38:08.767923 4754 scope.go:117] "RemoveContainer" containerID="2a52617105aca80bd6aa782680de53fad558b6014f1d373ef1592509df2dabc0" Jan 26 17:38:08 crc kubenswrapper[4754]: E0126 17:38:08.768866 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x65wv_openshift-machine-config-operator(8c3718a4-f354-4284-92e0-fdfb45a692bd)\"" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" Jan 26 17:38:21 crc kubenswrapper[4754]: I0126 17:38:21.767784 4754 scope.go:117] "RemoveContainer" containerID="2a52617105aca80bd6aa782680de53fad558b6014f1d373ef1592509df2dabc0" Jan 26 17:38:21 crc kubenswrapper[4754]: E0126 17:38:21.768621 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x65wv_openshift-machine-config-operator(8c3718a4-f354-4284-92e0-fdfb45a692bd)\"" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" Jan 26 17:38:32 crc kubenswrapper[4754]: I0126 17:38:32.767913 4754 scope.go:117] "RemoveContainer" containerID="2a52617105aca80bd6aa782680de53fad558b6014f1d373ef1592509df2dabc0" Jan 26 17:38:32 crc kubenswrapper[4754]: E0126 17:38:32.769176 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x65wv_openshift-machine-config-operator(8c3718a4-f354-4284-92e0-fdfb45a692bd)\"" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" Jan 26 17:38:45 crc kubenswrapper[4754]: I0126 17:38:45.871295 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-b8dnn"] Jan 26 17:38:45 crc kubenswrapper[4754]: I0126 17:38:45.873863 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-b8dnn" Jan 26 17:38:45 crc kubenswrapper[4754]: I0126 17:38:45.885200 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-b8dnn"] Jan 26 17:38:45 crc kubenswrapper[4754]: I0126 17:38:45.962103 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b1708bb9-e67e-4bab-96ea-cc492ea78d2b-catalog-content\") pod \"redhat-marketplace-b8dnn\" (UID: \"b1708bb9-e67e-4bab-96ea-cc492ea78d2b\") " pod="openshift-marketplace/redhat-marketplace-b8dnn" Jan 26 17:38:45 crc kubenswrapper[4754]: I0126 17:38:45.962156 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b1708bb9-e67e-4bab-96ea-cc492ea78d2b-utilities\") pod \"redhat-marketplace-b8dnn\" (UID: \"b1708bb9-e67e-4bab-96ea-cc492ea78d2b\") " pod="openshift-marketplace/redhat-marketplace-b8dnn" Jan 26 17:38:45 crc kubenswrapper[4754]: I0126 17:38:45.962245 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cxr8l\" (UniqueName: \"kubernetes.io/projected/b1708bb9-e67e-4bab-96ea-cc492ea78d2b-kube-api-access-cxr8l\") pod \"redhat-marketplace-b8dnn\" (UID: \"b1708bb9-e67e-4bab-96ea-cc492ea78d2b\") " pod="openshift-marketplace/redhat-marketplace-b8dnn" Jan 26 17:38:46 crc kubenswrapper[4754]: I0126 17:38:46.063679 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cxr8l\" (UniqueName: \"kubernetes.io/projected/b1708bb9-e67e-4bab-96ea-cc492ea78d2b-kube-api-access-cxr8l\") pod \"redhat-marketplace-b8dnn\" (UID: \"b1708bb9-e67e-4bab-96ea-cc492ea78d2b\") " pod="openshift-marketplace/redhat-marketplace-b8dnn" Jan 26 17:38:46 crc kubenswrapper[4754]: I0126 17:38:46.063770 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b1708bb9-e67e-4bab-96ea-cc492ea78d2b-catalog-content\") pod \"redhat-marketplace-b8dnn\" (UID: \"b1708bb9-e67e-4bab-96ea-cc492ea78d2b\") " pod="openshift-marketplace/redhat-marketplace-b8dnn" Jan 26 17:38:46 crc kubenswrapper[4754]: I0126 17:38:46.063802 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b1708bb9-e67e-4bab-96ea-cc492ea78d2b-utilities\") pod \"redhat-marketplace-b8dnn\" (UID: \"b1708bb9-e67e-4bab-96ea-cc492ea78d2b\") " pod="openshift-marketplace/redhat-marketplace-b8dnn" Jan 26 17:38:46 crc kubenswrapper[4754]: I0126 17:38:46.064264 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b1708bb9-e67e-4bab-96ea-cc492ea78d2b-utilities\") pod \"redhat-marketplace-b8dnn\" (UID: \"b1708bb9-e67e-4bab-96ea-cc492ea78d2b\") " pod="openshift-marketplace/redhat-marketplace-b8dnn" Jan 26 17:38:46 crc kubenswrapper[4754]: I0126 17:38:46.064822 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b1708bb9-e67e-4bab-96ea-cc492ea78d2b-catalog-content\") pod \"redhat-marketplace-b8dnn\" (UID: \"b1708bb9-e67e-4bab-96ea-cc492ea78d2b\") " pod="openshift-marketplace/redhat-marketplace-b8dnn" Jan 26 17:38:46 crc kubenswrapper[4754]: I0126 17:38:46.084168 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cxr8l\" (UniqueName: \"kubernetes.io/projected/b1708bb9-e67e-4bab-96ea-cc492ea78d2b-kube-api-access-cxr8l\") pod \"redhat-marketplace-b8dnn\" (UID: \"b1708bb9-e67e-4bab-96ea-cc492ea78d2b\") " pod="openshift-marketplace/redhat-marketplace-b8dnn" Jan 26 17:38:46 crc kubenswrapper[4754]: I0126 17:38:46.206040 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-b8dnn" Jan 26 17:38:46 crc kubenswrapper[4754]: I0126 17:38:46.709190 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-b8dnn"] Jan 26 17:38:47 crc kubenswrapper[4754]: I0126 17:38:47.179019 4754 generic.go:334] "Generic (PLEG): container finished" podID="b1708bb9-e67e-4bab-96ea-cc492ea78d2b" containerID="f7650431c265a3adf1c98bf8286f63f7460a76e6479a5ee98f870682cd958150" exitCode=0 Jan 26 17:38:47 crc kubenswrapper[4754]: I0126 17:38:47.179091 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b8dnn" event={"ID":"b1708bb9-e67e-4bab-96ea-cc492ea78d2b","Type":"ContainerDied","Data":"f7650431c265a3adf1c98bf8286f63f7460a76e6479a5ee98f870682cd958150"} Jan 26 17:38:47 crc kubenswrapper[4754]: I0126 17:38:47.179324 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b8dnn" event={"ID":"b1708bb9-e67e-4bab-96ea-cc492ea78d2b","Type":"ContainerStarted","Data":"fc23d2490e660ca2ae27a91109fb1d53792bacf5bbe1d747b5f0f563e99105f2"} Jan 26 17:38:47 crc kubenswrapper[4754]: I0126 17:38:47.768199 4754 scope.go:117] "RemoveContainer" containerID="2a52617105aca80bd6aa782680de53fad558b6014f1d373ef1592509df2dabc0" Jan 26 17:38:47 crc kubenswrapper[4754]: E0126 17:38:47.768426 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x65wv_openshift-machine-config-operator(8c3718a4-f354-4284-92e0-fdfb45a692bd)\"" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" Jan 26 17:38:48 crc kubenswrapper[4754]: I0126 17:38:48.189157 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b8dnn" event={"ID":"b1708bb9-e67e-4bab-96ea-cc492ea78d2b","Type":"ContainerStarted","Data":"33dd0687874e29c001be1eabc23ee486dc0b1d4af11767d9552c2b2de18b7c3c"} Jan 26 17:38:49 crc kubenswrapper[4754]: I0126 17:38:49.209161 4754 generic.go:334] "Generic (PLEG): container finished" podID="b1708bb9-e67e-4bab-96ea-cc492ea78d2b" containerID="33dd0687874e29c001be1eabc23ee486dc0b1d4af11767d9552c2b2de18b7c3c" exitCode=0 Jan 26 17:38:49 crc kubenswrapper[4754]: I0126 17:38:49.209215 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b8dnn" event={"ID":"b1708bb9-e67e-4bab-96ea-cc492ea78d2b","Type":"ContainerDied","Data":"33dd0687874e29c001be1eabc23ee486dc0b1d4af11767d9552c2b2de18b7c3c"} Jan 26 17:38:50 crc kubenswrapper[4754]: I0126 17:38:50.218512 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b8dnn" event={"ID":"b1708bb9-e67e-4bab-96ea-cc492ea78d2b","Type":"ContainerStarted","Data":"801de20aa212884bf68be881676ae7a2868c6698581556da2c200e19ed697cc0"} Jan 26 17:38:50 crc kubenswrapper[4754]: I0126 17:38:50.248438 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-b8dnn" podStartSLOduration=2.601808835 podStartE2EDuration="5.248413574s" podCreationTimestamp="2026-01-26 17:38:45 +0000 UTC" firstStartedPulling="2026-01-26 17:38:47.180720471 +0000 UTC m=+3093.704900905" lastFinishedPulling="2026-01-26 17:38:49.82732521 +0000 UTC m=+3096.351505644" observedRunningTime="2026-01-26 17:38:50.241844561 +0000 UTC m=+3096.766025005" watchObservedRunningTime="2026-01-26 17:38:50.248413574 +0000 UTC m=+3096.772594028" Jan 26 17:38:56 crc kubenswrapper[4754]: I0126 17:38:56.207501 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-b8dnn" Jan 26 17:38:56 crc kubenswrapper[4754]: I0126 17:38:56.208218 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-b8dnn" Jan 26 17:38:56 crc kubenswrapper[4754]: I0126 17:38:56.257292 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-b8dnn" Jan 26 17:38:56 crc kubenswrapper[4754]: I0126 17:38:56.324163 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-b8dnn" Jan 26 17:38:56 crc kubenswrapper[4754]: I0126 17:38:56.491518 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-b8dnn"] Jan 26 17:38:58 crc kubenswrapper[4754]: I0126 17:38:58.293357 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-b8dnn" podUID="b1708bb9-e67e-4bab-96ea-cc492ea78d2b" containerName="registry-server" containerID="cri-o://801de20aa212884bf68be881676ae7a2868c6698581556da2c200e19ed697cc0" gracePeriod=2 Jan 26 17:38:58 crc kubenswrapper[4754]: I0126 17:38:58.795772 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-b8dnn" Jan 26 17:38:58 crc kubenswrapper[4754]: I0126 17:38:58.907015 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cxr8l\" (UniqueName: \"kubernetes.io/projected/b1708bb9-e67e-4bab-96ea-cc492ea78d2b-kube-api-access-cxr8l\") pod \"b1708bb9-e67e-4bab-96ea-cc492ea78d2b\" (UID: \"b1708bb9-e67e-4bab-96ea-cc492ea78d2b\") " Jan 26 17:38:58 crc kubenswrapper[4754]: I0126 17:38:58.907124 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b1708bb9-e67e-4bab-96ea-cc492ea78d2b-catalog-content\") pod \"b1708bb9-e67e-4bab-96ea-cc492ea78d2b\" (UID: \"b1708bb9-e67e-4bab-96ea-cc492ea78d2b\") " Jan 26 17:38:58 crc kubenswrapper[4754]: I0126 17:38:58.907565 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b1708bb9-e67e-4bab-96ea-cc492ea78d2b-utilities\") pod \"b1708bb9-e67e-4bab-96ea-cc492ea78d2b\" (UID: \"b1708bb9-e67e-4bab-96ea-cc492ea78d2b\") " Jan 26 17:38:58 crc kubenswrapper[4754]: I0126 17:38:58.908286 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b1708bb9-e67e-4bab-96ea-cc492ea78d2b-utilities" (OuterVolumeSpecName: "utilities") pod "b1708bb9-e67e-4bab-96ea-cc492ea78d2b" (UID: "b1708bb9-e67e-4bab-96ea-cc492ea78d2b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:38:58 crc kubenswrapper[4754]: I0126 17:38:58.916256 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b1708bb9-e67e-4bab-96ea-cc492ea78d2b-kube-api-access-cxr8l" (OuterVolumeSpecName: "kube-api-access-cxr8l") pod "b1708bb9-e67e-4bab-96ea-cc492ea78d2b" (UID: "b1708bb9-e67e-4bab-96ea-cc492ea78d2b"). InnerVolumeSpecName "kube-api-access-cxr8l". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:38:58 crc kubenswrapper[4754]: I0126 17:38:58.940644 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b1708bb9-e67e-4bab-96ea-cc492ea78d2b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b1708bb9-e67e-4bab-96ea-cc492ea78d2b" (UID: "b1708bb9-e67e-4bab-96ea-cc492ea78d2b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:38:59 crc kubenswrapper[4754]: I0126 17:38:59.009844 4754 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b1708bb9-e67e-4bab-96ea-cc492ea78d2b-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 26 17:38:59 crc kubenswrapper[4754]: I0126 17:38:59.009880 4754 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b1708bb9-e67e-4bab-96ea-cc492ea78d2b-utilities\") on node \"crc\" DevicePath \"\"" Jan 26 17:38:59 crc kubenswrapper[4754]: I0126 17:38:59.009890 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cxr8l\" (UniqueName: \"kubernetes.io/projected/b1708bb9-e67e-4bab-96ea-cc492ea78d2b-kube-api-access-cxr8l\") on node \"crc\" DevicePath \"\"" Jan 26 17:38:59 crc kubenswrapper[4754]: I0126 17:38:59.307245 4754 generic.go:334] "Generic (PLEG): container finished" podID="b1708bb9-e67e-4bab-96ea-cc492ea78d2b" containerID="801de20aa212884bf68be881676ae7a2868c6698581556da2c200e19ed697cc0" exitCode=0 Jan 26 17:38:59 crc kubenswrapper[4754]: I0126 17:38:59.307391 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-b8dnn" Jan 26 17:38:59 crc kubenswrapper[4754]: I0126 17:38:59.307414 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b8dnn" event={"ID":"b1708bb9-e67e-4bab-96ea-cc492ea78d2b","Type":"ContainerDied","Data":"801de20aa212884bf68be881676ae7a2868c6698581556da2c200e19ed697cc0"} Jan 26 17:38:59 crc kubenswrapper[4754]: I0126 17:38:59.307903 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b8dnn" event={"ID":"b1708bb9-e67e-4bab-96ea-cc492ea78d2b","Type":"ContainerDied","Data":"fc23d2490e660ca2ae27a91109fb1d53792bacf5bbe1d747b5f0f563e99105f2"} Jan 26 17:38:59 crc kubenswrapper[4754]: I0126 17:38:59.307934 4754 scope.go:117] "RemoveContainer" containerID="801de20aa212884bf68be881676ae7a2868c6698581556da2c200e19ed697cc0" Jan 26 17:38:59 crc kubenswrapper[4754]: I0126 17:38:59.334418 4754 scope.go:117] "RemoveContainer" containerID="33dd0687874e29c001be1eabc23ee486dc0b1d4af11767d9552c2b2de18b7c3c" Jan 26 17:38:59 crc kubenswrapper[4754]: I0126 17:38:59.352062 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-b8dnn"] Jan 26 17:38:59 crc kubenswrapper[4754]: I0126 17:38:59.363423 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-b8dnn"] Jan 26 17:38:59 crc kubenswrapper[4754]: I0126 17:38:59.386960 4754 scope.go:117] "RemoveContainer" containerID="f7650431c265a3adf1c98bf8286f63f7460a76e6479a5ee98f870682cd958150" Jan 26 17:38:59 crc kubenswrapper[4754]: I0126 17:38:59.407910 4754 scope.go:117] "RemoveContainer" containerID="801de20aa212884bf68be881676ae7a2868c6698581556da2c200e19ed697cc0" Jan 26 17:38:59 crc kubenswrapper[4754]: E0126 17:38:59.408594 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"801de20aa212884bf68be881676ae7a2868c6698581556da2c200e19ed697cc0\": container with ID starting with 801de20aa212884bf68be881676ae7a2868c6698581556da2c200e19ed697cc0 not found: ID does not exist" containerID="801de20aa212884bf68be881676ae7a2868c6698581556da2c200e19ed697cc0" Jan 26 17:38:59 crc kubenswrapper[4754]: I0126 17:38:59.408636 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"801de20aa212884bf68be881676ae7a2868c6698581556da2c200e19ed697cc0"} err="failed to get container status \"801de20aa212884bf68be881676ae7a2868c6698581556da2c200e19ed697cc0\": rpc error: code = NotFound desc = could not find container \"801de20aa212884bf68be881676ae7a2868c6698581556da2c200e19ed697cc0\": container with ID starting with 801de20aa212884bf68be881676ae7a2868c6698581556da2c200e19ed697cc0 not found: ID does not exist" Jan 26 17:38:59 crc kubenswrapper[4754]: I0126 17:38:59.408862 4754 scope.go:117] "RemoveContainer" containerID="33dd0687874e29c001be1eabc23ee486dc0b1d4af11767d9552c2b2de18b7c3c" Jan 26 17:38:59 crc kubenswrapper[4754]: E0126 17:38:59.409218 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"33dd0687874e29c001be1eabc23ee486dc0b1d4af11767d9552c2b2de18b7c3c\": container with ID starting with 33dd0687874e29c001be1eabc23ee486dc0b1d4af11767d9552c2b2de18b7c3c not found: ID does not exist" containerID="33dd0687874e29c001be1eabc23ee486dc0b1d4af11767d9552c2b2de18b7c3c" Jan 26 17:38:59 crc kubenswrapper[4754]: I0126 17:38:59.409248 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"33dd0687874e29c001be1eabc23ee486dc0b1d4af11767d9552c2b2de18b7c3c"} err="failed to get container status \"33dd0687874e29c001be1eabc23ee486dc0b1d4af11767d9552c2b2de18b7c3c\": rpc error: code = NotFound desc = could not find container \"33dd0687874e29c001be1eabc23ee486dc0b1d4af11767d9552c2b2de18b7c3c\": container with ID starting with 33dd0687874e29c001be1eabc23ee486dc0b1d4af11767d9552c2b2de18b7c3c not found: ID does not exist" Jan 26 17:38:59 crc kubenswrapper[4754]: I0126 17:38:59.409262 4754 scope.go:117] "RemoveContainer" containerID="f7650431c265a3adf1c98bf8286f63f7460a76e6479a5ee98f870682cd958150" Jan 26 17:38:59 crc kubenswrapper[4754]: E0126 17:38:59.409468 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f7650431c265a3adf1c98bf8286f63f7460a76e6479a5ee98f870682cd958150\": container with ID starting with f7650431c265a3adf1c98bf8286f63f7460a76e6479a5ee98f870682cd958150 not found: ID does not exist" containerID="f7650431c265a3adf1c98bf8286f63f7460a76e6479a5ee98f870682cd958150" Jan 26 17:38:59 crc kubenswrapper[4754]: I0126 17:38:59.409496 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f7650431c265a3adf1c98bf8286f63f7460a76e6479a5ee98f870682cd958150"} err="failed to get container status \"f7650431c265a3adf1c98bf8286f63f7460a76e6479a5ee98f870682cd958150\": rpc error: code = NotFound desc = could not find container \"f7650431c265a3adf1c98bf8286f63f7460a76e6479a5ee98f870682cd958150\": container with ID starting with f7650431c265a3adf1c98bf8286f63f7460a76e6479a5ee98f870682cd958150 not found: ID does not exist" Jan 26 17:38:59 crc kubenswrapper[4754]: I0126 17:38:59.779709 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b1708bb9-e67e-4bab-96ea-cc492ea78d2b" path="/var/lib/kubelet/pods/b1708bb9-e67e-4bab-96ea-cc492ea78d2b/volumes" Jan 26 17:39:01 crc kubenswrapper[4754]: I0126 17:39:01.767657 4754 scope.go:117] "RemoveContainer" containerID="2a52617105aca80bd6aa782680de53fad558b6014f1d373ef1592509df2dabc0" Jan 26 17:39:01 crc kubenswrapper[4754]: E0126 17:39:01.768193 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x65wv_openshift-machine-config-operator(8c3718a4-f354-4284-92e0-fdfb45a692bd)\"" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" Jan 26 17:39:12 crc kubenswrapper[4754]: I0126 17:39:12.767981 4754 scope.go:117] "RemoveContainer" containerID="2a52617105aca80bd6aa782680de53fad558b6014f1d373ef1592509df2dabc0" Jan 26 17:39:12 crc kubenswrapper[4754]: E0126 17:39:12.768728 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x65wv_openshift-machine-config-operator(8c3718a4-f354-4284-92e0-fdfb45a692bd)\"" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" Jan 26 17:39:26 crc kubenswrapper[4754]: I0126 17:39:26.767303 4754 scope.go:117] "RemoveContainer" containerID="2a52617105aca80bd6aa782680de53fad558b6014f1d373ef1592509df2dabc0" Jan 26 17:39:26 crc kubenswrapper[4754]: E0126 17:39:26.768160 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x65wv_openshift-machine-config-operator(8c3718a4-f354-4284-92e0-fdfb45a692bd)\"" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" Jan 26 17:39:37 crc kubenswrapper[4754]: I0126 17:39:37.767239 4754 scope.go:117] "RemoveContainer" containerID="2a52617105aca80bd6aa782680de53fad558b6014f1d373ef1592509df2dabc0" Jan 26 17:39:37 crc kubenswrapper[4754]: E0126 17:39:37.767931 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x65wv_openshift-machine-config-operator(8c3718a4-f354-4284-92e0-fdfb45a692bd)\"" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" Jan 26 17:39:48 crc kubenswrapper[4754]: I0126 17:39:48.767485 4754 scope.go:117] "RemoveContainer" containerID="2a52617105aca80bd6aa782680de53fad558b6014f1d373ef1592509df2dabc0" Jan 26 17:39:48 crc kubenswrapper[4754]: E0126 17:39:48.768461 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x65wv_openshift-machine-config-operator(8c3718a4-f354-4284-92e0-fdfb45a692bd)\"" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" Jan 26 17:40:00 crc kubenswrapper[4754]: I0126 17:40:00.767886 4754 scope.go:117] "RemoveContainer" containerID="2a52617105aca80bd6aa782680de53fad558b6014f1d373ef1592509df2dabc0" Jan 26 17:40:00 crc kubenswrapper[4754]: E0126 17:40:00.768744 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x65wv_openshift-machine-config-operator(8c3718a4-f354-4284-92e0-fdfb45a692bd)\"" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" Jan 26 17:40:12 crc kubenswrapper[4754]: I0126 17:40:12.767534 4754 scope.go:117] "RemoveContainer" containerID="2a52617105aca80bd6aa782680de53fad558b6014f1d373ef1592509df2dabc0" Jan 26 17:40:12 crc kubenswrapper[4754]: E0126 17:40:12.768322 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x65wv_openshift-machine-config-operator(8c3718a4-f354-4284-92e0-fdfb45a692bd)\"" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" Jan 26 17:40:25 crc kubenswrapper[4754]: I0126 17:40:25.767940 4754 scope.go:117] "RemoveContainer" containerID="2a52617105aca80bd6aa782680de53fad558b6014f1d373ef1592509df2dabc0" Jan 26 17:40:25 crc kubenswrapper[4754]: E0126 17:40:25.768875 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x65wv_openshift-machine-config-operator(8c3718a4-f354-4284-92e0-fdfb45a692bd)\"" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" Jan 26 17:40:40 crc kubenswrapper[4754]: I0126 17:40:40.767844 4754 scope.go:117] "RemoveContainer" containerID="2a52617105aca80bd6aa782680de53fad558b6014f1d373ef1592509df2dabc0" Jan 26 17:40:40 crc kubenswrapper[4754]: E0126 17:40:40.768863 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x65wv_openshift-machine-config-operator(8c3718a4-f354-4284-92e0-fdfb45a692bd)\"" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" Jan 26 17:40:55 crc kubenswrapper[4754]: I0126 17:40:55.767900 4754 scope.go:117] "RemoveContainer" containerID="2a52617105aca80bd6aa782680de53fad558b6014f1d373ef1592509df2dabc0" Jan 26 17:40:55 crc kubenswrapper[4754]: E0126 17:40:55.768780 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x65wv_openshift-machine-config-operator(8c3718a4-f354-4284-92e0-fdfb45a692bd)\"" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" Jan 26 17:41:06 crc kubenswrapper[4754]: I0126 17:41:06.768025 4754 scope.go:117] "RemoveContainer" containerID="2a52617105aca80bd6aa782680de53fad558b6014f1d373ef1592509df2dabc0" Jan 26 17:41:06 crc kubenswrapper[4754]: E0126 17:41:06.768966 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x65wv_openshift-machine-config-operator(8c3718a4-f354-4284-92e0-fdfb45a692bd)\"" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" Jan 26 17:41:19 crc kubenswrapper[4754]: I0126 17:41:19.767335 4754 scope.go:117] "RemoveContainer" containerID="2a52617105aca80bd6aa782680de53fad558b6014f1d373ef1592509df2dabc0" Jan 26 17:41:19 crc kubenswrapper[4754]: E0126 17:41:19.768092 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x65wv_openshift-machine-config-operator(8c3718a4-f354-4284-92e0-fdfb45a692bd)\"" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" Jan 26 17:41:22 crc kubenswrapper[4754]: I0126 17:41:22.955098 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-vq6nk"] Jan 26 17:41:22 crc kubenswrapper[4754]: E0126 17:41:22.955993 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b1708bb9-e67e-4bab-96ea-cc492ea78d2b" containerName="registry-server" Jan 26 17:41:22 crc kubenswrapper[4754]: I0126 17:41:22.956007 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="b1708bb9-e67e-4bab-96ea-cc492ea78d2b" containerName="registry-server" Jan 26 17:41:22 crc kubenswrapper[4754]: E0126 17:41:22.956023 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b1708bb9-e67e-4bab-96ea-cc492ea78d2b" containerName="extract-content" Jan 26 17:41:22 crc kubenswrapper[4754]: I0126 17:41:22.956029 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="b1708bb9-e67e-4bab-96ea-cc492ea78d2b" containerName="extract-content" Jan 26 17:41:22 crc kubenswrapper[4754]: E0126 17:41:22.956039 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b1708bb9-e67e-4bab-96ea-cc492ea78d2b" containerName="extract-utilities" Jan 26 17:41:22 crc kubenswrapper[4754]: I0126 17:41:22.956046 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="b1708bb9-e67e-4bab-96ea-cc492ea78d2b" containerName="extract-utilities" Jan 26 17:41:22 crc kubenswrapper[4754]: I0126 17:41:22.956243 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="b1708bb9-e67e-4bab-96ea-cc492ea78d2b" containerName="registry-server" Jan 26 17:41:22 crc kubenswrapper[4754]: I0126 17:41:22.957725 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vq6nk" Jan 26 17:41:22 crc kubenswrapper[4754]: I0126 17:41:22.988369 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-vq6nk"] Jan 26 17:41:23 crc kubenswrapper[4754]: I0126 17:41:23.139933 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vccsq\" (UniqueName: \"kubernetes.io/projected/4ecf3cba-4837-4066-8f6e-11ecd9eaf828-kube-api-access-vccsq\") pod \"certified-operators-vq6nk\" (UID: \"4ecf3cba-4837-4066-8f6e-11ecd9eaf828\") " pod="openshift-marketplace/certified-operators-vq6nk" Jan 26 17:41:23 crc kubenswrapper[4754]: I0126 17:41:23.140069 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4ecf3cba-4837-4066-8f6e-11ecd9eaf828-catalog-content\") pod \"certified-operators-vq6nk\" (UID: \"4ecf3cba-4837-4066-8f6e-11ecd9eaf828\") " pod="openshift-marketplace/certified-operators-vq6nk" Jan 26 17:41:23 crc kubenswrapper[4754]: I0126 17:41:23.140139 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4ecf3cba-4837-4066-8f6e-11ecd9eaf828-utilities\") pod \"certified-operators-vq6nk\" (UID: \"4ecf3cba-4837-4066-8f6e-11ecd9eaf828\") " pod="openshift-marketplace/certified-operators-vq6nk" Jan 26 17:41:23 crc kubenswrapper[4754]: I0126 17:41:23.241631 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vccsq\" (UniqueName: \"kubernetes.io/projected/4ecf3cba-4837-4066-8f6e-11ecd9eaf828-kube-api-access-vccsq\") pod \"certified-operators-vq6nk\" (UID: \"4ecf3cba-4837-4066-8f6e-11ecd9eaf828\") " pod="openshift-marketplace/certified-operators-vq6nk" Jan 26 17:41:23 crc kubenswrapper[4754]: I0126 17:41:23.241936 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4ecf3cba-4837-4066-8f6e-11ecd9eaf828-catalog-content\") pod \"certified-operators-vq6nk\" (UID: \"4ecf3cba-4837-4066-8f6e-11ecd9eaf828\") " pod="openshift-marketplace/certified-operators-vq6nk" Jan 26 17:41:23 crc kubenswrapper[4754]: I0126 17:41:23.242028 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4ecf3cba-4837-4066-8f6e-11ecd9eaf828-utilities\") pod \"certified-operators-vq6nk\" (UID: \"4ecf3cba-4837-4066-8f6e-11ecd9eaf828\") " pod="openshift-marketplace/certified-operators-vq6nk" Jan 26 17:41:23 crc kubenswrapper[4754]: I0126 17:41:23.242640 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4ecf3cba-4837-4066-8f6e-11ecd9eaf828-catalog-content\") pod \"certified-operators-vq6nk\" (UID: \"4ecf3cba-4837-4066-8f6e-11ecd9eaf828\") " pod="openshift-marketplace/certified-operators-vq6nk" Jan 26 17:41:23 crc kubenswrapper[4754]: I0126 17:41:23.242718 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4ecf3cba-4837-4066-8f6e-11ecd9eaf828-utilities\") pod \"certified-operators-vq6nk\" (UID: \"4ecf3cba-4837-4066-8f6e-11ecd9eaf828\") " pod="openshift-marketplace/certified-operators-vq6nk" Jan 26 17:41:23 crc kubenswrapper[4754]: I0126 17:41:23.272713 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vccsq\" (UniqueName: \"kubernetes.io/projected/4ecf3cba-4837-4066-8f6e-11ecd9eaf828-kube-api-access-vccsq\") pod \"certified-operators-vq6nk\" (UID: \"4ecf3cba-4837-4066-8f6e-11ecd9eaf828\") " pod="openshift-marketplace/certified-operators-vq6nk" Jan 26 17:41:23 crc kubenswrapper[4754]: I0126 17:41:23.281903 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vq6nk" Jan 26 17:41:23 crc kubenswrapper[4754]: I0126 17:41:23.797538 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-vq6nk"] Jan 26 17:41:24 crc kubenswrapper[4754]: I0126 17:41:24.580975 4754 generic.go:334] "Generic (PLEG): container finished" podID="4ecf3cba-4837-4066-8f6e-11ecd9eaf828" containerID="eb99c4bb35ef7ed307b05df51ea3eeb680030ebfa56445cad14456a3a960f920" exitCode=0 Jan 26 17:41:24 crc kubenswrapper[4754]: I0126 17:41:24.581028 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vq6nk" event={"ID":"4ecf3cba-4837-4066-8f6e-11ecd9eaf828","Type":"ContainerDied","Data":"eb99c4bb35ef7ed307b05df51ea3eeb680030ebfa56445cad14456a3a960f920"} Jan 26 17:41:24 crc kubenswrapper[4754]: I0126 17:41:24.581056 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vq6nk" event={"ID":"4ecf3cba-4837-4066-8f6e-11ecd9eaf828","Type":"ContainerStarted","Data":"bcfe71a60bfa5920cb56e3f28ac267e321d9d57b99b6c16549375272337e542e"} Jan 26 17:41:27 crc kubenswrapper[4754]: I0126 17:41:27.613465 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vq6nk" event={"ID":"4ecf3cba-4837-4066-8f6e-11ecd9eaf828","Type":"ContainerStarted","Data":"01f5d56226501ae5505542ac5638255699dc4138007c80da376ba41befa76b58"} Jan 26 17:41:28 crc kubenswrapper[4754]: I0126 17:41:28.628006 4754 generic.go:334] "Generic (PLEG): container finished" podID="4ecf3cba-4837-4066-8f6e-11ecd9eaf828" containerID="01f5d56226501ae5505542ac5638255699dc4138007c80da376ba41befa76b58" exitCode=0 Jan 26 17:41:28 crc kubenswrapper[4754]: I0126 17:41:28.628078 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vq6nk" event={"ID":"4ecf3cba-4837-4066-8f6e-11ecd9eaf828","Type":"ContainerDied","Data":"01f5d56226501ae5505542ac5638255699dc4138007c80da376ba41befa76b58"} Jan 26 17:41:28 crc kubenswrapper[4754]: I0126 17:41:28.630226 4754 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 26 17:41:29 crc kubenswrapper[4754]: I0126 17:41:29.640608 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vq6nk" event={"ID":"4ecf3cba-4837-4066-8f6e-11ecd9eaf828","Type":"ContainerStarted","Data":"6ed71d7f9e62f95ef37b9ddc0b42ef44d296f1aa4ca0f732c8e20526b492e2f6"} Jan 26 17:41:29 crc kubenswrapper[4754]: I0126 17:41:29.658048 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-vq6nk" podStartSLOduration=3.220606778 podStartE2EDuration="7.658025014s" podCreationTimestamp="2026-01-26 17:41:22 +0000 UTC" firstStartedPulling="2026-01-26 17:41:24.582812473 +0000 UTC m=+3251.106992907" lastFinishedPulling="2026-01-26 17:41:29.020230709 +0000 UTC m=+3255.544411143" observedRunningTime="2026-01-26 17:41:29.65751148 +0000 UTC m=+3256.181691914" watchObservedRunningTime="2026-01-26 17:41:29.658025014 +0000 UTC m=+3256.182205458" Jan 26 17:41:32 crc kubenswrapper[4754]: I0126 17:41:32.767605 4754 scope.go:117] "RemoveContainer" containerID="2a52617105aca80bd6aa782680de53fad558b6014f1d373ef1592509df2dabc0" Jan 26 17:41:32 crc kubenswrapper[4754]: E0126 17:41:32.768510 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x65wv_openshift-machine-config-operator(8c3718a4-f354-4284-92e0-fdfb45a692bd)\"" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" Jan 26 17:41:33 crc kubenswrapper[4754]: I0126 17:41:33.282436 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-vq6nk" Jan 26 17:41:33 crc kubenswrapper[4754]: I0126 17:41:33.282771 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-vq6nk" Jan 26 17:41:33 crc kubenswrapper[4754]: I0126 17:41:33.326232 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-vq6nk" Jan 26 17:41:43 crc kubenswrapper[4754]: I0126 17:41:43.335848 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-vq6nk" Jan 26 17:41:43 crc kubenswrapper[4754]: I0126 17:41:43.392138 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-vq6nk"] Jan 26 17:41:43 crc kubenswrapper[4754]: I0126 17:41:43.800211 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-vq6nk" podUID="4ecf3cba-4837-4066-8f6e-11ecd9eaf828" containerName="registry-server" containerID="cri-o://6ed71d7f9e62f95ef37b9ddc0b42ef44d296f1aa4ca0f732c8e20526b492e2f6" gracePeriod=2 Jan 26 17:41:44 crc kubenswrapper[4754]: I0126 17:41:44.818792 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vq6nk" Jan 26 17:41:44 crc kubenswrapper[4754]: I0126 17:41:44.819037 4754 generic.go:334] "Generic (PLEG): container finished" podID="4ecf3cba-4837-4066-8f6e-11ecd9eaf828" containerID="6ed71d7f9e62f95ef37b9ddc0b42ef44d296f1aa4ca0f732c8e20526b492e2f6" exitCode=0 Jan 26 17:41:44 crc kubenswrapper[4754]: I0126 17:41:44.819067 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vq6nk" event={"ID":"4ecf3cba-4837-4066-8f6e-11ecd9eaf828","Type":"ContainerDied","Data":"6ed71d7f9e62f95ef37b9ddc0b42ef44d296f1aa4ca0f732c8e20526b492e2f6"} Jan 26 17:41:44 crc kubenswrapper[4754]: I0126 17:41:44.821346 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vq6nk" event={"ID":"4ecf3cba-4837-4066-8f6e-11ecd9eaf828","Type":"ContainerDied","Data":"bcfe71a60bfa5920cb56e3f28ac267e321d9d57b99b6c16549375272337e542e"} Jan 26 17:41:44 crc kubenswrapper[4754]: I0126 17:41:44.821507 4754 scope.go:117] "RemoveContainer" containerID="6ed71d7f9e62f95ef37b9ddc0b42ef44d296f1aa4ca0f732c8e20526b492e2f6" Jan 26 17:41:44 crc kubenswrapper[4754]: I0126 17:41:44.865045 4754 scope.go:117] "RemoveContainer" containerID="01f5d56226501ae5505542ac5638255699dc4138007c80da376ba41befa76b58" Jan 26 17:41:44 crc kubenswrapper[4754]: I0126 17:41:44.971179 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4ecf3cba-4837-4066-8f6e-11ecd9eaf828-utilities\") pod \"4ecf3cba-4837-4066-8f6e-11ecd9eaf828\" (UID: \"4ecf3cba-4837-4066-8f6e-11ecd9eaf828\") " Jan 26 17:41:44 crc kubenswrapper[4754]: I0126 17:41:44.971352 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4ecf3cba-4837-4066-8f6e-11ecd9eaf828-catalog-content\") pod \"4ecf3cba-4837-4066-8f6e-11ecd9eaf828\" (UID: \"4ecf3cba-4837-4066-8f6e-11ecd9eaf828\") " Jan 26 17:41:44 crc kubenswrapper[4754]: I0126 17:41:44.971551 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vccsq\" (UniqueName: \"kubernetes.io/projected/4ecf3cba-4837-4066-8f6e-11ecd9eaf828-kube-api-access-vccsq\") pod \"4ecf3cba-4837-4066-8f6e-11ecd9eaf828\" (UID: \"4ecf3cba-4837-4066-8f6e-11ecd9eaf828\") " Jan 26 17:41:44 crc kubenswrapper[4754]: I0126 17:41:44.972397 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4ecf3cba-4837-4066-8f6e-11ecd9eaf828-utilities" (OuterVolumeSpecName: "utilities") pod "4ecf3cba-4837-4066-8f6e-11ecd9eaf828" (UID: "4ecf3cba-4837-4066-8f6e-11ecd9eaf828"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:41:44 crc kubenswrapper[4754]: I0126 17:41:44.977651 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4ecf3cba-4837-4066-8f6e-11ecd9eaf828-kube-api-access-vccsq" (OuterVolumeSpecName: "kube-api-access-vccsq") pod "4ecf3cba-4837-4066-8f6e-11ecd9eaf828" (UID: "4ecf3cba-4837-4066-8f6e-11ecd9eaf828"). InnerVolumeSpecName "kube-api-access-vccsq". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:41:45 crc kubenswrapper[4754]: I0126 17:41:45.026600 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4ecf3cba-4837-4066-8f6e-11ecd9eaf828-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4ecf3cba-4837-4066-8f6e-11ecd9eaf828" (UID: "4ecf3cba-4837-4066-8f6e-11ecd9eaf828"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:41:45 crc kubenswrapper[4754]: I0126 17:41:45.073948 4754 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4ecf3cba-4837-4066-8f6e-11ecd9eaf828-utilities\") on node \"crc\" DevicePath \"\"" Jan 26 17:41:45 crc kubenswrapper[4754]: I0126 17:41:45.073996 4754 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4ecf3cba-4837-4066-8f6e-11ecd9eaf828-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 26 17:41:45 crc kubenswrapper[4754]: I0126 17:41:45.074006 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vccsq\" (UniqueName: \"kubernetes.io/projected/4ecf3cba-4837-4066-8f6e-11ecd9eaf828-kube-api-access-vccsq\") on node \"crc\" DevicePath \"\"" Jan 26 17:41:45 crc kubenswrapper[4754]: I0126 17:41:45.835331 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vq6nk" Jan 26 17:41:45 crc kubenswrapper[4754]: I0126 17:41:45.871113 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-vq6nk"] Jan 26 17:41:45 crc kubenswrapper[4754]: I0126 17:41:45.881877 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-vq6nk"] Jan 26 17:41:45 crc kubenswrapper[4754]: I0126 17:41:45.985280 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-q846q"] Jan 26 17:41:45 crc kubenswrapper[4754]: E0126 17:41:45.985824 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ecf3cba-4837-4066-8f6e-11ecd9eaf828" containerName="extract-utilities" Jan 26 17:41:45 crc kubenswrapper[4754]: I0126 17:41:45.985846 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ecf3cba-4837-4066-8f6e-11ecd9eaf828" containerName="extract-utilities" Jan 26 17:41:45 crc kubenswrapper[4754]: E0126 17:41:45.985862 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ecf3cba-4837-4066-8f6e-11ecd9eaf828" containerName="registry-server" Jan 26 17:41:45 crc kubenswrapper[4754]: I0126 17:41:45.985871 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ecf3cba-4837-4066-8f6e-11ecd9eaf828" containerName="registry-server" Jan 26 17:41:45 crc kubenswrapper[4754]: E0126 17:41:45.985893 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ecf3cba-4837-4066-8f6e-11ecd9eaf828" containerName="extract-content" Jan 26 17:41:45 crc kubenswrapper[4754]: I0126 17:41:45.985902 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ecf3cba-4837-4066-8f6e-11ecd9eaf828" containerName="extract-content" Jan 26 17:41:45 crc kubenswrapper[4754]: I0126 17:41:45.986132 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="4ecf3cba-4837-4066-8f6e-11ecd9eaf828" containerName="registry-server" Jan 26 17:41:45 crc kubenswrapper[4754]: I0126 17:41:45.987531 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-q846q" Jan 26 17:41:45 crc kubenswrapper[4754]: I0126 17:41:45.995874 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-q846q"] Jan 26 17:41:46 crc kubenswrapper[4754]: I0126 17:41:46.094331 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fz594\" (UniqueName: \"kubernetes.io/projected/66ae83ce-79e7-46a0-9660-a1b18fa2d693-kube-api-access-fz594\") pod \"redhat-operators-q846q\" (UID: \"66ae83ce-79e7-46a0-9660-a1b18fa2d693\") " pod="openshift-marketplace/redhat-operators-q846q" Jan 26 17:41:46 crc kubenswrapper[4754]: I0126 17:41:46.094386 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/66ae83ce-79e7-46a0-9660-a1b18fa2d693-catalog-content\") pod \"redhat-operators-q846q\" (UID: \"66ae83ce-79e7-46a0-9660-a1b18fa2d693\") " pod="openshift-marketplace/redhat-operators-q846q" Jan 26 17:41:46 crc kubenswrapper[4754]: I0126 17:41:46.094816 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/66ae83ce-79e7-46a0-9660-a1b18fa2d693-utilities\") pod \"redhat-operators-q846q\" (UID: \"66ae83ce-79e7-46a0-9660-a1b18fa2d693\") " pod="openshift-marketplace/redhat-operators-q846q" Jan 26 17:41:46 crc kubenswrapper[4754]: I0126 17:41:46.197050 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fz594\" (UniqueName: \"kubernetes.io/projected/66ae83ce-79e7-46a0-9660-a1b18fa2d693-kube-api-access-fz594\") pod \"redhat-operators-q846q\" (UID: \"66ae83ce-79e7-46a0-9660-a1b18fa2d693\") " pod="openshift-marketplace/redhat-operators-q846q" Jan 26 17:41:46 crc kubenswrapper[4754]: I0126 17:41:46.197104 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/66ae83ce-79e7-46a0-9660-a1b18fa2d693-catalog-content\") pod \"redhat-operators-q846q\" (UID: \"66ae83ce-79e7-46a0-9660-a1b18fa2d693\") " pod="openshift-marketplace/redhat-operators-q846q" Jan 26 17:41:46 crc kubenswrapper[4754]: I0126 17:41:46.197213 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/66ae83ce-79e7-46a0-9660-a1b18fa2d693-utilities\") pod \"redhat-operators-q846q\" (UID: \"66ae83ce-79e7-46a0-9660-a1b18fa2d693\") " pod="openshift-marketplace/redhat-operators-q846q" Jan 26 17:41:46 crc kubenswrapper[4754]: I0126 17:41:46.197738 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/66ae83ce-79e7-46a0-9660-a1b18fa2d693-utilities\") pod \"redhat-operators-q846q\" (UID: \"66ae83ce-79e7-46a0-9660-a1b18fa2d693\") " pod="openshift-marketplace/redhat-operators-q846q" Jan 26 17:41:46 crc kubenswrapper[4754]: I0126 17:41:46.197738 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/66ae83ce-79e7-46a0-9660-a1b18fa2d693-catalog-content\") pod \"redhat-operators-q846q\" (UID: \"66ae83ce-79e7-46a0-9660-a1b18fa2d693\") " pod="openshift-marketplace/redhat-operators-q846q" Jan 26 17:41:46 crc kubenswrapper[4754]: I0126 17:41:46.218427 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fz594\" (UniqueName: \"kubernetes.io/projected/66ae83ce-79e7-46a0-9660-a1b18fa2d693-kube-api-access-fz594\") pod \"redhat-operators-q846q\" (UID: \"66ae83ce-79e7-46a0-9660-a1b18fa2d693\") " pod="openshift-marketplace/redhat-operators-q846q" Jan 26 17:41:46 crc kubenswrapper[4754]: I0126 17:41:46.320790 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-q846q" Jan 26 17:41:46 crc kubenswrapper[4754]: I0126 17:41:46.387016 4754 scope.go:117] "RemoveContainer" containerID="eb99c4bb35ef7ed307b05df51ea3eeb680030ebfa56445cad14456a3a960f920" Jan 26 17:41:46 crc kubenswrapper[4754]: I0126 17:41:46.412091 4754 scope.go:117] "RemoveContainer" containerID="6ed71d7f9e62f95ef37b9ddc0b42ef44d296f1aa4ca0f732c8e20526b492e2f6" Jan 26 17:41:46 crc kubenswrapper[4754]: E0126 17:41:46.415783 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6ed71d7f9e62f95ef37b9ddc0b42ef44d296f1aa4ca0f732c8e20526b492e2f6\": container with ID starting with 6ed71d7f9e62f95ef37b9ddc0b42ef44d296f1aa4ca0f732c8e20526b492e2f6 not found: ID does not exist" containerID="6ed71d7f9e62f95ef37b9ddc0b42ef44d296f1aa4ca0f732c8e20526b492e2f6" Jan 26 17:41:46 crc kubenswrapper[4754]: I0126 17:41:46.415836 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6ed71d7f9e62f95ef37b9ddc0b42ef44d296f1aa4ca0f732c8e20526b492e2f6"} err="failed to get container status \"6ed71d7f9e62f95ef37b9ddc0b42ef44d296f1aa4ca0f732c8e20526b492e2f6\": rpc error: code = NotFound desc = could not find container \"6ed71d7f9e62f95ef37b9ddc0b42ef44d296f1aa4ca0f732c8e20526b492e2f6\": container with ID starting with 6ed71d7f9e62f95ef37b9ddc0b42ef44d296f1aa4ca0f732c8e20526b492e2f6 not found: ID does not exist" Jan 26 17:41:46 crc kubenswrapper[4754]: I0126 17:41:46.415861 4754 scope.go:117] "RemoveContainer" containerID="01f5d56226501ae5505542ac5638255699dc4138007c80da376ba41befa76b58" Jan 26 17:41:46 crc kubenswrapper[4754]: E0126 17:41:46.416206 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"01f5d56226501ae5505542ac5638255699dc4138007c80da376ba41befa76b58\": container with ID starting with 01f5d56226501ae5505542ac5638255699dc4138007c80da376ba41befa76b58 not found: ID does not exist" containerID="01f5d56226501ae5505542ac5638255699dc4138007c80da376ba41befa76b58" Jan 26 17:41:46 crc kubenswrapper[4754]: I0126 17:41:46.416261 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"01f5d56226501ae5505542ac5638255699dc4138007c80da376ba41befa76b58"} err="failed to get container status \"01f5d56226501ae5505542ac5638255699dc4138007c80da376ba41befa76b58\": rpc error: code = NotFound desc = could not find container \"01f5d56226501ae5505542ac5638255699dc4138007c80da376ba41befa76b58\": container with ID starting with 01f5d56226501ae5505542ac5638255699dc4138007c80da376ba41befa76b58 not found: ID does not exist" Jan 26 17:41:46 crc kubenswrapper[4754]: I0126 17:41:46.416289 4754 scope.go:117] "RemoveContainer" containerID="eb99c4bb35ef7ed307b05df51ea3eeb680030ebfa56445cad14456a3a960f920" Jan 26 17:41:46 crc kubenswrapper[4754]: E0126 17:41:46.416756 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eb99c4bb35ef7ed307b05df51ea3eeb680030ebfa56445cad14456a3a960f920\": container with ID starting with eb99c4bb35ef7ed307b05df51ea3eeb680030ebfa56445cad14456a3a960f920 not found: ID does not exist" containerID="eb99c4bb35ef7ed307b05df51ea3eeb680030ebfa56445cad14456a3a960f920" Jan 26 17:41:46 crc kubenswrapper[4754]: I0126 17:41:46.416820 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eb99c4bb35ef7ed307b05df51ea3eeb680030ebfa56445cad14456a3a960f920"} err="failed to get container status \"eb99c4bb35ef7ed307b05df51ea3eeb680030ebfa56445cad14456a3a960f920\": rpc error: code = NotFound desc = could not find container \"eb99c4bb35ef7ed307b05df51ea3eeb680030ebfa56445cad14456a3a960f920\": container with ID starting with eb99c4bb35ef7ed307b05df51ea3eeb680030ebfa56445cad14456a3a960f920 not found: ID does not exist" Jan 26 17:41:46 crc kubenswrapper[4754]: I0126 17:41:46.768251 4754 scope.go:117] "RemoveContainer" containerID="2a52617105aca80bd6aa782680de53fad558b6014f1d373ef1592509df2dabc0" Jan 26 17:41:46 crc kubenswrapper[4754]: I0126 17:41:46.929126 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-q846q"] Jan 26 17:41:47 crc kubenswrapper[4754]: I0126 17:41:47.784514 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4ecf3cba-4837-4066-8f6e-11ecd9eaf828" path="/var/lib/kubelet/pods/4ecf3cba-4837-4066-8f6e-11ecd9eaf828/volumes" Jan 26 17:41:47 crc kubenswrapper[4754]: I0126 17:41:47.864905 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" event={"ID":"8c3718a4-f354-4284-92e0-fdfb45a692bd","Type":"ContainerStarted","Data":"b8ddaeea1adb9fa522ab28e802b3b1b023114966ebf93bdd7941ca64dfbc9362"} Jan 26 17:41:47 crc kubenswrapper[4754]: I0126 17:41:47.868781 4754 generic.go:334] "Generic (PLEG): container finished" podID="66ae83ce-79e7-46a0-9660-a1b18fa2d693" containerID="e4e79c84e7e9a4bd3bd6ebcd0264319bbfbcd496dbfed27be623f48f917deac6" exitCode=0 Jan 26 17:41:47 crc kubenswrapper[4754]: I0126 17:41:47.868838 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q846q" event={"ID":"66ae83ce-79e7-46a0-9660-a1b18fa2d693","Type":"ContainerDied","Data":"e4e79c84e7e9a4bd3bd6ebcd0264319bbfbcd496dbfed27be623f48f917deac6"} Jan 26 17:41:47 crc kubenswrapper[4754]: I0126 17:41:47.868869 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q846q" event={"ID":"66ae83ce-79e7-46a0-9660-a1b18fa2d693","Type":"ContainerStarted","Data":"97683b112c2268073747fd2ac6e583f4f059c40578fd77ca8cdeb93b57495e44"} Jan 26 17:41:49 crc kubenswrapper[4754]: I0126 17:41:49.887776 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q846q" event={"ID":"66ae83ce-79e7-46a0-9660-a1b18fa2d693","Type":"ContainerStarted","Data":"ee0b7f1a0b15026d4a7851c3ff7f0a0e9e2a6532c63aaf32318c699b15288004"} Jan 26 17:41:52 crc kubenswrapper[4754]: I0126 17:41:52.914430 4754 generic.go:334] "Generic (PLEG): container finished" podID="66ae83ce-79e7-46a0-9660-a1b18fa2d693" containerID="ee0b7f1a0b15026d4a7851c3ff7f0a0e9e2a6532c63aaf32318c699b15288004" exitCode=0 Jan 26 17:41:52 crc kubenswrapper[4754]: I0126 17:41:52.914516 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q846q" event={"ID":"66ae83ce-79e7-46a0-9660-a1b18fa2d693","Type":"ContainerDied","Data":"ee0b7f1a0b15026d4a7851c3ff7f0a0e9e2a6532c63aaf32318c699b15288004"} Jan 26 17:41:55 crc kubenswrapper[4754]: I0126 17:41:55.942347 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q846q" event={"ID":"66ae83ce-79e7-46a0-9660-a1b18fa2d693","Type":"ContainerStarted","Data":"50db7544cffa07834ce9bb905a3e27021eabebd0e88d732544e408fc69662350"} Jan 26 17:41:55 crc kubenswrapper[4754]: I0126 17:41:55.969170 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-q846q" podStartSLOduration=4.002907654 podStartE2EDuration="10.969150143s" podCreationTimestamp="2026-01-26 17:41:45 +0000 UTC" firstStartedPulling="2026-01-26 17:41:47.871750767 +0000 UTC m=+3274.395931221" lastFinishedPulling="2026-01-26 17:41:54.837993266 +0000 UTC m=+3281.362173710" observedRunningTime="2026-01-26 17:41:55.961754428 +0000 UTC m=+3282.485934872" watchObservedRunningTime="2026-01-26 17:41:55.969150143 +0000 UTC m=+3282.493330577" Jan 26 17:41:56 crc kubenswrapper[4754]: I0126 17:41:56.322033 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-q846q" Jan 26 17:41:56 crc kubenswrapper[4754]: I0126 17:41:56.322152 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-q846q" Jan 26 17:41:57 crc kubenswrapper[4754]: I0126 17:41:57.371417 4754 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-q846q" podUID="66ae83ce-79e7-46a0-9660-a1b18fa2d693" containerName="registry-server" probeResult="failure" output=< Jan 26 17:41:57 crc kubenswrapper[4754]: timeout: failed to connect service ":50051" within 1s Jan 26 17:41:57 crc kubenswrapper[4754]: > Jan 26 17:42:06 crc kubenswrapper[4754]: I0126 17:42:06.375897 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-q846q" Jan 26 17:42:06 crc kubenswrapper[4754]: I0126 17:42:06.422487 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-q846q" Jan 26 17:42:06 crc kubenswrapper[4754]: I0126 17:42:06.613995 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-q846q"] Jan 26 17:42:08 crc kubenswrapper[4754]: I0126 17:42:08.044693 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-q846q" podUID="66ae83ce-79e7-46a0-9660-a1b18fa2d693" containerName="registry-server" containerID="cri-o://50db7544cffa07834ce9bb905a3e27021eabebd0e88d732544e408fc69662350" gracePeriod=2 Jan 26 17:42:08 crc kubenswrapper[4754]: I0126 17:42:08.545058 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-q846q" Jan 26 17:42:08 crc kubenswrapper[4754]: I0126 17:42:08.666893 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fz594\" (UniqueName: \"kubernetes.io/projected/66ae83ce-79e7-46a0-9660-a1b18fa2d693-kube-api-access-fz594\") pod \"66ae83ce-79e7-46a0-9660-a1b18fa2d693\" (UID: \"66ae83ce-79e7-46a0-9660-a1b18fa2d693\") " Jan 26 17:42:08 crc kubenswrapper[4754]: I0126 17:42:08.667096 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/66ae83ce-79e7-46a0-9660-a1b18fa2d693-catalog-content\") pod \"66ae83ce-79e7-46a0-9660-a1b18fa2d693\" (UID: \"66ae83ce-79e7-46a0-9660-a1b18fa2d693\") " Jan 26 17:42:08 crc kubenswrapper[4754]: I0126 17:42:08.667130 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/66ae83ce-79e7-46a0-9660-a1b18fa2d693-utilities\") pod \"66ae83ce-79e7-46a0-9660-a1b18fa2d693\" (UID: \"66ae83ce-79e7-46a0-9660-a1b18fa2d693\") " Jan 26 17:42:08 crc kubenswrapper[4754]: I0126 17:42:08.668206 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/66ae83ce-79e7-46a0-9660-a1b18fa2d693-utilities" (OuterVolumeSpecName: "utilities") pod "66ae83ce-79e7-46a0-9660-a1b18fa2d693" (UID: "66ae83ce-79e7-46a0-9660-a1b18fa2d693"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:42:08 crc kubenswrapper[4754]: I0126 17:42:08.673349 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/66ae83ce-79e7-46a0-9660-a1b18fa2d693-kube-api-access-fz594" (OuterVolumeSpecName: "kube-api-access-fz594") pod "66ae83ce-79e7-46a0-9660-a1b18fa2d693" (UID: "66ae83ce-79e7-46a0-9660-a1b18fa2d693"). InnerVolumeSpecName "kube-api-access-fz594". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:42:08 crc kubenswrapper[4754]: I0126 17:42:08.769004 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fz594\" (UniqueName: \"kubernetes.io/projected/66ae83ce-79e7-46a0-9660-a1b18fa2d693-kube-api-access-fz594\") on node \"crc\" DevicePath \"\"" Jan 26 17:42:08 crc kubenswrapper[4754]: I0126 17:42:08.769283 4754 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/66ae83ce-79e7-46a0-9660-a1b18fa2d693-utilities\") on node \"crc\" DevicePath \"\"" Jan 26 17:42:08 crc kubenswrapper[4754]: I0126 17:42:08.783331 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/66ae83ce-79e7-46a0-9660-a1b18fa2d693-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "66ae83ce-79e7-46a0-9660-a1b18fa2d693" (UID: "66ae83ce-79e7-46a0-9660-a1b18fa2d693"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:42:08 crc kubenswrapper[4754]: I0126 17:42:08.871377 4754 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/66ae83ce-79e7-46a0-9660-a1b18fa2d693-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 26 17:42:09 crc kubenswrapper[4754]: I0126 17:42:09.055145 4754 generic.go:334] "Generic (PLEG): container finished" podID="66ae83ce-79e7-46a0-9660-a1b18fa2d693" containerID="50db7544cffa07834ce9bb905a3e27021eabebd0e88d732544e408fc69662350" exitCode=0 Jan 26 17:42:09 crc kubenswrapper[4754]: I0126 17:42:09.055194 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q846q" event={"ID":"66ae83ce-79e7-46a0-9660-a1b18fa2d693","Type":"ContainerDied","Data":"50db7544cffa07834ce9bb905a3e27021eabebd0e88d732544e408fc69662350"} Jan 26 17:42:09 crc kubenswrapper[4754]: I0126 17:42:09.055203 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-q846q" Jan 26 17:42:09 crc kubenswrapper[4754]: I0126 17:42:09.055223 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q846q" event={"ID":"66ae83ce-79e7-46a0-9660-a1b18fa2d693","Type":"ContainerDied","Data":"97683b112c2268073747fd2ac6e583f4f059c40578fd77ca8cdeb93b57495e44"} Jan 26 17:42:09 crc kubenswrapper[4754]: I0126 17:42:09.055247 4754 scope.go:117] "RemoveContainer" containerID="50db7544cffa07834ce9bb905a3e27021eabebd0e88d732544e408fc69662350" Jan 26 17:42:09 crc kubenswrapper[4754]: I0126 17:42:09.075915 4754 scope.go:117] "RemoveContainer" containerID="ee0b7f1a0b15026d4a7851c3ff7f0a0e9e2a6532c63aaf32318c699b15288004" Jan 26 17:42:09 crc kubenswrapper[4754]: I0126 17:42:09.097140 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-q846q"] Jan 26 17:42:09 crc kubenswrapper[4754]: I0126 17:42:09.107766 4754 scope.go:117] "RemoveContainer" containerID="e4e79c84e7e9a4bd3bd6ebcd0264319bbfbcd496dbfed27be623f48f917deac6" Jan 26 17:42:09 crc kubenswrapper[4754]: I0126 17:42:09.109382 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-q846q"] Jan 26 17:42:09 crc kubenswrapper[4754]: I0126 17:42:09.147851 4754 scope.go:117] "RemoveContainer" containerID="50db7544cffa07834ce9bb905a3e27021eabebd0e88d732544e408fc69662350" Jan 26 17:42:09 crc kubenswrapper[4754]: E0126 17:42:09.148237 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"50db7544cffa07834ce9bb905a3e27021eabebd0e88d732544e408fc69662350\": container with ID starting with 50db7544cffa07834ce9bb905a3e27021eabebd0e88d732544e408fc69662350 not found: ID does not exist" containerID="50db7544cffa07834ce9bb905a3e27021eabebd0e88d732544e408fc69662350" Jan 26 17:42:09 crc kubenswrapper[4754]: I0126 17:42:09.148306 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"50db7544cffa07834ce9bb905a3e27021eabebd0e88d732544e408fc69662350"} err="failed to get container status \"50db7544cffa07834ce9bb905a3e27021eabebd0e88d732544e408fc69662350\": rpc error: code = NotFound desc = could not find container \"50db7544cffa07834ce9bb905a3e27021eabebd0e88d732544e408fc69662350\": container with ID starting with 50db7544cffa07834ce9bb905a3e27021eabebd0e88d732544e408fc69662350 not found: ID does not exist" Jan 26 17:42:09 crc kubenswrapper[4754]: I0126 17:42:09.148333 4754 scope.go:117] "RemoveContainer" containerID="ee0b7f1a0b15026d4a7851c3ff7f0a0e9e2a6532c63aaf32318c699b15288004" Jan 26 17:42:09 crc kubenswrapper[4754]: E0126 17:42:09.148846 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ee0b7f1a0b15026d4a7851c3ff7f0a0e9e2a6532c63aaf32318c699b15288004\": container with ID starting with ee0b7f1a0b15026d4a7851c3ff7f0a0e9e2a6532c63aaf32318c699b15288004 not found: ID does not exist" containerID="ee0b7f1a0b15026d4a7851c3ff7f0a0e9e2a6532c63aaf32318c699b15288004" Jan 26 17:42:09 crc kubenswrapper[4754]: I0126 17:42:09.148886 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ee0b7f1a0b15026d4a7851c3ff7f0a0e9e2a6532c63aaf32318c699b15288004"} err="failed to get container status \"ee0b7f1a0b15026d4a7851c3ff7f0a0e9e2a6532c63aaf32318c699b15288004\": rpc error: code = NotFound desc = could not find container \"ee0b7f1a0b15026d4a7851c3ff7f0a0e9e2a6532c63aaf32318c699b15288004\": container with ID starting with ee0b7f1a0b15026d4a7851c3ff7f0a0e9e2a6532c63aaf32318c699b15288004 not found: ID does not exist" Jan 26 17:42:09 crc kubenswrapper[4754]: I0126 17:42:09.148916 4754 scope.go:117] "RemoveContainer" containerID="e4e79c84e7e9a4bd3bd6ebcd0264319bbfbcd496dbfed27be623f48f917deac6" Jan 26 17:42:09 crc kubenswrapper[4754]: E0126 17:42:09.149139 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e4e79c84e7e9a4bd3bd6ebcd0264319bbfbcd496dbfed27be623f48f917deac6\": container with ID starting with e4e79c84e7e9a4bd3bd6ebcd0264319bbfbcd496dbfed27be623f48f917deac6 not found: ID does not exist" containerID="e4e79c84e7e9a4bd3bd6ebcd0264319bbfbcd496dbfed27be623f48f917deac6" Jan 26 17:42:09 crc kubenswrapper[4754]: I0126 17:42:09.149164 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e4e79c84e7e9a4bd3bd6ebcd0264319bbfbcd496dbfed27be623f48f917deac6"} err="failed to get container status \"e4e79c84e7e9a4bd3bd6ebcd0264319bbfbcd496dbfed27be623f48f917deac6\": rpc error: code = NotFound desc = could not find container \"e4e79c84e7e9a4bd3bd6ebcd0264319bbfbcd496dbfed27be623f48f917deac6\": container with ID starting with e4e79c84e7e9a4bd3bd6ebcd0264319bbfbcd496dbfed27be623f48f917deac6 not found: ID does not exist" Jan 26 17:42:09 crc kubenswrapper[4754]: I0126 17:42:09.785105 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="66ae83ce-79e7-46a0-9660-a1b18fa2d693" path="/var/lib/kubelet/pods/66ae83ce-79e7-46a0-9660-a1b18fa2d693/volumes" Jan 26 17:44:07 crc kubenswrapper[4754]: I0126 17:44:07.129185 4754 patch_prober.go:28] interesting pod/machine-config-daemon-x65wv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 26 17:44:07 crc kubenswrapper[4754]: I0126 17:44:07.129657 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 26 17:44:21 crc kubenswrapper[4754]: I0126 17:44:21.167802 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-5j59d"] Jan 26 17:44:21 crc kubenswrapper[4754]: E0126 17:44:21.169385 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66ae83ce-79e7-46a0-9660-a1b18fa2d693" containerName="registry-server" Jan 26 17:44:21 crc kubenswrapper[4754]: I0126 17:44:21.169417 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="66ae83ce-79e7-46a0-9660-a1b18fa2d693" containerName="registry-server" Jan 26 17:44:21 crc kubenswrapper[4754]: E0126 17:44:21.169464 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66ae83ce-79e7-46a0-9660-a1b18fa2d693" containerName="extract-content" Jan 26 17:44:21 crc kubenswrapper[4754]: I0126 17:44:21.169482 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="66ae83ce-79e7-46a0-9660-a1b18fa2d693" containerName="extract-content" Jan 26 17:44:21 crc kubenswrapper[4754]: E0126 17:44:21.169521 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66ae83ce-79e7-46a0-9660-a1b18fa2d693" containerName="extract-utilities" Jan 26 17:44:21 crc kubenswrapper[4754]: I0126 17:44:21.169538 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="66ae83ce-79e7-46a0-9660-a1b18fa2d693" containerName="extract-utilities" Jan 26 17:44:21 crc kubenswrapper[4754]: I0126 17:44:21.170042 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="66ae83ce-79e7-46a0-9660-a1b18fa2d693" containerName="registry-server" Jan 26 17:44:21 crc kubenswrapper[4754]: I0126 17:44:21.196110 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-5j59d"] Jan 26 17:44:21 crc kubenswrapper[4754]: I0126 17:44:21.196279 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5j59d" Jan 26 17:44:21 crc kubenswrapper[4754]: I0126 17:44:21.296085 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d44be81c-a456-4822-9ff9-ce584324a994-catalog-content\") pod \"community-operators-5j59d\" (UID: \"d44be81c-a456-4822-9ff9-ce584324a994\") " pod="openshift-marketplace/community-operators-5j59d" Jan 26 17:44:21 crc kubenswrapper[4754]: I0126 17:44:21.296176 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qpjq8\" (UniqueName: \"kubernetes.io/projected/d44be81c-a456-4822-9ff9-ce584324a994-kube-api-access-qpjq8\") pod \"community-operators-5j59d\" (UID: \"d44be81c-a456-4822-9ff9-ce584324a994\") " pod="openshift-marketplace/community-operators-5j59d" Jan 26 17:44:21 crc kubenswrapper[4754]: I0126 17:44:21.296554 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d44be81c-a456-4822-9ff9-ce584324a994-utilities\") pod \"community-operators-5j59d\" (UID: \"d44be81c-a456-4822-9ff9-ce584324a994\") " pod="openshift-marketplace/community-operators-5j59d" Jan 26 17:44:21 crc kubenswrapper[4754]: I0126 17:44:21.398515 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d44be81c-a456-4822-9ff9-ce584324a994-utilities\") pod \"community-operators-5j59d\" (UID: \"d44be81c-a456-4822-9ff9-ce584324a994\") " pod="openshift-marketplace/community-operators-5j59d" Jan 26 17:44:21 crc kubenswrapper[4754]: I0126 17:44:21.398609 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d44be81c-a456-4822-9ff9-ce584324a994-catalog-content\") pod \"community-operators-5j59d\" (UID: \"d44be81c-a456-4822-9ff9-ce584324a994\") " pod="openshift-marketplace/community-operators-5j59d" Jan 26 17:44:21 crc kubenswrapper[4754]: I0126 17:44:21.398681 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qpjq8\" (UniqueName: \"kubernetes.io/projected/d44be81c-a456-4822-9ff9-ce584324a994-kube-api-access-qpjq8\") pod \"community-operators-5j59d\" (UID: \"d44be81c-a456-4822-9ff9-ce584324a994\") " pod="openshift-marketplace/community-operators-5j59d" Jan 26 17:44:21 crc kubenswrapper[4754]: I0126 17:44:21.399170 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d44be81c-a456-4822-9ff9-ce584324a994-utilities\") pod \"community-operators-5j59d\" (UID: \"d44be81c-a456-4822-9ff9-ce584324a994\") " pod="openshift-marketplace/community-operators-5j59d" Jan 26 17:44:21 crc kubenswrapper[4754]: I0126 17:44:21.399251 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d44be81c-a456-4822-9ff9-ce584324a994-catalog-content\") pod \"community-operators-5j59d\" (UID: \"d44be81c-a456-4822-9ff9-ce584324a994\") " pod="openshift-marketplace/community-operators-5j59d" Jan 26 17:44:21 crc kubenswrapper[4754]: I0126 17:44:21.420241 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qpjq8\" (UniqueName: \"kubernetes.io/projected/d44be81c-a456-4822-9ff9-ce584324a994-kube-api-access-qpjq8\") pod \"community-operators-5j59d\" (UID: \"d44be81c-a456-4822-9ff9-ce584324a994\") " pod="openshift-marketplace/community-operators-5j59d" Jan 26 17:44:21 crc kubenswrapper[4754]: I0126 17:44:21.530341 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5j59d" Jan 26 17:44:22 crc kubenswrapper[4754]: I0126 17:44:22.020648 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-5j59d"] Jan 26 17:44:22 crc kubenswrapper[4754]: W0126 17:44:22.028847 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd44be81c_a456_4822_9ff9_ce584324a994.slice/crio-710d235de07dc160c66519f6822263591fbf47c8ee6b788a3b3b31ad34a2a4bc WatchSource:0}: Error finding container 710d235de07dc160c66519f6822263591fbf47c8ee6b788a3b3b31ad34a2a4bc: Status 404 returned error can't find the container with id 710d235de07dc160c66519f6822263591fbf47c8ee6b788a3b3b31ad34a2a4bc Jan 26 17:44:22 crc kubenswrapper[4754]: I0126 17:44:22.218807 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5j59d" event={"ID":"d44be81c-a456-4822-9ff9-ce584324a994","Type":"ContainerStarted","Data":"710d235de07dc160c66519f6822263591fbf47c8ee6b788a3b3b31ad34a2a4bc"} Jan 26 17:44:23 crc kubenswrapper[4754]: I0126 17:44:23.234458 4754 generic.go:334] "Generic (PLEG): container finished" podID="d44be81c-a456-4822-9ff9-ce584324a994" containerID="e608faa06ec84c4f01f700532a7b3af9c40a6a082b841df9eb4c7da41ded05e9" exitCode=0 Jan 26 17:44:23 crc kubenswrapper[4754]: I0126 17:44:23.234530 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5j59d" event={"ID":"d44be81c-a456-4822-9ff9-ce584324a994","Type":"ContainerDied","Data":"e608faa06ec84c4f01f700532a7b3af9c40a6a082b841df9eb4c7da41ded05e9"} Jan 26 17:44:24 crc kubenswrapper[4754]: E0126 17:44:24.726379 4754 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd44be81c_a456_4822_9ff9_ce584324a994.slice/crio-ea549cc5ae7fd1fdea50c4d89a6bf4ab53249661eb60fc673c5256421f7bd5ef.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd44be81c_a456_4822_9ff9_ce584324a994.slice/crio-conmon-ea549cc5ae7fd1fdea50c4d89a6bf4ab53249661eb60fc673c5256421f7bd5ef.scope\": RecentStats: unable to find data in memory cache]" Jan 26 17:44:25 crc kubenswrapper[4754]: I0126 17:44:25.262954 4754 generic.go:334] "Generic (PLEG): container finished" podID="d44be81c-a456-4822-9ff9-ce584324a994" containerID="ea549cc5ae7fd1fdea50c4d89a6bf4ab53249661eb60fc673c5256421f7bd5ef" exitCode=0 Jan 26 17:44:25 crc kubenswrapper[4754]: I0126 17:44:25.263004 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5j59d" event={"ID":"d44be81c-a456-4822-9ff9-ce584324a994","Type":"ContainerDied","Data":"ea549cc5ae7fd1fdea50c4d89a6bf4ab53249661eb60fc673c5256421f7bd5ef"} Jan 26 17:44:26 crc kubenswrapper[4754]: I0126 17:44:26.274135 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5j59d" event={"ID":"d44be81c-a456-4822-9ff9-ce584324a994","Type":"ContainerStarted","Data":"d8de7bcfc1e4e60b10c7cd62b60962fe4d021613c28757a946d2f30f08161786"} Jan 26 17:44:26 crc kubenswrapper[4754]: I0126 17:44:26.301087 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-5j59d" podStartSLOduration=2.892037489 podStartE2EDuration="5.301062339s" podCreationTimestamp="2026-01-26 17:44:21 +0000 UTC" firstStartedPulling="2026-01-26 17:44:23.237273377 +0000 UTC m=+3429.761453811" lastFinishedPulling="2026-01-26 17:44:25.646298237 +0000 UTC m=+3432.170478661" observedRunningTime="2026-01-26 17:44:26.294693534 +0000 UTC m=+3432.818873998" watchObservedRunningTime="2026-01-26 17:44:26.301062339 +0000 UTC m=+3432.825242783" Jan 26 17:44:31 crc kubenswrapper[4754]: I0126 17:44:31.531166 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-5j59d" Jan 26 17:44:31 crc kubenswrapper[4754]: I0126 17:44:31.531869 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-5j59d" Jan 26 17:44:31 crc kubenswrapper[4754]: I0126 17:44:31.591833 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-5j59d" Jan 26 17:44:32 crc kubenswrapper[4754]: I0126 17:44:32.401344 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-5j59d" Jan 26 17:44:32 crc kubenswrapper[4754]: I0126 17:44:32.455931 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-5j59d"] Jan 26 17:44:34 crc kubenswrapper[4754]: I0126 17:44:34.341417 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-5j59d" podUID="d44be81c-a456-4822-9ff9-ce584324a994" containerName="registry-server" containerID="cri-o://d8de7bcfc1e4e60b10c7cd62b60962fe4d021613c28757a946d2f30f08161786" gracePeriod=2 Jan 26 17:44:34 crc kubenswrapper[4754]: I0126 17:44:34.794800 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5j59d" Jan 26 17:44:34 crc kubenswrapper[4754]: I0126 17:44:34.962322 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d44be81c-a456-4822-9ff9-ce584324a994-catalog-content\") pod \"d44be81c-a456-4822-9ff9-ce584324a994\" (UID: \"d44be81c-a456-4822-9ff9-ce584324a994\") " Jan 26 17:44:34 crc kubenswrapper[4754]: I0126 17:44:34.962638 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d44be81c-a456-4822-9ff9-ce584324a994-utilities\") pod \"d44be81c-a456-4822-9ff9-ce584324a994\" (UID: \"d44be81c-a456-4822-9ff9-ce584324a994\") " Jan 26 17:44:34 crc kubenswrapper[4754]: I0126 17:44:34.962724 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qpjq8\" (UniqueName: \"kubernetes.io/projected/d44be81c-a456-4822-9ff9-ce584324a994-kube-api-access-qpjq8\") pod \"d44be81c-a456-4822-9ff9-ce584324a994\" (UID: \"d44be81c-a456-4822-9ff9-ce584324a994\") " Jan 26 17:44:34 crc kubenswrapper[4754]: I0126 17:44:34.963588 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d44be81c-a456-4822-9ff9-ce584324a994-utilities" (OuterVolumeSpecName: "utilities") pod "d44be81c-a456-4822-9ff9-ce584324a994" (UID: "d44be81c-a456-4822-9ff9-ce584324a994"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:44:34 crc kubenswrapper[4754]: I0126 17:44:34.971357 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d44be81c-a456-4822-9ff9-ce584324a994-kube-api-access-qpjq8" (OuterVolumeSpecName: "kube-api-access-qpjq8") pod "d44be81c-a456-4822-9ff9-ce584324a994" (UID: "d44be81c-a456-4822-9ff9-ce584324a994"). InnerVolumeSpecName "kube-api-access-qpjq8". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:44:35 crc kubenswrapper[4754]: I0126 17:44:35.018518 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d44be81c-a456-4822-9ff9-ce584324a994-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d44be81c-a456-4822-9ff9-ce584324a994" (UID: "d44be81c-a456-4822-9ff9-ce584324a994"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:44:35 crc kubenswrapper[4754]: I0126 17:44:35.064974 4754 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d44be81c-a456-4822-9ff9-ce584324a994-utilities\") on node \"crc\" DevicePath \"\"" Jan 26 17:44:35 crc kubenswrapper[4754]: I0126 17:44:35.065233 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qpjq8\" (UniqueName: \"kubernetes.io/projected/d44be81c-a456-4822-9ff9-ce584324a994-kube-api-access-qpjq8\") on node \"crc\" DevicePath \"\"" Jan 26 17:44:35 crc kubenswrapper[4754]: I0126 17:44:35.065300 4754 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d44be81c-a456-4822-9ff9-ce584324a994-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 26 17:44:35 crc kubenswrapper[4754]: I0126 17:44:35.353363 4754 generic.go:334] "Generic (PLEG): container finished" podID="d44be81c-a456-4822-9ff9-ce584324a994" containerID="d8de7bcfc1e4e60b10c7cd62b60962fe4d021613c28757a946d2f30f08161786" exitCode=0 Jan 26 17:44:35 crc kubenswrapper[4754]: I0126 17:44:35.353912 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5j59d" event={"ID":"d44be81c-a456-4822-9ff9-ce584324a994","Type":"ContainerDied","Data":"d8de7bcfc1e4e60b10c7cd62b60962fe4d021613c28757a946d2f30f08161786"} Jan 26 17:44:35 crc kubenswrapper[4754]: I0126 17:44:35.353953 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5j59d" Jan 26 17:44:35 crc kubenswrapper[4754]: I0126 17:44:35.354422 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5j59d" event={"ID":"d44be81c-a456-4822-9ff9-ce584324a994","Type":"ContainerDied","Data":"710d235de07dc160c66519f6822263591fbf47c8ee6b788a3b3b31ad34a2a4bc"} Jan 26 17:44:35 crc kubenswrapper[4754]: I0126 17:44:35.354548 4754 scope.go:117] "RemoveContainer" containerID="d8de7bcfc1e4e60b10c7cd62b60962fe4d021613c28757a946d2f30f08161786" Jan 26 17:44:35 crc kubenswrapper[4754]: I0126 17:44:35.376279 4754 scope.go:117] "RemoveContainer" containerID="ea549cc5ae7fd1fdea50c4d89a6bf4ab53249661eb60fc673c5256421f7bd5ef" Jan 26 17:44:35 crc kubenswrapper[4754]: I0126 17:44:35.396395 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-5j59d"] Jan 26 17:44:35 crc kubenswrapper[4754]: I0126 17:44:35.402584 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-5j59d"] Jan 26 17:44:35 crc kubenswrapper[4754]: I0126 17:44:35.427849 4754 scope.go:117] "RemoveContainer" containerID="e608faa06ec84c4f01f700532a7b3af9c40a6a082b841df9eb4c7da41ded05e9" Jan 26 17:44:35 crc kubenswrapper[4754]: I0126 17:44:35.451482 4754 scope.go:117] "RemoveContainer" containerID="d8de7bcfc1e4e60b10c7cd62b60962fe4d021613c28757a946d2f30f08161786" Jan 26 17:44:35 crc kubenswrapper[4754]: E0126 17:44:35.452069 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d8de7bcfc1e4e60b10c7cd62b60962fe4d021613c28757a946d2f30f08161786\": container with ID starting with d8de7bcfc1e4e60b10c7cd62b60962fe4d021613c28757a946d2f30f08161786 not found: ID does not exist" containerID="d8de7bcfc1e4e60b10c7cd62b60962fe4d021613c28757a946d2f30f08161786" Jan 26 17:44:35 crc kubenswrapper[4754]: I0126 17:44:35.452123 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d8de7bcfc1e4e60b10c7cd62b60962fe4d021613c28757a946d2f30f08161786"} err="failed to get container status \"d8de7bcfc1e4e60b10c7cd62b60962fe4d021613c28757a946d2f30f08161786\": rpc error: code = NotFound desc = could not find container \"d8de7bcfc1e4e60b10c7cd62b60962fe4d021613c28757a946d2f30f08161786\": container with ID starting with d8de7bcfc1e4e60b10c7cd62b60962fe4d021613c28757a946d2f30f08161786 not found: ID does not exist" Jan 26 17:44:35 crc kubenswrapper[4754]: I0126 17:44:35.452176 4754 scope.go:117] "RemoveContainer" containerID="ea549cc5ae7fd1fdea50c4d89a6bf4ab53249661eb60fc673c5256421f7bd5ef" Jan 26 17:44:35 crc kubenswrapper[4754]: E0126 17:44:35.452658 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ea549cc5ae7fd1fdea50c4d89a6bf4ab53249661eb60fc673c5256421f7bd5ef\": container with ID starting with ea549cc5ae7fd1fdea50c4d89a6bf4ab53249661eb60fc673c5256421f7bd5ef not found: ID does not exist" containerID="ea549cc5ae7fd1fdea50c4d89a6bf4ab53249661eb60fc673c5256421f7bd5ef" Jan 26 17:44:35 crc kubenswrapper[4754]: I0126 17:44:35.452714 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ea549cc5ae7fd1fdea50c4d89a6bf4ab53249661eb60fc673c5256421f7bd5ef"} err="failed to get container status \"ea549cc5ae7fd1fdea50c4d89a6bf4ab53249661eb60fc673c5256421f7bd5ef\": rpc error: code = NotFound desc = could not find container \"ea549cc5ae7fd1fdea50c4d89a6bf4ab53249661eb60fc673c5256421f7bd5ef\": container with ID starting with ea549cc5ae7fd1fdea50c4d89a6bf4ab53249661eb60fc673c5256421f7bd5ef not found: ID does not exist" Jan 26 17:44:35 crc kubenswrapper[4754]: I0126 17:44:35.452738 4754 scope.go:117] "RemoveContainer" containerID="e608faa06ec84c4f01f700532a7b3af9c40a6a082b841df9eb4c7da41ded05e9" Jan 26 17:44:35 crc kubenswrapper[4754]: E0126 17:44:35.453164 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e608faa06ec84c4f01f700532a7b3af9c40a6a082b841df9eb4c7da41ded05e9\": container with ID starting with e608faa06ec84c4f01f700532a7b3af9c40a6a082b841df9eb4c7da41ded05e9 not found: ID does not exist" containerID="e608faa06ec84c4f01f700532a7b3af9c40a6a082b841df9eb4c7da41ded05e9" Jan 26 17:44:35 crc kubenswrapper[4754]: I0126 17:44:35.453192 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e608faa06ec84c4f01f700532a7b3af9c40a6a082b841df9eb4c7da41ded05e9"} err="failed to get container status \"e608faa06ec84c4f01f700532a7b3af9c40a6a082b841df9eb4c7da41ded05e9\": rpc error: code = NotFound desc = could not find container \"e608faa06ec84c4f01f700532a7b3af9c40a6a082b841df9eb4c7da41ded05e9\": container with ID starting with e608faa06ec84c4f01f700532a7b3af9c40a6a082b841df9eb4c7da41ded05e9 not found: ID does not exist" Jan 26 17:44:35 crc kubenswrapper[4754]: I0126 17:44:35.786793 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d44be81c-a456-4822-9ff9-ce584324a994" path="/var/lib/kubelet/pods/d44be81c-a456-4822-9ff9-ce584324a994/volumes" Jan 26 17:44:37 crc kubenswrapper[4754]: I0126 17:44:37.128847 4754 patch_prober.go:28] interesting pod/machine-config-daemon-x65wv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 26 17:44:37 crc kubenswrapper[4754]: I0126 17:44:37.129108 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 26 17:45:00 crc kubenswrapper[4754]: I0126 17:45:00.148554 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29490825-mpdwg"] Jan 26 17:45:00 crc kubenswrapper[4754]: E0126 17:45:00.149410 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d44be81c-a456-4822-9ff9-ce584324a994" containerName="extract-content" Jan 26 17:45:00 crc kubenswrapper[4754]: I0126 17:45:00.149423 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="d44be81c-a456-4822-9ff9-ce584324a994" containerName="extract-content" Jan 26 17:45:00 crc kubenswrapper[4754]: E0126 17:45:00.149441 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d44be81c-a456-4822-9ff9-ce584324a994" containerName="registry-server" Jan 26 17:45:00 crc kubenswrapper[4754]: I0126 17:45:00.149446 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="d44be81c-a456-4822-9ff9-ce584324a994" containerName="registry-server" Jan 26 17:45:00 crc kubenswrapper[4754]: E0126 17:45:00.149473 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d44be81c-a456-4822-9ff9-ce584324a994" containerName="extract-utilities" Jan 26 17:45:00 crc kubenswrapper[4754]: I0126 17:45:00.149479 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="d44be81c-a456-4822-9ff9-ce584324a994" containerName="extract-utilities" Jan 26 17:45:00 crc kubenswrapper[4754]: I0126 17:45:00.149647 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="d44be81c-a456-4822-9ff9-ce584324a994" containerName="registry-server" Jan 26 17:45:00 crc kubenswrapper[4754]: I0126 17:45:00.150303 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29490825-mpdwg" Jan 26 17:45:00 crc kubenswrapper[4754]: I0126 17:45:00.152812 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Jan 26 17:45:00 crc kubenswrapper[4754]: I0126 17:45:00.162132 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29490825-mpdwg"] Jan 26 17:45:00 crc kubenswrapper[4754]: I0126 17:45:00.182188 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Jan 26 17:45:00 crc kubenswrapper[4754]: I0126 17:45:00.234931 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/db7c55e8-5cf4-48d5-ab91-4795c5e52311-config-volume\") pod \"collect-profiles-29490825-mpdwg\" (UID: \"db7c55e8-5cf4-48d5-ab91-4795c5e52311\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29490825-mpdwg" Jan 26 17:45:00 crc kubenswrapper[4754]: I0126 17:45:00.234978 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rqdp7\" (UniqueName: \"kubernetes.io/projected/db7c55e8-5cf4-48d5-ab91-4795c5e52311-kube-api-access-rqdp7\") pod \"collect-profiles-29490825-mpdwg\" (UID: \"db7c55e8-5cf4-48d5-ab91-4795c5e52311\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29490825-mpdwg" Jan 26 17:45:00 crc kubenswrapper[4754]: I0126 17:45:00.235445 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/db7c55e8-5cf4-48d5-ab91-4795c5e52311-secret-volume\") pod \"collect-profiles-29490825-mpdwg\" (UID: \"db7c55e8-5cf4-48d5-ab91-4795c5e52311\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29490825-mpdwg" Jan 26 17:45:00 crc kubenswrapper[4754]: I0126 17:45:00.337075 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/db7c55e8-5cf4-48d5-ab91-4795c5e52311-secret-volume\") pod \"collect-profiles-29490825-mpdwg\" (UID: \"db7c55e8-5cf4-48d5-ab91-4795c5e52311\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29490825-mpdwg" Jan 26 17:45:00 crc kubenswrapper[4754]: I0126 17:45:00.337171 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/db7c55e8-5cf4-48d5-ab91-4795c5e52311-config-volume\") pod \"collect-profiles-29490825-mpdwg\" (UID: \"db7c55e8-5cf4-48d5-ab91-4795c5e52311\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29490825-mpdwg" Jan 26 17:45:00 crc kubenswrapper[4754]: I0126 17:45:00.337204 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rqdp7\" (UniqueName: \"kubernetes.io/projected/db7c55e8-5cf4-48d5-ab91-4795c5e52311-kube-api-access-rqdp7\") pod \"collect-profiles-29490825-mpdwg\" (UID: \"db7c55e8-5cf4-48d5-ab91-4795c5e52311\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29490825-mpdwg" Jan 26 17:45:00 crc kubenswrapper[4754]: I0126 17:45:00.338796 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/db7c55e8-5cf4-48d5-ab91-4795c5e52311-config-volume\") pod \"collect-profiles-29490825-mpdwg\" (UID: \"db7c55e8-5cf4-48d5-ab91-4795c5e52311\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29490825-mpdwg" Jan 26 17:45:00 crc kubenswrapper[4754]: I0126 17:45:00.344502 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/db7c55e8-5cf4-48d5-ab91-4795c5e52311-secret-volume\") pod \"collect-profiles-29490825-mpdwg\" (UID: \"db7c55e8-5cf4-48d5-ab91-4795c5e52311\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29490825-mpdwg" Jan 26 17:45:00 crc kubenswrapper[4754]: I0126 17:45:00.354852 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rqdp7\" (UniqueName: \"kubernetes.io/projected/db7c55e8-5cf4-48d5-ab91-4795c5e52311-kube-api-access-rqdp7\") pod \"collect-profiles-29490825-mpdwg\" (UID: \"db7c55e8-5cf4-48d5-ab91-4795c5e52311\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29490825-mpdwg" Jan 26 17:45:00 crc kubenswrapper[4754]: I0126 17:45:00.488835 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29490825-mpdwg" Jan 26 17:45:00 crc kubenswrapper[4754]: I0126 17:45:00.929302 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29490825-mpdwg"] Jan 26 17:45:01 crc kubenswrapper[4754]: I0126 17:45:01.565635 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29490825-mpdwg" event={"ID":"db7c55e8-5cf4-48d5-ab91-4795c5e52311","Type":"ContainerStarted","Data":"35cf69838a106f00bd63b4cd8d315032bd6e152fdb467f0a455244ddeba92ac5"} Jan 26 17:45:01 crc kubenswrapper[4754]: I0126 17:45:01.566031 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29490825-mpdwg" event={"ID":"db7c55e8-5cf4-48d5-ab91-4795c5e52311","Type":"ContainerStarted","Data":"2cf9e3e04a25356888f554eb8bcdb8fdb1a2775b02f6ffc25721fe8af60c3c0a"} Jan 26 17:45:01 crc kubenswrapper[4754]: I0126 17:45:01.583181 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29490825-mpdwg" podStartSLOduration=1.583164337 podStartE2EDuration="1.583164337s" podCreationTimestamp="2026-01-26 17:45:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 17:45:01.581046129 +0000 UTC m=+3468.105226563" watchObservedRunningTime="2026-01-26 17:45:01.583164337 +0000 UTC m=+3468.107344771" Jan 26 17:45:02 crc kubenswrapper[4754]: I0126 17:45:02.578286 4754 generic.go:334] "Generic (PLEG): container finished" podID="db7c55e8-5cf4-48d5-ab91-4795c5e52311" containerID="35cf69838a106f00bd63b4cd8d315032bd6e152fdb467f0a455244ddeba92ac5" exitCode=0 Jan 26 17:45:02 crc kubenswrapper[4754]: I0126 17:45:02.578352 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29490825-mpdwg" event={"ID":"db7c55e8-5cf4-48d5-ab91-4795c5e52311","Type":"ContainerDied","Data":"35cf69838a106f00bd63b4cd8d315032bd6e152fdb467f0a455244ddeba92ac5"} Jan 26 17:45:03 crc kubenswrapper[4754]: I0126 17:45:03.921507 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29490825-mpdwg" Jan 26 17:45:04 crc kubenswrapper[4754]: I0126 17:45:04.011296 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/db7c55e8-5cf4-48d5-ab91-4795c5e52311-config-volume\") pod \"db7c55e8-5cf4-48d5-ab91-4795c5e52311\" (UID: \"db7c55e8-5cf4-48d5-ab91-4795c5e52311\") " Jan 26 17:45:04 crc kubenswrapper[4754]: I0126 17:45:04.011409 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/db7c55e8-5cf4-48d5-ab91-4795c5e52311-secret-volume\") pod \"db7c55e8-5cf4-48d5-ab91-4795c5e52311\" (UID: \"db7c55e8-5cf4-48d5-ab91-4795c5e52311\") " Jan 26 17:45:04 crc kubenswrapper[4754]: I0126 17:45:04.011571 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rqdp7\" (UniqueName: \"kubernetes.io/projected/db7c55e8-5cf4-48d5-ab91-4795c5e52311-kube-api-access-rqdp7\") pod \"db7c55e8-5cf4-48d5-ab91-4795c5e52311\" (UID: \"db7c55e8-5cf4-48d5-ab91-4795c5e52311\") " Jan 26 17:45:04 crc kubenswrapper[4754]: I0126 17:45:04.012594 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/db7c55e8-5cf4-48d5-ab91-4795c5e52311-config-volume" (OuterVolumeSpecName: "config-volume") pod "db7c55e8-5cf4-48d5-ab91-4795c5e52311" (UID: "db7c55e8-5cf4-48d5-ab91-4795c5e52311"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:45:04 crc kubenswrapper[4754]: I0126 17:45:04.017703 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/db7c55e8-5cf4-48d5-ab91-4795c5e52311-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "db7c55e8-5cf4-48d5-ab91-4795c5e52311" (UID: "db7c55e8-5cf4-48d5-ab91-4795c5e52311"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:45:04 crc kubenswrapper[4754]: I0126 17:45:04.018094 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/db7c55e8-5cf4-48d5-ab91-4795c5e52311-kube-api-access-rqdp7" (OuterVolumeSpecName: "kube-api-access-rqdp7") pod "db7c55e8-5cf4-48d5-ab91-4795c5e52311" (UID: "db7c55e8-5cf4-48d5-ab91-4795c5e52311"). InnerVolumeSpecName "kube-api-access-rqdp7". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:45:04 crc kubenswrapper[4754]: I0126 17:45:04.114065 4754 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/db7c55e8-5cf4-48d5-ab91-4795c5e52311-config-volume\") on node \"crc\" DevicePath \"\"" Jan 26 17:45:04 crc kubenswrapper[4754]: I0126 17:45:04.114118 4754 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/db7c55e8-5cf4-48d5-ab91-4795c5e52311-secret-volume\") on node \"crc\" DevicePath \"\"" Jan 26 17:45:04 crc kubenswrapper[4754]: I0126 17:45:04.114135 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rqdp7\" (UniqueName: \"kubernetes.io/projected/db7c55e8-5cf4-48d5-ab91-4795c5e52311-kube-api-access-rqdp7\") on node \"crc\" DevicePath \"\"" Jan 26 17:45:04 crc kubenswrapper[4754]: I0126 17:45:04.595273 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29490825-mpdwg" event={"ID":"db7c55e8-5cf4-48d5-ab91-4795c5e52311","Type":"ContainerDied","Data":"2cf9e3e04a25356888f554eb8bcdb8fdb1a2775b02f6ffc25721fe8af60c3c0a"} Jan 26 17:45:04 crc kubenswrapper[4754]: I0126 17:45:04.595602 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2cf9e3e04a25356888f554eb8bcdb8fdb1a2775b02f6ffc25721fe8af60c3c0a" Jan 26 17:45:04 crc kubenswrapper[4754]: I0126 17:45:04.595327 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29490825-mpdwg" Jan 26 17:45:04 crc kubenswrapper[4754]: I0126 17:45:04.663537 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29490780-7wcg8"] Jan 26 17:45:04 crc kubenswrapper[4754]: I0126 17:45:04.672964 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29490780-7wcg8"] Jan 26 17:45:05 crc kubenswrapper[4754]: I0126 17:45:05.777482 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4f2d7a6e-760a-4cd5-b9d0-7483d4f12933" path="/var/lib/kubelet/pods/4f2d7a6e-760a-4cd5-b9d0-7483d4f12933/volumes" Jan 26 17:45:07 crc kubenswrapper[4754]: I0126 17:45:07.129332 4754 patch_prober.go:28] interesting pod/machine-config-daemon-x65wv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 26 17:45:07 crc kubenswrapper[4754]: I0126 17:45:07.129766 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 26 17:45:07 crc kubenswrapper[4754]: I0126 17:45:07.129833 4754 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" Jan 26 17:45:07 crc kubenswrapper[4754]: I0126 17:45:07.130891 4754 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"b8ddaeea1adb9fa522ab28e802b3b1b023114966ebf93bdd7941ca64dfbc9362"} pod="openshift-machine-config-operator/machine-config-daemon-x65wv" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 26 17:45:07 crc kubenswrapper[4754]: I0126 17:45:07.131001 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" containerName="machine-config-daemon" containerID="cri-o://b8ddaeea1adb9fa522ab28e802b3b1b023114966ebf93bdd7941ca64dfbc9362" gracePeriod=600 Jan 26 17:45:08 crc kubenswrapper[4754]: I0126 17:45:08.631178 4754 generic.go:334] "Generic (PLEG): container finished" podID="8c3718a4-f354-4284-92e0-fdfb45a692bd" containerID="b8ddaeea1adb9fa522ab28e802b3b1b023114966ebf93bdd7941ca64dfbc9362" exitCode=0 Jan 26 17:45:08 crc kubenswrapper[4754]: I0126 17:45:08.631246 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" event={"ID":"8c3718a4-f354-4284-92e0-fdfb45a692bd","Type":"ContainerDied","Data":"b8ddaeea1adb9fa522ab28e802b3b1b023114966ebf93bdd7941ca64dfbc9362"} Jan 26 17:45:08 crc kubenswrapper[4754]: I0126 17:45:08.632934 4754 scope.go:117] "RemoveContainer" containerID="2a52617105aca80bd6aa782680de53fad558b6014f1d373ef1592509df2dabc0" Jan 26 17:45:09 crc kubenswrapper[4754]: I0126 17:45:09.648729 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" event={"ID":"8c3718a4-f354-4284-92e0-fdfb45a692bd","Type":"ContainerStarted","Data":"7b1482a92488015f91f13a86a1aa38c65e69fa9f874f9482c4d8f0f34463d2c2"} Jan 26 17:45:27 crc kubenswrapper[4754]: I0126 17:45:27.788893 4754 scope.go:117] "RemoveContainer" containerID="b6c77d399b6d69676f8c557d43ec9baff18e7851b914517948580ae3bb687f5a" Jan 26 17:47:37 crc kubenswrapper[4754]: I0126 17:47:37.128848 4754 patch_prober.go:28] interesting pod/machine-config-daemon-x65wv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 26 17:47:37 crc kubenswrapper[4754]: I0126 17:47:37.130630 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 26 17:48:07 crc kubenswrapper[4754]: I0126 17:48:07.129519 4754 patch_prober.go:28] interesting pod/machine-config-daemon-x65wv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 26 17:48:07 crc kubenswrapper[4754]: I0126 17:48:07.130137 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 26 17:48:37 crc kubenswrapper[4754]: I0126 17:48:37.128935 4754 patch_prober.go:28] interesting pod/machine-config-daemon-x65wv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 26 17:48:37 crc kubenswrapper[4754]: I0126 17:48:37.130934 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 26 17:48:37 crc kubenswrapper[4754]: I0126 17:48:37.131066 4754 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" Jan 26 17:48:37 crc kubenswrapper[4754]: I0126 17:48:37.132410 4754 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"7b1482a92488015f91f13a86a1aa38c65e69fa9f874f9482c4d8f0f34463d2c2"} pod="openshift-machine-config-operator/machine-config-daemon-x65wv" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 26 17:48:37 crc kubenswrapper[4754]: I0126 17:48:37.132489 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" containerName="machine-config-daemon" containerID="cri-o://7b1482a92488015f91f13a86a1aa38c65e69fa9f874f9482c4d8f0f34463d2c2" gracePeriod=600 Jan 26 17:48:37 crc kubenswrapper[4754]: E0126 17:48:37.259507 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x65wv_openshift-machine-config-operator(8c3718a4-f354-4284-92e0-fdfb45a692bd)\"" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" Jan 26 17:48:37 crc kubenswrapper[4754]: I0126 17:48:37.415858 4754 generic.go:334] "Generic (PLEG): container finished" podID="8c3718a4-f354-4284-92e0-fdfb45a692bd" containerID="7b1482a92488015f91f13a86a1aa38c65e69fa9f874f9482c4d8f0f34463d2c2" exitCode=0 Jan 26 17:48:37 crc kubenswrapper[4754]: I0126 17:48:37.415909 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" event={"ID":"8c3718a4-f354-4284-92e0-fdfb45a692bd","Type":"ContainerDied","Data":"7b1482a92488015f91f13a86a1aa38c65e69fa9f874f9482c4d8f0f34463d2c2"} Jan 26 17:48:37 crc kubenswrapper[4754]: I0126 17:48:37.415953 4754 scope.go:117] "RemoveContainer" containerID="b8ddaeea1adb9fa522ab28e802b3b1b023114966ebf93bdd7941ca64dfbc9362" Jan 26 17:48:37 crc kubenswrapper[4754]: I0126 17:48:37.416703 4754 scope.go:117] "RemoveContainer" containerID="7b1482a92488015f91f13a86a1aa38c65e69fa9f874f9482c4d8f0f34463d2c2" Jan 26 17:48:37 crc kubenswrapper[4754]: E0126 17:48:37.416993 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x65wv_openshift-machine-config-operator(8c3718a4-f354-4284-92e0-fdfb45a692bd)\"" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" Jan 26 17:48:50 crc kubenswrapper[4754]: I0126 17:48:50.767320 4754 scope.go:117] "RemoveContainer" containerID="7b1482a92488015f91f13a86a1aa38c65e69fa9f874f9482c4d8f0f34463d2c2" Jan 26 17:48:50 crc kubenswrapper[4754]: E0126 17:48:50.768176 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x65wv_openshift-machine-config-operator(8c3718a4-f354-4284-92e0-fdfb45a692bd)\"" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" Jan 26 17:48:54 crc kubenswrapper[4754]: I0126 17:48:54.270840 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-bqgqj"] Jan 26 17:48:54 crc kubenswrapper[4754]: E0126 17:48:54.273558 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db7c55e8-5cf4-48d5-ab91-4795c5e52311" containerName="collect-profiles" Jan 26 17:48:54 crc kubenswrapper[4754]: I0126 17:48:54.273761 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="db7c55e8-5cf4-48d5-ab91-4795c5e52311" containerName="collect-profiles" Jan 26 17:48:54 crc kubenswrapper[4754]: I0126 17:48:54.274277 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="db7c55e8-5cf4-48d5-ab91-4795c5e52311" containerName="collect-profiles" Jan 26 17:48:54 crc kubenswrapper[4754]: I0126 17:48:54.276655 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bqgqj" Jan 26 17:48:54 crc kubenswrapper[4754]: I0126 17:48:54.280309 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-bqgqj"] Jan 26 17:48:54 crc kubenswrapper[4754]: I0126 17:48:54.404607 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ecc926d0-b740-4cf8-8bef-6b01f74a111d-catalog-content\") pod \"redhat-marketplace-bqgqj\" (UID: \"ecc926d0-b740-4cf8-8bef-6b01f74a111d\") " pod="openshift-marketplace/redhat-marketplace-bqgqj" Jan 26 17:48:54 crc kubenswrapper[4754]: I0126 17:48:54.404882 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ecc926d0-b740-4cf8-8bef-6b01f74a111d-utilities\") pod \"redhat-marketplace-bqgqj\" (UID: \"ecc926d0-b740-4cf8-8bef-6b01f74a111d\") " pod="openshift-marketplace/redhat-marketplace-bqgqj" Jan 26 17:48:54 crc kubenswrapper[4754]: I0126 17:48:54.404929 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vc8gv\" (UniqueName: \"kubernetes.io/projected/ecc926d0-b740-4cf8-8bef-6b01f74a111d-kube-api-access-vc8gv\") pod \"redhat-marketplace-bqgqj\" (UID: \"ecc926d0-b740-4cf8-8bef-6b01f74a111d\") " pod="openshift-marketplace/redhat-marketplace-bqgqj" Jan 26 17:48:54 crc kubenswrapper[4754]: I0126 17:48:54.506069 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ecc926d0-b740-4cf8-8bef-6b01f74a111d-catalog-content\") pod \"redhat-marketplace-bqgqj\" (UID: \"ecc926d0-b740-4cf8-8bef-6b01f74a111d\") " pod="openshift-marketplace/redhat-marketplace-bqgqj" Jan 26 17:48:54 crc kubenswrapper[4754]: I0126 17:48:54.506246 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ecc926d0-b740-4cf8-8bef-6b01f74a111d-utilities\") pod \"redhat-marketplace-bqgqj\" (UID: \"ecc926d0-b740-4cf8-8bef-6b01f74a111d\") " pod="openshift-marketplace/redhat-marketplace-bqgqj" Jan 26 17:48:54 crc kubenswrapper[4754]: I0126 17:48:54.506269 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vc8gv\" (UniqueName: \"kubernetes.io/projected/ecc926d0-b740-4cf8-8bef-6b01f74a111d-kube-api-access-vc8gv\") pod \"redhat-marketplace-bqgqj\" (UID: \"ecc926d0-b740-4cf8-8bef-6b01f74a111d\") " pod="openshift-marketplace/redhat-marketplace-bqgqj" Jan 26 17:48:54 crc kubenswrapper[4754]: I0126 17:48:54.506813 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ecc926d0-b740-4cf8-8bef-6b01f74a111d-catalog-content\") pod \"redhat-marketplace-bqgqj\" (UID: \"ecc926d0-b740-4cf8-8bef-6b01f74a111d\") " pod="openshift-marketplace/redhat-marketplace-bqgqj" Jan 26 17:48:54 crc kubenswrapper[4754]: I0126 17:48:54.506819 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ecc926d0-b740-4cf8-8bef-6b01f74a111d-utilities\") pod \"redhat-marketplace-bqgqj\" (UID: \"ecc926d0-b740-4cf8-8bef-6b01f74a111d\") " pod="openshift-marketplace/redhat-marketplace-bqgqj" Jan 26 17:48:54 crc kubenswrapper[4754]: I0126 17:48:54.532547 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vc8gv\" (UniqueName: \"kubernetes.io/projected/ecc926d0-b740-4cf8-8bef-6b01f74a111d-kube-api-access-vc8gv\") pod \"redhat-marketplace-bqgqj\" (UID: \"ecc926d0-b740-4cf8-8bef-6b01f74a111d\") " pod="openshift-marketplace/redhat-marketplace-bqgqj" Jan 26 17:48:54 crc kubenswrapper[4754]: I0126 17:48:54.600844 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bqgqj" Jan 26 17:48:55 crc kubenswrapper[4754]: I0126 17:48:55.083743 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-bqgqj"] Jan 26 17:48:55 crc kubenswrapper[4754]: W0126 17:48:55.088496 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podecc926d0_b740_4cf8_8bef_6b01f74a111d.slice/crio-a4805e4ba438004a02ec1f508e23147cb5bc168a36e329fba6e3cecb618299d3 WatchSource:0}: Error finding container a4805e4ba438004a02ec1f508e23147cb5bc168a36e329fba6e3cecb618299d3: Status 404 returned error can't find the container with id a4805e4ba438004a02ec1f508e23147cb5bc168a36e329fba6e3cecb618299d3 Jan 26 17:48:55 crc kubenswrapper[4754]: I0126 17:48:55.562781 4754 generic.go:334] "Generic (PLEG): container finished" podID="ecc926d0-b740-4cf8-8bef-6b01f74a111d" containerID="af1c92d8e0f25f6a346556ef59cc39e9b160af03908d759b720a0589f52544a2" exitCode=0 Jan 26 17:48:55 crc kubenswrapper[4754]: I0126 17:48:55.562865 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bqgqj" event={"ID":"ecc926d0-b740-4cf8-8bef-6b01f74a111d","Type":"ContainerDied","Data":"af1c92d8e0f25f6a346556ef59cc39e9b160af03908d759b720a0589f52544a2"} Jan 26 17:48:55 crc kubenswrapper[4754]: I0126 17:48:55.563040 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bqgqj" event={"ID":"ecc926d0-b740-4cf8-8bef-6b01f74a111d","Type":"ContainerStarted","Data":"a4805e4ba438004a02ec1f508e23147cb5bc168a36e329fba6e3cecb618299d3"} Jan 26 17:48:55 crc kubenswrapper[4754]: I0126 17:48:55.564641 4754 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 26 17:48:57 crc kubenswrapper[4754]: I0126 17:48:57.584862 4754 generic.go:334] "Generic (PLEG): container finished" podID="ecc926d0-b740-4cf8-8bef-6b01f74a111d" containerID="329e198f896fe22d3fdf87a94103f8eb31f20a2298557408164b62bf27f7e148" exitCode=0 Jan 26 17:48:57 crc kubenswrapper[4754]: I0126 17:48:57.584950 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bqgqj" event={"ID":"ecc926d0-b740-4cf8-8bef-6b01f74a111d","Type":"ContainerDied","Data":"329e198f896fe22d3fdf87a94103f8eb31f20a2298557408164b62bf27f7e148"} Jan 26 17:48:58 crc kubenswrapper[4754]: I0126 17:48:58.594490 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bqgqj" event={"ID":"ecc926d0-b740-4cf8-8bef-6b01f74a111d","Type":"ContainerStarted","Data":"72db35c0308711a101674a53f604ff9bed3fab50833ecea4ebd50d143005083d"} Jan 26 17:48:58 crc kubenswrapper[4754]: I0126 17:48:58.616095 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-bqgqj" podStartSLOduration=2.136048538 podStartE2EDuration="4.61607667s" podCreationTimestamp="2026-01-26 17:48:54 +0000 UTC" firstStartedPulling="2026-01-26 17:48:55.564402734 +0000 UTC m=+3702.088583158" lastFinishedPulling="2026-01-26 17:48:58.044430856 +0000 UTC m=+3704.568611290" observedRunningTime="2026-01-26 17:48:58.610106916 +0000 UTC m=+3705.134287370" watchObservedRunningTime="2026-01-26 17:48:58.61607667 +0000 UTC m=+3705.140257104" Jan 26 17:49:03 crc kubenswrapper[4754]: I0126 17:49:03.773499 4754 scope.go:117] "RemoveContainer" containerID="7b1482a92488015f91f13a86a1aa38c65e69fa9f874f9482c4d8f0f34463d2c2" Jan 26 17:49:03 crc kubenswrapper[4754]: E0126 17:49:03.774320 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x65wv_openshift-machine-config-operator(8c3718a4-f354-4284-92e0-fdfb45a692bd)\"" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" Jan 26 17:49:04 crc kubenswrapper[4754]: I0126 17:49:04.601420 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-bqgqj" Jan 26 17:49:04 crc kubenswrapper[4754]: I0126 17:49:04.601475 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-bqgqj" Jan 26 17:49:04 crc kubenswrapper[4754]: I0126 17:49:04.666066 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-bqgqj" Jan 26 17:49:04 crc kubenswrapper[4754]: I0126 17:49:04.710333 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-bqgqj" Jan 26 17:49:04 crc kubenswrapper[4754]: I0126 17:49:04.903030 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-bqgqj"] Jan 26 17:49:06 crc kubenswrapper[4754]: I0126 17:49:06.662258 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-bqgqj" podUID="ecc926d0-b740-4cf8-8bef-6b01f74a111d" containerName="registry-server" containerID="cri-o://72db35c0308711a101674a53f604ff9bed3fab50833ecea4ebd50d143005083d" gracePeriod=2 Jan 26 17:49:07 crc kubenswrapper[4754]: I0126 17:49:07.198020 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bqgqj" Jan 26 17:49:07 crc kubenswrapper[4754]: I0126 17:49:07.339536 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vc8gv\" (UniqueName: \"kubernetes.io/projected/ecc926d0-b740-4cf8-8bef-6b01f74a111d-kube-api-access-vc8gv\") pod \"ecc926d0-b740-4cf8-8bef-6b01f74a111d\" (UID: \"ecc926d0-b740-4cf8-8bef-6b01f74a111d\") " Jan 26 17:49:07 crc kubenswrapper[4754]: I0126 17:49:07.340097 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ecc926d0-b740-4cf8-8bef-6b01f74a111d-catalog-content\") pod \"ecc926d0-b740-4cf8-8bef-6b01f74a111d\" (UID: \"ecc926d0-b740-4cf8-8bef-6b01f74a111d\") " Jan 26 17:49:07 crc kubenswrapper[4754]: I0126 17:49:07.340219 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ecc926d0-b740-4cf8-8bef-6b01f74a111d-utilities\") pod \"ecc926d0-b740-4cf8-8bef-6b01f74a111d\" (UID: \"ecc926d0-b740-4cf8-8bef-6b01f74a111d\") " Jan 26 17:49:07 crc kubenswrapper[4754]: I0126 17:49:07.340871 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ecc926d0-b740-4cf8-8bef-6b01f74a111d-utilities" (OuterVolumeSpecName: "utilities") pod "ecc926d0-b740-4cf8-8bef-6b01f74a111d" (UID: "ecc926d0-b740-4cf8-8bef-6b01f74a111d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:49:07 crc kubenswrapper[4754]: I0126 17:49:07.345080 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ecc926d0-b740-4cf8-8bef-6b01f74a111d-kube-api-access-vc8gv" (OuterVolumeSpecName: "kube-api-access-vc8gv") pod "ecc926d0-b740-4cf8-8bef-6b01f74a111d" (UID: "ecc926d0-b740-4cf8-8bef-6b01f74a111d"). InnerVolumeSpecName "kube-api-access-vc8gv". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:49:07 crc kubenswrapper[4754]: I0126 17:49:07.366075 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ecc926d0-b740-4cf8-8bef-6b01f74a111d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ecc926d0-b740-4cf8-8bef-6b01f74a111d" (UID: "ecc926d0-b740-4cf8-8bef-6b01f74a111d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:49:07 crc kubenswrapper[4754]: I0126 17:49:07.442873 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vc8gv\" (UniqueName: \"kubernetes.io/projected/ecc926d0-b740-4cf8-8bef-6b01f74a111d-kube-api-access-vc8gv\") on node \"crc\" DevicePath \"\"" Jan 26 17:49:07 crc kubenswrapper[4754]: I0126 17:49:07.442903 4754 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ecc926d0-b740-4cf8-8bef-6b01f74a111d-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 26 17:49:07 crc kubenswrapper[4754]: I0126 17:49:07.442913 4754 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ecc926d0-b740-4cf8-8bef-6b01f74a111d-utilities\") on node \"crc\" DevicePath \"\"" Jan 26 17:49:07 crc kubenswrapper[4754]: I0126 17:49:07.672141 4754 generic.go:334] "Generic (PLEG): container finished" podID="ecc926d0-b740-4cf8-8bef-6b01f74a111d" containerID="72db35c0308711a101674a53f604ff9bed3fab50833ecea4ebd50d143005083d" exitCode=0 Jan 26 17:49:07 crc kubenswrapper[4754]: I0126 17:49:07.672188 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bqgqj" event={"ID":"ecc926d0-b740-4cf8-8bef-6b01f74a111d","Type":"ContainerDied","Data":"72db35c0308711a101674a53f604ff9bed3fab50833ecea4ebd50d143005083d"} Jan 26 17:49:07 crc kubenswrapper[4754]: I0126 17:49:07.672261 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bqgqj" event={"ID":"ecc926d0-b740-4cf8-8bef-6b01f74a111d","Type":"ContainerDied","Data":"a4805e4ba438004a02ec1f508e23147cb5bc168a36e329fba6e3cecb618299d3"} Jan 26 17:49:07 crc kubenswrapper[4754]: I0126 17:49:07.672308 4754 scope.go:117] "RemoveContainer" containerID="72db35c0308711a101674a53f604ff9bed3fab50833ecea4ebd50d143005083d" Jan 26 17:49:07 crc kubenswrapper[4754]: I0126 17:49:07.672214 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bqgqj" Jan 26 17:49:07 crc kubenswrapper[4754]: I0126 17:49:07.691740 4754 scope.go:117] "RemoveContainer" containerID="329e198f896fe22d3fdf87a94103f8eb31f20a2298557408164b62bf27f7e148" Jan 26 17:49:07 crc kubenswrapper[4754]: I0126 17:49:07.721051 4754 scope.go:117] "RemoveContainer" containerID="af1c92d8e0f25f6a346556ef59cc39e9b160af03908d759b720a0589f52544a2" Jan 26 17:49:07 crc kubenswrapper[4754]: I0126 17:49:07.728161 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-bqgqj"] Jan 26 17:49:07 crc kubenswrapper[4754]: I0126 17:49:07.735628 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-bqgqj"] Jan 26 17:49:07 crc kubenswrapper[4754]: I0126 17:49:07.767408 4754 scope.go:117] "RemoveContainer" containerID="72db35c0308711a101674a53f604ff9bed3fab50833ecea4ebd50d143005083d" Jan 26 17:49:07 crc kubenswrapper[4754]: E0126 17:49:07.768000 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"72db35c0308711a101674a53f604ff9bed3fab50833ecea4ebd50d143005083d\": container with ID starting with 72db35c0308711a101674a53f604ff9bed3fab50833ecea4ebd50d143005083d not found: ID does not exist" containerID="72db35c0308711a101674a53f604ff9bed3fab50833ecea4ebd50d143005083d" Jan 26 17:49:07 crc kubenswrapper[4754]: I0126 17:49:07.768045 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"72db35c0308711a101674a53f604ff9bed3fab50833ecea4ebd50d143005083d"} err="failed to get container status \"72db35c0308711a101674a53f604ff9bed3fab50833ecea4ebd50d143005083d\": rpc error: code = NotFound desc = could not find container \"72db35c0308711a101674a53f604ff9bed3fab50833ecea4ebd50d143005083d\": container with ID starting with 72db35c0308711a101674a53f604ff9bed3fab50833ecea4ebd50d143005083d not found: ID does not exist" Jan 26 17:49:07 crc kubenswrapper[4754]: I0126 17:49:07.768074 4754 scope.go:117] "RemoveContainer" containerID="329e198f896fe22d3fdf87a94103f8eb31f20a2298557408164b62bf27f7e148" Jan 26 17:49:07 crc kubenswrapper[4754]: E0126 17:49:07.768579 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"329e198f896fe22d3fdf87a94103f8eb31f20a2298557408164b62bf27f7e148\": container with ID starting with 329e198f896fe22d3fdf87a94103f8eb31f20a2298557408164b62bf27f7e148 not found: ID does not exist" containerID="329e198f896fe22d3fdf87a94103f8eb31f20a2298557408164b62bf27f7e148" Jan 26 17:49:07 crc kubenswrapper[4754]: I0126 17:49:07.768611 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"329e198f896fe22d3fdf87a94103f8eb31f20a2298557408164b62bf27f7e148"} err="failed to get container status \"329e198f896fe22d3fdf87a94103f8eb31f20a2298557408164b62bf27f7e148\": rpc error: code = NotFound desc = could not find container \"329e198f896fe22d3fdf87a94103f8eb31f20a2298557408164b62bf27f7e148\": container with ID starting with 329e198f896fe22d3fdf87a94103f8eb31f20a2298557408164b62bf27f7e148 not found: ID does not exist" Jan 26 17:49:07 crc kubenswrapper[4754]: I0126 17:49:07.768626 4754 scope.go:117] "RemoveContainer" containerID="af1c92d8e0f25f6a346556ef59cc39e9b160af03908d759b720a0589f52544a2" Jan 26 17:49:07 crc kubenswrapper[4754]: E0126 17:49:07.768971 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"af1c92d8e0f25f6a346556ef59cc39e9b160af03908d759b720a0589f52544a2\": container with ID starting with af1c92d8e0f25f6a346556ef59cc39e9b160af03908d759b720a0589f52544a2 not found: ID does not exist" containerID="af1c92d8e0f25f6a346556ef59cc39e9b160af03908d759b720a0589f52544a2" Jan 26 17:49:07 crc kubenswrapper[4754]: I0126 17:49:07.768999 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"af1c92d8e0f25f6a346556ef59cc39e9b160af03908d759b720a0589f52544a2"} err="failed to get container status \"af1c92d8e0f25f6a346556ef59cc39e9b160af03908d759b720a0589f52544a2\": rpc error: code = NotFound desc = could not find container \"af1c92d8e0f25f6a346556ef59cc39e9b160af03908d759b720a0589f52544a2\": container with ID starting with af1c92d8e0f25f6a346556ef59cc39e9b160af03908d759b720a0589f52544a2 not found: ID does not exist" Jan 26 17:49:07 crc kubenswrapper[4754]: I0126 17:49:07.779911 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ecc926d0-b740-4cf8-8bef-6b01f74a111d" path="/var/lib/kubelet/pods/ecc926d0-b740-4cf8-8bef-6b01f74a111d/volumes" Jan 26 17:49:14 crc kubenswrapper[4754]: I0126 17:49:14.767801 4754 scope.go:117] "RemoveContainer" containerID="7b1482a92488015f91f13a86a1aa38c65e69fa9f874f9482c4d8f0f34463d2c2" Jan 26 17:49:14 crc kubenswrapper[4754]: E0126 17:49:14.768545 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x65wv_openshift-machine-config-operator(8c3718a4-f354-4284-92e0-fdfb45a692bd)\"" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" Jan 26 17:49:26 crc kubenswrapper[4754]: I0126 17:49:26.190477 4754 generic.go:334] "Generic (PLEG): container finished" podID="6f413c96-a0cb-4d11-9def-df8c3218db98" containerID="3426d5689db28b717decc25263f3ffdc456630c56b24555f60c60ddab335bf44" exitCode=0 Jan 26 17:49:26 crc kubenswrapper[4754]: I0126 17:49:26.190605 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"6f413c96-a0cb-4d11-9def-df8c3218db98","Type":"ContainerDied","Data":"3426d5689db28b717decc25263f3ffdc456630c56b24555f60c60ddab335bf44"} Jan 26 17:49:26 crc kubenswrapper[4754]: I0126 17:49:26.767287 4754 scope.go:117] "RemoveContainer" containerID="7b1482a92488015f91f13a86a1aa38c65e69fa9f874f9482c4d8f0f34463d2c2" Jan 26 17:49:26 crc kubenswrapper[4754]: E0126 17:49:26.768004 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x65wv_openshift-machine-config-operator(8c3718a4-f354-4284-92e0-fdfb45a692bd)\"" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" Jan 26 17:49:27 crc kubenswrapper[4754]: I0126 17:49:27.514136 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Jan 26 17:49:27 crc kubenswrapper[4754]: I0126 17:49:27.620840 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"6f413c96-a0cb-4d11-9def-df8c3218db98\" (UID: \"6f413c96-a0cb-4d11-9def-df8c3218db98\") " Jan 26 17:49:27 crc kubenswrapper[4754]: I0126 17:49:27.620929 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/6f413c96-a0cb-4d11-9def-df8c3218db98-openstack-config\") pod \"6f413c96-a0cb-4d11-9def-df8c3218db98\" (UID: \"6f413c96-a0cb-4d11-9def-df8c3218db98\") " Jan 26 17:49:27 crc kubenswrapper[4754]: I0126 17:49:27.620957 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p72sf\" (UniqueName: \"kubernetes.io/projected/6f413c96-a0cb-4d11-9def-df8c3218db98-kube-api-access-p72sf\") pod \"6f413c96-a0cb-4d11-9def-df8c3218db98\" (UID: \"6f413c96-a0cb-4d11-9def-df8c3218db98\") " Jan 26 17:49:27 crc kubenswrapper[4754]: I0126 17:49:27.621030 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6f413c96-a0cb-4d11-9def-df8c3218db98-config-data\") pod \"6f413c96-a0cb-4d11-9def-df8c3218db98\" (UID: \"6f413c96-a0cb-4d11-9def-df8c3218db98\") " Jan 26 17:49:27 crc kubenswrapper[4754]: I0126 17:49:27.621114 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/6f413c96-a0cb-4d11-9def-df8c3218db98-ca-certs\") pod \"6f413c96-a0cb-4d11-9def-df8c3218db98\" (UID: \"6f413c96-a0cb-4d11-9def-df8c3218db98\") " Jan 26 17:49:27 crc kubenswrapper[4754]: I0126 17:49:27.621161 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/6f413c96-a0cb-4d11-9def-df8c3218db98-test-operator-ephemeral-temporary\") pod \"6f413c96-a0cb-4d11-9def-df8c3218db98\" (UID: \"6f413c96-a0cb-4d11-9def-df8c3218db98\") " Jan 26 17:49:27 crc kubenswrapper[4754]: I0126 17:49:27.621186 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/6f413c96-a0cb-4d11-9def-df8c3218db98-openstack-config-secret\") pod \"6f413c96-a0cb-4d11-9def-df8c3218db98\" (UID: \"6f413c96-a0cb-4d11-9def-df8c3218db98\") " Jan 26 17:49:27 crc kubenswrapper[4754]: I0126 17:49:27.621279 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/6f413c96-a0cb-4d11-9def-df8c3218db98-test-operator-ephemeral-workdir\") pod \"6f413c96-a0cb-4d11-9def-df8c3218db98\" (UID: \"6f413c96-a0cb-4d11-9def-df8c3218db98\") " Jan 26 17:49:27 crc kubenswrapper[4754]: I0126 17:49:27.621316 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6f413c96-a0cb-4d11-9def-df8c3218db98-ssh-key\") pod \"6f413c96-a0cb-4d11-9def-df8c3218db98\" (UID: \"6f413c96-a0cb-4d11-9def-df8c3218db98\") " Jan 26 17:49:27 crc kubenswrapper[4754]: I0126 17:49:27.622391 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6f413c96-a0cb-4d11-9def-df8c3218db98-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "6f413c96-a0cb-4d11-9def-df8c3218db98" (UID: "6f413c96-a0cb-4d11-9def-df8c3218db98"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:49:27 crc kubenswrapper[4754]: I0126 17:49:27.623241 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6f413c96-a0cb-4d11-9def-df8c3218db98-config-data" (OuterVolumeSpecName: "config-data") pod "6f413c96-a0cb-4d11-9def-df8c3218db98" (UID: "6f413c96-a0cb-4d11-9def-df8c3218db98"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:49:27 crc kubenswrapper[4754]: I0126 17:49:27.626595 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6f413c96-a0cb-4d11-9def-df8c3218db98-kube-api-access-p72sf" (OuterVolumeSpecName: "kube-api-access-p72sf") pod "6f413c96-a0cb-4d11-9def-df8c3218db98" (UID: "6f413c96-a0cb-4d11-9def-df8c3218db98"). InnerVolumeSpecName "kube-api-access-p72sf". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:49:27 crc kubenswrapper[4754]: I0126 17:49:27.626781 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage03-crc" (OuterVolumeSpecName: "test-operator-logs") pod "6f413c96-a0cb-4d11-9def-df8c3218db98" (UID: "6f413c96-a0cb-4d11-9def-df8c3218db98"). InnerVolumeSpecName "local-storage03-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Jan 26 17:49:27 crc kubenswrapper[4754]: I0126 17:49:27.627747 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6f413c96-a0cb-4d11-9def-df8c3218db98-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "6f413c96-a0cb-4d11-9def-df8c3218db98" (UID: "6f413c96-a0cb-4d11-9def-df8c3218db98"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:49:27 crc kubenswrapper[4754]: I0126 17:49:27.650047 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6f413c96-a0cb-4d11-9def-df8c3218db98-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "6f413c96-a0cb-4d11-9def-df8c3218db98" (UID: "6f413c96-a0cb-4d11-9def-df8c3218db98"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:49:27 crc kubenswrapper[4754]: I0126 17:49:27.651707 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6f413c96-a0cb-4d11-9def-df8c3218db98-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "6f413c96-a0cb-4d11-9def-df8c3218db98" (UID: "6f413c96-a0cb-4d11-9def-df8c3218db98"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:49:27 crc kubenswrapper[4754]: I0126 17:49:27.652890 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6f413c96-a0cb-4d11-9def-df8c3218db98-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "6f413c96-a0cb-4d11-9def-df8c3218db98" (UID: "6f413c96-a0cb-4d11-9def-df8c3218db98"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:49:27 crc kubenswrapper[4754]: I0126 17:49:27.667612 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6f413c96-a0cb-4d11-9def-df8c3218db98-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "6f413c96-a0cb-4d11-9def-df8c3218db98" (UID: "6f413c96-a0cb-4d11-9def-df8c3218db98"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:49:27 crc kubenswrapper[4754]: I0126 17:49:27.722989 4754 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" " Jan 26 17:49:27 crc kubenswrapper[4754]: I0126 17:49:27.723034 4754 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/6f413c96-a0cb-4d11-9def-df8c3218db98-openstack-config\") on node \"crc\" DevicePath \"\"" Jan 26 17:49:27 crc kubenswrapper[4754]: I0126 17:49:27.723049 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p72sf\" (UniqueName: \"kubernetes.io/projected/6f413c96-a0cb-4d11-9def-df8c3218db98-kube-api-access-p72sf\") on node \"crc\" DevicePath \"\"" Jan 26 17:49:27 crc kubenswrapper[4754]: I0126 17:49:27.723058 4754 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6f413c96-a0cb-4d11-9def-df8c3218db98-config-data\") on node \"crc\" DevicePath \"\"" Jan 26 17:49:27 crc kubenswrapper[4754]: I0126 17:49:27.723066 4754 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/6f413c96-a0cb-4d11-9def-df8c3218db98-ca-certs\") on node \"crc\" DevicePath \"\"" Jan 26 17:49:27 crc kubenswrapper[4754]: I0126 17:49:27.723077 4754 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/6f413c96-a0cb-4d11-9def-df8c3218db98-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Jan 26 17:49:27 crc kubenswrapper[4754]: I0126 17:49:27.723085 4754 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/6f413c96-a0cb-4d11-9def-df8c3218db98-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Jan 26 17:49:27 crc kubenswrapper[4754]: I0126 17:49:27.723094 4754 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/6f413c96-a0cb-4d11-9def-df8c3218db98-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Jan 26 17:49:27 crc kubenswrapper[4754]: I0126 17:49:27.723101 4754 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6f413c96-a0cb-4d11-9def-df8c3218db98-ssh-key\") on node \"crc\" DevicePath \"\"" Jan 26 17:49:27 crc kubenswrapper[4754]: I0126 17:49:27.740629 4754 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage03-crc" (UniqueName: "kubernetes.io/local-volume/local-storage03-crc") on node "crc" Jan 26 17:49:27 crc kubenswrapper[4754]: I0126 17:49:27.825549 4754 reconciler_common.go:293] "Volume detached for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" DevicePath \"\"" Jan 26 17:49:28 crc kubenswrapper[4754]: I0126 17:49:28.213001 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"6f413c96-a0cb-4d11-9def-df8c3218db98","Type":"ContainerDied","Data":"d60a26d5912f2c0f9acdb318d8eec23b93eea368db920c2db06859102ea6d69a"} Jan 26 17:49:28 crc kubenswrapper[4754]: I0126 17:49:28.213045 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d60a26d5912f2c0f9acdb318d8eec23b93eea368db920c2db06859102ea6d69a" Jan 26 17:49:28 crc kubenswrapper[4754]: I0126 17:49:28.213088 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Jan 26 17:49:34 crc kubenswrapper[4754]: I0126 17:49:34.550352 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Jan 26 17:49:34 crc kubenswrapper[4754]: E0126 17:49:34.551294 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ecc926d0-b740-4cf8-8bef-6b01f74a111d" containerName="registry-server" Jan 26 17:49:34 crc kubenswrapper[4754]: I0126 17:49:34.551313 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="ecc926d0-b740-4cf8-8bef-6b01f74a111d" containerName="registry-server" Jan 26 17:49:34 crc kubenswrapper[4754]: E0126 17:49:34.551338 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6f413c96-a0cb-4d11-9def-df8c3218db98" containerName="tempest-tests-tempest-tests-runner" Jan 26 17:49:34 crc kubenswrapper[4754]: I0126 17:49:34.551348 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="6f413c96-a0cb-4d11-9def-df8c3218db98" containerName="tempest-tests-tempest-tests-runner" Jan 26 17:49:34 crc kubenswrapper[4754]: E0126 17:49:34.551362 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ecc926d0-b740-4cf8-8bef-6b01f74a111d" containerName="extract-content" Jan 26 17:49:34 crc kubenswrapper[4754]: I0126 17:49:34.551370 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="ecc926d0-b740-4cf8-8bef-6b01f74a111d" containerName="extract-content" Jan 26 17:49:34 crc kubenswrapper[4754]: E0126 17:49:34.551392 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ecc926d0-b740-4cf8-8bef-6b01f74a111d" containerName="extract-utilities" Jan 26 17:49:34 crc kubenswrapper[4754]: I0126 17:49:34.551401 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="ecc926d0-b740-4cf8-8bef-6b01f74a111d" containerName="extract-utilities" Jan 26 17:49:34 crc kubenswrapper[4754]: I0126 17:49:34.551610 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="6f413c96-a0cb-4d11-9def-df8c3218db98" containerName="tempest-tests-tempest-tests-runner" Jan 26 17:49:34 crc kubenswrapper[4754]: I0126 17:49:34.551649 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="ecc926d0-b740-4cf8-8bef-6b01f74a111d" containerName="registry-server" Jan 26 17:49:34 crc kubenswrapper[4754]: I0126 17:49:34.552371 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Jan 26 17:49:34 crc kubenswrapper[4754]: I0126 17:49:34.561017 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-hm66p" Jan 26 17:49:34 crc kubenswrapper[4754]: I0126 17:49:34.567734 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Jan 26 17:49:34 crc kubenswrapper[4754]: I0126 17:49:34.746462 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"b323e46e-1ad6-41a9-9e7f-ad13e473f49e\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Jan 26 17:49:34 crc kubenswrapper[4754]: I0126 17:49:34.746771 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s8lqp\" (UniqueName: \"kubernetes.io/projected/b323e46e-1ad6-41a9-9e7f-ad13e473f49e-kube-api-access-s8lqp\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"b323e46e-1ad6-41a9-9e7f-ad13e473f49e\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Jan 26 17:49:34 crc kubenswrapper[4754]: I0126 17:49:34.849336 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"b323e46e-1ad6-41a9-9e7f-ad13e473f49e\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Jan 26 17:49:34 crc kubenswrapper[4754]: I0126 17:49:34.849442 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s8lqp\" (UniqueName: \"kubernetes.io/projected/b323e46e-1ad6-41a9-9e7f-ad13e473f49e-kube-api-access-s8lqp\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"b323e46e-1ad6-41a9-9e7f-ad13e473f49e\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Jan 26 17:49:34 crc kubenswrapper[4754]: I0126 17:49:34.850012 4754 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"b323e46e-1ad6-41a9-9e7f-ad13e473f49e\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Jan 26 17:49:34 crc kubenswrapper[4754]: I0126 17:49:34.874796 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s8lqp\" (UniqueName: \"kubernetes.io/projected/b323e46e-1ad6-41a9-9e7f-ad13e473f49e-kube-api-access-s8lqp\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"b323e46e-1ad6-41a9-9e7f-ad13e473f49e\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Jan 26 17:49:34 crc kubenswrapper[4754]: I0126 17:49:34.879604 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"b323e46e-1ad6-41a9-9e7f-ad13e473f49e\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Jan 26 17:49:35 crc kubenswrapper[4754]: I0126 17:49:35.177512 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Jan 26 17:49:35 crc kubenswrapper[4754]: I0126 17:49:35.612718 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Jan 26 17:49:36 crc kubenswrapper[4754]: I0126 17:49:36.276902 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"b323e46e-1ad6-41a9-9e7f-ad13e473f49e","Type":"ContainerStarted","Data":"43dd5a9718a67fcfacdc8e71eb1fe53e7a8b3a43a4538d42d9083f2b66be6599"} Jan 26 17:49:38 crc kubenswrapper[4754]: I0126 17:49:38.296072 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"b323e46e-1ad6-41a9-9e7f-ad13e473f49e","Type":"ContainerStarted","Data":"f2b1713504af3da69b4be6a4e5a4e69378ba8e0956d27c78b7df835afc3ecaa7"} Jan 26 17:49:38 crc kubenswrapper[4754]: I0126 17:49:38.313454 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" podStartSLOduration=2.571443512 podStartE2EDuration="4.313435996s" podCreationTimestamp="2026-01-26 17:49:34 +0000 UTC" firstStartedPulling="2026-01-26 17:49:35.625859269 +0000 UTC m=+3742.150039703" lastFinishedPulling="2026-01-26 17:49:37.367851753 +0000 UTC m=+3743.892032187" observedRunningTime="2026-01-26 17:49:38.30958575 +0000 UTC m=+3744.833766184" watchObservedRunningTime="2026-01-26 17:49:38.313435996 +0000 UTC m=+3744.837616430" Jan 26 17:49:38 crc kubenswrapper[4754]: I0126 17:49:38.767196 4754 scope.go:117] "RemoveContainer" containerID="7b1482a92488015f91f13a86a1aa38c65e69fa9f874f9482c4d8f0f34463d2c2" Jan 26 17:49:38 crc kubenswrapper[4754]: E0126 17:49:38.767460 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x65wv_openshift-machine-config-operator(8c3718a4-f354-4284-92e0-fdfb45a692bd)\"" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" Jan 26 17:49:51 crc kubenswrapper[4754]: I0126 17:49:51.768083 4754 scope.go:117] "RemoveContainer" containerID="7b1482a92488015f91f13a86a1aa38c65e69fa9f874f9482c4d8f0f34463d2c2" Jan 26 17:49:51 crc kubenswrapper[4754]: E0126 17:49:51.769095 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x65wv_openshift-machine-config-operator(8c3718a4-f354-4284-92e0-fdfb45a692bd)\"" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" Jan 26 17:49:59 crc kubenswrapper[4754]: I0126 17:49:59.505524 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-ldztm/must-gather-rxq7n"] Jan 26 17:49:59 crc kubenswrapper[4754]: I0126 17:49:59.508300 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ldztm/must-gather-rxq7n" Jan 26 17:49:59 crc kubenswrapper[4754]: I0126 17:49:59.516487 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-ldztm"/"openshift-service-ca.crt" Jan 26 17:49:59 crc kubenswrapper[4754]: I0126 17:49:59.533948 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-ldztm/must-gather-rxq7n"] Jan 26 17:49:59 crc kubenswrapper[4754]: I0126 17:49:59.542309 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-ldztm"/"kube-root-ca.crt" Jan 26 17:49:59 crc kubenswrapper[4754]: I0126 17:49:59.618876 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/8fd36f5c-ba61-4968-80fd-d82a19020b87-must-gather-output\") pod \"must-gather-rxq7n\" (UID: \"8fd36f5c-ba61-4968-80fd-d82a19020b87\") " pod="openshift-must-gather-ldztm/must-gather-rxq7n" Jan 26 17:49:59 crc kubenswrapper[4754]: I0126 17:49:59.619195 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5d4rs\" (UniqueName: \"kubernetes.io/projected/8fd36f5c-ba61-4968-80fd-d82a19020b87-kube-api-access-5d4rs\") pod \"must-gather-rxq7n\" (UID: \"8fd36f5c-ba61-4968-80fd-d82a19020b87\") " pod="openshift-must-gather-ldztm/must-gather-rxq7n" Jan 26 17:49:59 crc kubenswrapper[4754]: I0126 17:49:59.724372 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5d4rs\" (UniqueName: \"kubernetes.io/projected/8fd36f5c-ba61-4968-80fd-d82a19020b87-kube-api-access-5d4rs\") pod \"must-gather-rxq7n\" (UID: \"8fd36f5c-ba61-4968-80fd-d82a19020b87\") " pod="openshift-must-gather-ldztm/must-gather-rxq7n" Jan 26 17:49:59 crc kubenswrapper[4754]: I0126 17:49:59.725308 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/8fd36f5c-ba61-4968-80fd-d82a19020b87-must-gather-output\") pod \"must-gather-rxq7n\" (UID: \"8fd36f5c-ba61-4968-80fd-d82a19020b87\") " pod="openshift-must-gather-ldztm/must-gather-rxq7n" Jan 26 17:49:59 crc kubenswrapper[4754]: I0126 17:49:59.727908 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/8fd36f5c-ba61-4968-80fd-d82a19020b87-must-gather-output\") pod \"must-gather-rxq7n\" (UID: \"8fd36f5c-ba61-4968-80fd-d82a19020b87\") " pod="openshift-must-gather-ldztm/must-gather-rxq7n" Jan 26 17:49:59 crc kubenswrapper[4754]: I0126 17:49:59.769153 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5d4rs\" (UniqueName: \"kubernetes.io/projected/8fd36f5c-ba61-4968-80fd-d82a19020b87-kube-api-access-5d4rs\") pod \"must-gather-rxq7n\" (UID: \"8fd36f5c-ba61-4968-80fd-d82a19020b87\") " pod="openshift-must-gather-ldztm/must-gather-rxq7n" Jan 26 17:49:59 crc kubenswrapper[4754]: I0126 17:49:59.844088 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ldztm/must-gather-rxq7n" Jan 26 17:50:00 crc kubenswrapper[4754]: I0126 17:50:00.316036 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-ldztm/must-gather-rxq7n"] Jan 26 17:50:00 crc kubenswrapper[4754]: W0126 17:50:00.725418 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8fd36f5c_ba61_4968_80fd_d82a19020b87.slice/crio-b7b075239bae4c0742cafec4e98677cea3792f9200872f24b69376f636d1f9bf WatchSource:0}: Error finding container b7b075239bae4c0742cafec4e98677cea3792f9200872f24b69376f636d1f9bf: Status 404 returned error can't find the container with id b7b075239bae4c0742cafec4e98677cea3792f9200872f24b69376f636d1f9bf Jan 26 17:50:01 crc kubenswrapper[4754]: I0126 17:50:01.494821 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-ldztm/must-gather-rxq7n" event={"ID":"8fd36f5c-ba61-4968-80fd-d82a19020b87","Type":"ContainerStarted","Data":"b7b075239bae4c0742cafec4e98677cea3792f9200872f24b69376f636d1f9bf"} Jan 26 17:50:02 crc kubenswrapper[4754]: I0126 17:50:02.768759 4754 scope.go:117] "RemoveContainer" containerID="7b1482a92488015f91f13a86a1aa38c65e69fa9f874f9482c4d8f0f34463d2c2" Jan 26 17:50:02 crc kubenswrapper[4754]: E0126 17:50:02.769792 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x65wv_openshift-machine-config-operator(8c3718a4-f354-4284-92e0-fdfb45a692bd)\"" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" Jan 26 17:50:07 crc kubenswrapper[4754]: I0126 17:50:07.551577 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-ldztm/must-gather-rxq7n" event={"ID":"8fd36f5c-ba61-4968-80fd-d82a19020b87","Type":"ContainerStarted","Data":"cfa489b848ca1551df79275c29bbd3d9ff8fa27da4b862e0bc1d45b7822fdaee"} Jan 26 17:50:07 crc kubenswrapper[4754]: I0126 17:50:07.552162 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-ldztm/must-gather-rxq7n" event={"ID":"8fd36f5c-ba61-4968-80fd-d82a19020b87","Type":"ContainerStarted","Data":"16ff16dad3135c7cfd616f48c6518a11f6b888cdc8cb783551fbefa71546d0da"} Jan 26 17:50:07 crc kubenswrapper[4754]: I0126 17:50:07.576843 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-ldztm/must-gather-rxq7n" podStartSLOduration=2.58506691 podStartE2EDuration="8.576819672s" podCreationTimestamp="2026-01-26 17:49:59 +0000 UTC" firstStartedPulling="2026-01-26 17:50:00.727903663 +0000 UTC m=+3767.252084097" lastFinishedPulling="2026-01-26 17:50:06.719656425 +0000 UTC m=+3773.243836859" observedRunningTime="2026-01-26 17:50:07.566352265 +0000 UTC m=+3774.090532709" watchObservedRunningTime="2026-01-26 17:50:07.576819672 +0000 UTC m=+3774.101000106" Jan 26 17:50:10 crc kubenswrapper[4754]: I0126 17:50:10.531608 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-ldztm/crc-debug-mwrvs"] Jan 26 17:50:10 crc kubenswrapper[4754]: I0126 17:50:10.534365 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ldztm/crc-debug-mwrvs" Jan 26 17:50:10 crc kubenswrapper[4754]: I0126 17:50:10.536503 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-ldztm"/"default-dockercfg-lbdr5" Jan 26 17:50:10 crc kubenswrapper[4754]: I0126 17:50:10.633224 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f2914d98-8777-4180-bcb9-fba3a6dbfed8-host\") pod \"crc-debug-mwrvs\" (UID: \"f2914d98-8777-4180-bcb9-fba3a6dbfed8\") " pod="openshift-must-gather-ldztm/crc-debug-mwrvs" Jan 26 17:50:10 crc kubenswrapper[4754]: I0126 17:50:10.633342 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8nvgb\" (UniqueName: \"kubernetes.io/projected/f2914d98-8777-4180-bcb9-fba3a6dbfed8-kube-api-access-8nvgb\") pod \"crc-debug-mwrvs\" (UID: \"f2914d98-8777-4180-bcb9-fba3a6dbfed8\") " pod="openshift-must-gather-ldztm/crc-debug-mwrvs" Jan 26 17:50:10 crc kubenswrapper[4754]: I0126 17:50:10.735580 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8nvgb\" (UniqueName: \"kubernetes.io/projected/f2914d98-8777-4180-bcb9-fba3a6dbfed8-kube-api-access-8nvgb\") pod \"crc-debug-mwrvs\" (UID: \"f2914d98-8777-4180-bcb9-fba3a6dbfed8\") " pod="openshift-must-gather-ldztm/crc-debug-mwrvs" Jan 26 17:50:10 crc kubenswrapper[4754]: I0126 17:50:10.735949 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f2914d98-8777-4180-bcb9-fba3a6dbfed8-host\") pod \"crc-debug-mwrvs\" (UID: \"f2914d98-8777-4180-bcb9-fba3a6dbfed8\") " pod="openshift-must-gather-ldztm/crc-debug-mwrvs" Jan 26 17:50:10 crc kubenswrapper[4754]: I0126 17:50:10.736130 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f2914d98-8777-4180-bcb9-fba3a6dbfed8-host\") pod \"crc-debug-mwrvs\" (UID: \"f2914d98-8777-4180-bcb9-fba3a6dbfed8\") " pod="openshift-must-gather-ldztm/crc-debug-mwrvs" Jan 26 17:50:10 crc kubenswrapper[4754]: I0126 17:50:10.760735 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8nvgb\" (UniqueName: \"kubernetes.io/projected/f2914d98-8777-4180-bcb9-fba3a6dbfed8-kube-api-access-8nvgb\") pod \"crc-debug-mwrvs\" (UID: \"f2914d98-8777-4180-bcb9-fba3a6dbfed8\") " pod="openshift-must-gather-ldztm/crc-debug-mwrvs" Jan 26 17:50:10 crc kubenswrapper[4754]: I0126 17:50:10.853111 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ldztm/crc-debug-mwrvs" Jan 26 17:50:11 crc kubenswrapper[4754]: I0126 17:50:11.586007 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-ldztm/crc-debug-mwrvs" event={"ID":"f2914d98-8777-4180-bcb9-fba3a6dbfed8","Type":"ContainerStarted","Data":"6eaed46a1c8ee134101fc887f9d19923d7ab1c0af32d428971468858d35342bb"} Jan 26 17:50:14 crc kubenswrapper[4754]: I0126 17:50:14.767687 4754 scope.go:117] "RemoveContainer" containerID="7b1482a92488015f91f13a86a1aa38c65e69fa9f874f9482c4d8f0f34463d2c2" Jan 26 17:50:14 crc kubenswrapper[4754]: E0126 17:50:14.768942 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x65wv_openshift-machine-config-operator(8c3718a4-f354-4284-92e0-fdfb45a692bd)\"" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" Jan 26 17:50:26 crc kubenswrapper[4754]: E0126 17:50:26.805052 4754 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:6ab858aed98e4fe57e6b144da8e90ad5d6698bb4cc5521206f5c05809f0f9296" Jan 26 17:50:26 crc kubenswrapper[4754]: E0126 17:50:26.805665 4754 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:container-00,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:6ab858aed98e4fe57e6b144da8e90ad5d6698bb4cc5521206f5c05809f0f9296,Command:[chroot /host bash -c echo 'TOOLBOX_NAME=toolbox-osp' > /root/.toolboxrc ; rm -rf \"/var/tmp/sos-osp\" && mkdir -p \"/var/tmp/sos-osp\" && sudo podman rm --force toolbox-osp; sudo --preserve-env podman pull --authfile /var/lib/kubelet/config.json registry.redhat.io/rhel9/support-tools && toolbox sos report --batch --all-logs --only-plugins block,cifs,crio,devicemapper,devices,firewall_tables,firewalld,iscsi,lvm2,memory,multipath,nfs,nis,nvme,podman,process,processor,selinux,scsi,udev,logs,crypto --tmp-dir=\"/var/tmp/sos-osp\" && if [[ \"$(ls /var/log/pods/*/{*.log.*,*/*.log.*} 2>/dev/null)\" != '' ]]; then tar --ignore-failed-read --warning=no-file-changed -cJf \"/var/tmp/sos-osp/podlogs.tar.xz\" --transform 's,^,podlogs/,' /var/log/pods/*/{*.log.*,*/*.log.*} || true; fi],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:TMOUT,Value:900,ValueFrom:nil,},EnvVar{Name:HOST,Value:/host,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:host,ReadOnly:false,MountPath:/host,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-8nvgb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:*true,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod crc-debug-mwrvs_openshift-must-gather-ldztm(f2914d98-8777-4180-bcb9-fba3a6dbfed8): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Jan 26 17:50:26 crc kubenswrapper[4754]: E0126 17:50:26.806954 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"container-00\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openshift-must-gather-ldztm/crc-debug-mwrvs" podUID="f2914d98-8777-4180-bcb9-fba3a6dbfed8" Jan 26 17:50:27 crc kubenswrapper[4754]: E0126 17:50:27.755381 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"container-00\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:6ab858aed98e4fe57e6b144da8e90ad5d6698bb4cc5521206f5c05809f0f9296\\\"\"" pod="openshift-must-gather-ldztm/crc-debug-mwrvs" podUID="f2914d98-8777-4180-bcb9-fba3a6dbfed8" Jan 26 17:50:27 crc kubenswrapper[4754]: I0126 17:50:27.767562 4754 scope.go:117] "RemoveContainer" containerID="7b1482a92488015f91f13a86a1aa38c65e69fa9f874f9482c4d8f0f34463d2c2" Jan 26 17:50:27 crc kubenswrapper[4754]: E0126 17:50:27.767837 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x65wv_openshift-machine-config-operator(8c3718a4-f354-4284-92e0-fdfb45a692bd)\"" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" Jan 26 17:50:38 crc kubenswrapper[4754]: I0126 17:50:38.767471 4754 scope.go:117] "RemoveContainer" containerID="7b1482a92488015f91f13a86a1aa38c65e69fa9f874f9482c4d8f0f34463d2c2" Jan 26 17:50:38 crc kubenswrapper[4754]: E0126 17:50:38.768292 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x65wv_openshift-machine-config-operator(8c3718a4-f354-4284-92e0-fdfb45a692bd)\"" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" Jan 26 17:50:40 crc kubenswrapper[4754]: I0126 17:50:40.882513 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-ldztm/crc-debug-mwrvs" event={"ID":"f2914d98-8777-4180-bcb9-fba3a6dbfed8","Type":"ContainerStarted","Data":"9763fe4f8cd2d2d229ae61d0290c1fe2bfc0e075232bde8c244f8cbbf16b7b72"} Jan 26 17:50:41 crc kubenswrapper[4754]: I0126 17:50:41.911294 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-ldztm/crc-debug-mwrvs" podStartSLOduration=2.602857859 podStartE2EDuration="31.911276679s" podCreationTimestamp="2026-01-26 17:50:10 +0000 UTC" firstStartedPulling="2026-01-26 17:50:10.897428327 +0000 UTC m=+3777.421608761" lastFinishedPulling="2026-01-26 17:50:40.205847147 +0000 UTC m=+3806.730027581" observedRunningTime="2026-01-26 17:50:41.902764815 +0000 UTC m=+3808.426945259" watchObservedRunningTime="2026-01-26 17:50:41.911276679 +0000 UTC m=+3808.435457113" Jan 26 17:50:49 crc kubenswrapper[4754]: I0126 17:50:49.767589 4754 scope.go:117] "RemoveContainer" containerID="7b1482a92488015f91f13a86a1aa38c65e69fa9f874f9482c4d8f0f34463d2c2" Jan 26 17:50:50 crc kubenswrapper[4754]: E0126 17:50:50.014958 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x65wv_openshift-machine-config-operator(8c3718a4-f354-4284-92e0-fdfb45a692bd)\"" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" Jan 26 17:51:01 crc kubenswrapper[4754]: I0126 17:51:01.767580 4754 scope.go:117] "RemoveContainer" containerID="7b1482a92488015f91f13a86a1aa38c65e69fa9f874f9482c4d8f0f34463d2c2" Jan 26 17:51:01 crc kubenswrapper[4754]: E0126 17:51:01.768482 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x65wv_openshift-machine-config-operator(8c3718a4-f354-4284-92e0-fdfb45a692bd)\"" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" Jan 26 17:51:15 crc kubenswrapper[4754]: I0126 17:51:15.767966 4754 scope.go:117] "RemoveContainer" containerID="7b1482a92488015f91f13a86a1aa38c65e69fa9f874f9482c4d8f0f34463d2c2" Jan 26 17:51:15 crc kubenswrapper[4754]: E0126 17:51:15.768807 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x65wv_openshift-machine-config-operator(8c3718a4-f354-4284-92e0-fdfb45a692bd)\"" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" Jan 26 17:51:23 crc kubenswrapper[4754]: I0126 17:51:23.244144 4754 generic.go:334] "Generic (PLEG): container finished" podID="f2914d98-8777-4180-bcb9-fba3a6dbfed8" containerID="9763fe4f8cd2d2d229ae61d0290c1fe2bfc0e075232bde8c244f8cbbf16b7b72" exitCode=0 Jan 26 17:51:23 crc kubenswrapper[4754]: I0126 17:51:23.244236 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-ldztm/crc-debug-mwrvs" event={"ID":"f2914d98-8777-4180-bcb9-fba3a6dbfed8","Type":"ContainerDied","Data":"9763fe4f8cd2d2d229ae61d0290c1fe2bfc0e075232bde8c244f8cbbf16b7b72"} Jan 26 17:51:24 crc kubenswrapper[4754]: I0126 17:51:24.377985 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ldztm/crc-debug-mwrvs" Jan 26 17:51:24 crc kubenswrapper[4754]: I0126 17:51:24.420277 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-ldztm/crc-debug-mwrvs"] Jan 26 17:51:24 crc kubenswrapper[4754]: I0126 17:51:24.431748 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-ldztm/crc-debug-mwrvs"] Jan 26 17:51:24 crc kubenswrapper[4754]: I0126 17:51:24.511766 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8nvgb\" (UniqueName: \"kubernetes.io/projected/f2914d98-8777-4180-bcb9-fba3a6dbfed8-kube-api-access-8nvgb\") pod \"f2914d98-8777-4180-bcb9-fba3a6dbfed8\" (UID: \"f2914d98-8777-4180-bcb9-fba3a6dbfed8\") " Jan 26 17:51:24 crc kubenswrapper[4754]: I0126 17:51:24.511931 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f2914d98-8777-4180-bcb9-fba3a6dbfed8-host\") pod \"f2914d98-8777-4180-bcb9-fba3a6dbfed8\" (UID: \"f2914d98-8777-4180-bcb9-fba3a6dbfed8\") " Jan 26 17:51:24 crc kubenswrapper[4754]: I0126 17:51:24.512125 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f2914d98-8777-4180-bcb9-fba3a6dbfed8-host" (OuterVolumeSpecName: "host") pod "f2914d98-8777-4180-bcb9-fba3a6dbfed8" (UID: "f2914d98-8777-4180-bcb9-fba3a6dbfed8"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 17:51:24 crc kubenswrapper[4754]: I0126 17:51:24.512377 4754 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f2914d98-8777-4180-bcb9-fba3a6dbfed8-host\") on node \"crc\" DevicePath \"\"" Jan 26 17:51:24 crc kubenswrapper[4754]: I0126 17:51:24.517658 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f2914d98-8777-4180-bcb9-fba3a6dbfed8-kube-api-access-8nvgb" (OuterVolumeSpecName: "kube-api-access-8nvgb") pod "f2914d98-8777-4180-bcb9-fba3a6dbfed8" (UID: "f2914d98-8777-4180-bcb9-fba3a6dbfed8"). InnerVolumeSpecName "kube-api-access-8nvgb". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:51:24 crc kubenswrapper[4754]: I0126 17:51:24.613831 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8nvgb\" (UniqueName: \"kubernetes.io/projected/f2914d98-8777-4180-bcb9-fba3a6dbfed8-kube-api-access-8nvgb\") on node \"crc\" DevicePath \"\"" Jan 26 17:51:25 crc kubenswrapper[4754]: I0126 17:51:25.263467 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6eaed46a1c8ee134101fc887f9d19923d7ab1c0af32d428971468858d35342bb" Jan 26 17:51:25 crc kubenswrapper[4754]: I0126 17:51:25.263553 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ldztm/crc-debug-mwrvs" Jan 26 17:51:25 crc kubenswrapper[4754]: I0126 17:51:25.573360 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-ldztm/crc-debug-d8tnl"] Jan 26 17:51:25 crc kubenswrapper[4754]: E0126 17:51:25.573872 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f2914d98-8777-4180-bcb9-fba3a6dbfed8" containerName="container-00" Jan 26 17:51:25 crc kubenswrapper[4754]: I0126 17:51:25.573888 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="f2914d98-8777-4180-bcb9-fba3a6dbfed8" containerName="container-00" Jan 26 17:51:25 crc kubenswrapper[4754]: I0126 17:51:25.574115 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="f2914d98-8777-4180-bcb9-fba3a6dbfed8" containerName="container-00" Jan 26 17:51:25 crc kubenswrapper[4754]: I0126 17:51:25.574919 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ldztm/crc-debug-d8tnl" Jan 26 17:51:25 crc kubenswrapper[4754]: I0126 17:51:25.578776 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-ldztm"/"default-dockercfg-lbdr5" Jan 26 17:51:25 crc kubenswrapper[4754]: I0126 17:51:25.632390 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/3aaec58a-3af9-4c5a-97ba-416d56c3a6a4-host\") pod \"crc-debug-d8tnl\" (UID: \"3aaec58a-3af9-4c5a-97ba-416d56c3a6a4\") " pod="openshift-must-gather-ldztm/crc-debug-d8tnl" Jan 26 17:51:25 crc kubenswrapper[4754]: I0126 17:51:25.632514 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pccrh\" (UniqueName: \"kubernetes.io/projected/3aaec58a-3af9-4c5a-97ba-416d56c3a6a4-kube-api-access-pccrh\") pod \"crc-debug-d8tnl\" (UID: \"3aaec58a-3af9-4c5a-97ba-416d56c3a6a4\") " pod="openshift-must-gather-ldztm/crc-debug-d8tnl" Jan 26 17:51:25 crc kubenswrapper[4754]: I0126 17:51:25.733811 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pccrh\" (UniqueName: \"kubernetes.io/projected/3aaec58a-3af9-4c5a-97ba-416d56c3a6a4-kube-api-access-pccrh\") pod \"crc-debug-d8tnl\" (UID: \"3aaec58a-3af9-4c5a-97ba-416d56c3a6a4\") " pod="openshift-must-gather-ldztm/crc-debug-d8tnl" Jan 26 17:51:25 crc kubenswrapper[4754]: I0126 17:51:25.734018 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/3aaec58a-3af9-4c5a-97ba-416d56c3a6a4-host\") pod \"crc-debug-d8tnl\" (UID: \"3aaec58a-3af9-4c5a-97ba-416d56c3a6a4\") " pod="openshift-must-gather-ldztm/crc-debug-d8tnl" Jan 26 17:51:25 crc kubenswrapper[4754]: I0126 17:51:25.734133 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/3aaec58a-3af9-4c5a-97ba-416d56c3a6a4-host\") pod \"crc-debug-d8tnl\" (UID: \"3aaec58a-3af9-4c5a-97ba-416d56c3a6a4\") " pod="openshift-must-gather-ldztm/crc-debug-d8tnl" Jan 26 17:51:25 crc kubenswrapper[4754]: I0126 17:51:25.750083 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pccrh\" (UniqueName: \"kubernetes.io/projected/3aaec58a-3af9-4c5a-97ba-416d56c3a6a4-kube-api-access-pccrh\") pod \"crc-debug-d8tnl\" (UID: \"3aaec58a-3af9-4c5a-97ba-416d56c3a6a4\") " pod="openshift-must-gather-ldztm/crc-debug-d8tnl" Jan 26 17:51:25 crc kubenswrapper[4754]: I0126 17:51:25.777842 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f2914d98-8777-4180-bcb9-fba3a6dbfed8" path="/var/lib/kubelet/pods/f2914d98-8777-4180-bcb9-fba3a6dbfed8/volumes" Jan 26 17:51:25 crc kubenswrapper[4754]: I0126 17:51:25.894828 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ldztm/crc-debug-d8tnl" Jan 26 17:51:26 crc kubenswrapper[4754]: I0126 17:51:26.274183 4754 generic.go:334] "Generic (PLEG): container finished" podID="3aaec58a-3af9-4c5a-97ba-416d56c3a6a4" containerID="b9b30debc34721139f192d189ec3db967ad94b14831d970e5e44915564802d8a" exitCode=0 Jan 26 17:51:26 crc kubenswrapper[4754]: I0126 17:51:26.274301 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-ldztm/crc-debug-d8tnl" event={"ID":"3aaec58a-3af9-4c5a-97ba-416d56c3a6a4","Type":"ContainerDied","Data":"b9b30debc34721139f192d189ec3db967ad94b14831d970e5e44915564802d8a"} Jan 26 17:51:26 crc kubenswrapper[4754]: I0126 17:51:26.274582 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-ldztm/crc-debug-d8tnl" event={"ID":"3aaec58a-3af9-4c5a-97ba-416d56c3a6a4","Type":"ContainerStarted","Data":"23cacc3e3372cda0386e68338c37f30dac91f1773fe4bbdbf62fe53e2b1e1058"} Jan 26 17:51:26 crc kubenswrapper[4754]: I0126 17:51:26.764222 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-ldztm/crc-debug-d8tnl"] Jan 26 17:51:26 crc kubenswrapper[4754]: I0126 17:51:26.772293 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-ldztm/crc-debug-d8tnl"] Jan 26 17:51:27 crc kubenswrapper[4754]: I0126 17:51:27.382278 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ldztm/crc-debug-d8tnl" Jan 26 17:51:27 crc kubenswrapper[4754]: I0126 17:51:27.467469 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pccrh\" (UniqueName: \"kubernetes.io/projected/3aaec58a-3af9-4c5a-97ba-416d56c3a6a4-kube-api-access-pccrh\") pod \"3aaec58a-3af9-4c5a-97ba-416d56c3a6a4\" (UID: \"3aaec58a-3af9-4c5a-97ba-416d56c3a6a4\") " Jan 26 17:51:27 crc kubenswrapper[4754]: I0126 17:51:27.467653 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/3aaec58a-3af9-4c5a-97ba-416d56c3a6a4-host\") pod \"3aaec58a-3af9-4c5a-97ba-416d56c3a6a4\" (UID: \"3aaec58a-3af9-4c5a-97ba-416d56c3a6a4\") " Jan 26 17:51:27 crc kubenswrapper[4754]: I0126 17:51:27.467874 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3aaec58a-3af9-4c5a-97ba-416d56c3a6a4-host" (OuterVolumeSpecName: "host") pod "3aaec58a-3af9-4c5a-97ba-416d56c3a6a4" (UID: "3aaec58a-3af9-4c5a-97ba-416d56c3a6a4"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 17:51:27 crc kubenswrapper[4754]: I0126 17:51:27.468330 4754 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/3aaec58a-3af9-4c5a-97ba-416d56c3a6a4-host\") on node \"crc\" DevicePath \"\"" Jan 26 17:51:27 crc kubenswrapper[4754]: I0126 17:51:27.473540 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3aaec58a-3af9-4c5a-97ba-416d56c3a6a4-kube-api-access-pccrh" (OuterVolumeSpecName: "kube-api-access-pccrh") pod "3aaec58a-3af9-4c5a-97ba-416d56c3a6a4" (UID: "3aaec58a-3af9-4c5a-97ba-416d56c3a6a4"). InnerVolumeSpecName "kube-api-access-pccrh". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:51:27 crc kubenswrapper[4754]: I0126 17:51:27.570150 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pccrh\" (UniqueName: \"kubernetes.io/projected/3aaec58a-3af9-4c5a-97ba-416d56c3a6a4-kube-api-access-pccrh\") on node \"crc\" DevicePath \"\"" Jan 26 17:51:27 crc kubenswrapper[4754]: I0126 17:51:27.779329 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3aaec58a-3af9-4c5a-97ba-416d56c3a6a4" path="/var/lib/kubelet/pods/3aaec58a-3af9-4c5a-97ba-416d56c3a6a4/volumes" Jan 26 17:51:27 crc kubenswrapper[4754]: I0126 17:51:27.914516 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-ldztm/crc-debug-fhwtm"] Jan 26 17:51:27 crc kubenswrapper[4754]: E0126 17:51:27.914916 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3aaec58a-3af9-4c5a-97ba-416d56c3a6a4" containerName="container-00" Jan 26 17:51:27 crc kubenswrapper[4754]: I0126 17:51:27.914932 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="3aaec58a-3af9-4c5a-97ba-416d56c3a6a4" containerName="container-00" Jan 26 17:51:27 crc kubenswrapper[4754]: I0126 17:51:27.915135 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="3aaec58a-3af9-4c5a-97ba-416d56c3a6a4" containerName="container-00" Jan 26 17:51:27 crc kubenswrapper[4754]: I0126 17:51:27.915694 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ldztm/crc-debug-fhwtm" Jan 26 17:51:27 crc kubenswrapper[4754]: I0126 17:51:27.977211 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-596gt\" (UniqueName: \"kubernetes.io/projected/2abb48e6-d767-43d1-a7e2-ea739fd9dd82-kube-api-access-596gt\") pod \"crc-debug-fhwtm\" (UID: \"2abb48e6-d767-43d1-a7e2-ea739fd9dd82\") " pod="openshift-must-gather-ldztm/crc-debug-fhwtm" Jan 26 17:51:27 crc kubenswrapper[4754]: I0126 17:51:27.977329 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/2abb48e6-d767-43d1-a7e2-ea739fd9dd82-host\") pod \"crc-debug-fhwtm\" (UID: \"2abb48e6-d767-43d1-a7e2-ea739fd9dd82\") " pod="openshift-must-gather-ldztm/crc-debug-fhwtm" Jan 26 17:51:28 crc kubenswrapper[4754]: I0126 17:51:28.082493 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/2abb48e6-d767-43d1-a7e2-ea739fd9dd82-host\") pod \"crc-debug-fhwtm\" (UID: \"2abb48e6-d767-43d1-a7e2-ea739fd9dd82\") " pod="openshift-must-gather-ldztm/crc-debug-fhwtm" Jan 26 17:51:28 crc kubenswrapper[4754]: I0126 17:51:28.082685 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/2abb48e6-d767-43d1-a7e2-ea739fd9dd82-host\") pod \"crc-debug-fhwtm\" (UID: \"2abb48e6-d767-43d1-a7e2-ea739fd9dd82\") " pod="openshift-must-gather-ldztm/crc-debug-fhwtm" Jan 26 17:51:28 crc kubenswrapper[4754]: I0126 17:51:28.082734 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-596gt\" (UniqueName: \"kubernetes.io/projected/2abb48e6-d767-43d1-a7e2-ea739fd9dd82-kube-api-access-596gt\") pod \"crc-debug-fhwtm\" (UID: \"2abb48e6-d767-43d1-a7e2-ea739fd9dd82\") " pod="openshift-must-gather-ldztm/crc-debug-fhwtm" Jan 26 17:51:28 crc kubenswrapper[4754]: I0126 17:51:28.102517 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-596gt\" (UniqueName: \"kubernetes.io/projected/2abb48e6-d767-43d1-a7e2-ea739fd9dd82-kube-api-access-596gt\") pod \"crc-debug-fhwtm\" (UID: \"2abb48e6-d767-43d1-a7e2-ea739fd9dd82\") " pod="openshift-must-gather-ldztm/crc-debug-fhwtm" Jan 26 17:51:28 crc kubenswrapper[4754]: I0126 17:51:28.246862 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ldztm/crc-debug-fhwtm" Jan 26 17:51:28 crc kubenswrapper[4754]: W0126 17:51:28.277173 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2abb48e6_d767_43d1_a7e2_ea739fd9dd82.slice/crio-16bb33e0a3b1b178cc8ae522b94f859ee2fe9a8697c334ba866a196c448e2d4f WatchSource:0}: Error finding container 16bb33e0a3b1b178cc8ae522b94f859ee2fe9a8697c334ba866a196c448e2d4f: Status 404 returned error can't find the container with id 16bb33e0a3b1b178cc8ae522b94f859ee2fe9a8697c334ba866a196c448e2d4f Jan 26 17:51:28 crc kubenswrapper[4754]: I0126 17:51:28.291162 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-ldztm/crc-debug-fhwtm" event={"ID":"2abb48e6-d767-43d1-a7e2-ea739fd9dd82","Type":"ContainerStarted","Data":"16bb33e0a3b1b178cc8ae522b94f859ee2fe9a8697c334ba866a196c448e2d4f"} Jan 26 17:51:28 crc kubenswrapper[4754]: I0126 17:51:28.292757 4754 scope.go:117] "RemoveContainer" containerID="b9b30debc34721139f192d189ec3db967ad94b14831d970e5e44915564802d8a" Jan 26 17:51:28 crc kubenswrapper[4754]: I0126 17:51:28.292910 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ldztm/crc-debug-d8tnl" Jan 26 17:51:29 crc kubenswrapper[4754]: I0126 17:51:29.304089 4754 generic.go:334] "Generic (PLEG): container finished" podID="2abb48e6-d767-43d1-a7e2-ea739fd9dd82" containerID="dbbf2dafd2905277c477a43a512f34f4a5fc6b05d065050d629e8811ba277690" exitCode=0 Jan 26 17:51:29 crc kubenswrapper[4754]: I0126 17:51:29.304174 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-ldztm/crc-debug-fhwtm" event={"ID":"2abb48e6-d767-43d1-a7e2-ea739fd9dd82","Type":"ContainerDied","Data":"dbbf2dafd2905277c477a43a512f34f4a5fc6b05d065050d629e8811ba277690"} Jan 26 17:51:29 crc kubenswrapper[4754]: I0126 17:51:29.343201 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-ldztm/crc-debug-fhwtm"] Jan 26 17:51:29 crc kubenswrapper[4754]: I0126 17:51:29.350803 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-ldztm/crc-debug-fhwtm"] Jan 26 17:51:29 crc kubenswrapper[4754]: I0126 17:51:29.767738 4754 scope.go:117] "RemoveContainer" containerID="7b1482a92488015f91f13a86a1aa38c65e69fa9f874f9482c4d8f0f34463d2c2" Jan 26 17:51:29 crc kubenswrapper[4754]: E0126 17:51:29.768317 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x65wv_openshift-machine-config-operator(8c3718a4-f354-4284-92e0-fdfb45a692bd)\"" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" Jan 26 17:51:30 crc kubenswrapper[4754]: I0126 17:51:30.447744 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ldztm/crc-debug-fhwtm" Jan 26 17:51:30 crc kubenswrapper[4754]: I0126 17:51:30.528049 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-596gt\" (UniqueName: \"kubernetes.io/projected/2abb48e6-d767-43d1-a7e2-ea739fd9dd82-kube-api-access-596gt\") pod \"2abb48e6-d767-43d1-a7e2-ea739fd9dd82\" (UID: \"2abb48e6-d767-43d1-a7e2-ea739fd9dd82\") " Jan 26 17:51:30 crc kubenswrapper[4754]: I0126 17:51:30.528247 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/2abb48e6-d767-43d1-a7e2-ea739fd9dd82-host\") pod \"2abb48e6-d767-43d1-a7e2-ea739fd9dd82\" (UID: \"2abb48e6-d767-43d1-a7e2-ea739fd9dd82\") " Jan 26 17:51:30 crc kubenswrapper[4754]: I0126 17:51:30.528380 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2abb48e6-d767-43d1-a7e2-ea739fd9dd82-host" (OuterVolumeSpecName: "host") pod "2abb48e6-d767-43d1-a7e2-ea739fd9dd82" (UID: "2abb48e6-d767-43d1-a7e2-ea739fd9dd82"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 17:51:30 crc kubenswrapper[4754]: I0126 17:51:30.528732 4754 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/2abb48e6-d767-43d1-a7e2-ea739fd9dd82-host\") on node \"crc\" DevicePath \"\"" Jan 26 17:51:30 crc kubenswrapper[4754]: I0126 17:51:30.541952 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2abb48e6-d767-43d1-a7e2-ea739fd9dd82-kube-api-access-596gt" (OuterVolumeSpecName: "kube-api-access-596gt") pod "2abb48e6-d767-43d1-a7e2-ea739fd9dd82" (UID: "2abb48e6-d767-43d1-a7e2-ea739fd9dd82"). InnerVolumeSpecName "kube-api-access-596gt". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:51:30 crc kubenswrapper[4754]: I0126 17:51:30.630647 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-596gt\" (UniqueName: \"kubernetes.io/projected/2abb48e6-d767-43d1-a7e2-ea739fd9dd82-kube-api-access-596gt\") on node \"crc\" DevicePath \"\"" Jan 26 17:51:31 crc kubenswrapper[4754]: I0126 17:51:31.331453 4754 scope.go:117] "RemoveContainer" containerID="dbbf2dafd2905277c477a43a512f34f4a5fc6b05d065050d629e8811ba277690" Jan 26 17:51:31 crc kubenswrapper[4754]: I0126 17:51:31.331490 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ldztm/crc-debug-fhwtm" Jan 26 17:51:31 crc kubenswrapper[4754]: I0126 17:51:31.778771 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2abb48e6-d767-43d1-a7e2-ea739fd9dd82" path="/var/lib/kubelet/pods/2abb48e6-d767-43d1-a7e2-ea739fd9dd82/volumes" Jan 26 17:51:42 crc kubenswrapper[4754]: I0126 17:51:42.767357 4754 scope.go:117] "RemoveContainer" containerID="7b1482a92488015f91f13a86a1aa38c65e69fa9f874f9482c4d8f0f34463d2c2" Jan 26 17:51:42 crc kubenswrapper[4754]: E0126 17:51:42.768179 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x65wv_openshift-machine-config-operator(8c3718a4-f354-4284-92e0-fdfb45a692bd)\"" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" Jan 26 17:51:45 crc kubenswrapper[4754]: I0126 17:51:45.022167 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-5d4bd978-rj87r_8c7859b2-be6e-49c3-8392-47ed649a1f68/barbican-api/0.log" Jan 26 17:51:45 crc kubenswrapper[4754]: I0126 17:51:45.174195 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-5d4bd978-rj87r_8c7859b2-be6e-49c3-8392-47ed649a1f68/barbican-api-log/0.log" Jan 26 17:51:45 crc kubenswrapper[4754]: I0126 17:51:45.216744 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-99649c64-sldzz_0b004ba0-01f6-49f9-8d8d-56033dcd3533/barbican-keystone-listener/0.log" Jan 26 17:51:45 crc kubenswrapper[4754]: I0126 17:51:45.329516 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-99649c64-sldzz_0b004ba0-01f6-49f9-8d8d-56033dcd3533/barbican-keystone-listener-log/0.log" Jan 26 17:51:45 crc kubenswrapper[4754]: I0126 17:51:45.451992 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-5677764859-tswcr_d64625ff-6586-4c80-a720-b2febd49a966/barbican-worker/0.log" Jan 26 17:51:45 crc kubenswrapper[4754]: I0126 17:51:45.482137 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-5677764859-tswcr_d64625ff-6586-4c80-a720-b2febd49a966/barbican-worker-log/0.log" Jan 26 17:51:45 crc kubenswrapper[4754]: I0126 17:51:45.645324 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-nng92_fc4da904-699d-44a4-995d-d5ac9b05695b/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Jan 26 17:51:45 crc kubenswrapper[4754]: I0126 17:51:45.714988 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_b41f5e0e-1829-41f9-926c-ce15178e16c8/ceilometer-central-agent/0.log" Jan 26 17:51:45 crc kubenswrapper[4754]: I0126 17:51:45.781063 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_b41f5e0e-1829-41f9-926c-ce15178e16c8/ceilometer-notification-agent/0.log" Jan 26 17:51:45 crc kubenswrapper[4754]: I0126 17:51:45.847703 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_b41f5e0e-1829-41f9-926c-ce15178e16c8/proxy-httpd/0.log" Jan 26 17:51:45 crc kubenswrapper[4754]: I0126 17:51:45.883494 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_b41f5e0e-1829-41f9-926c-ce15178e16c8/sg-core/0.log" Jan 26 17:51:46 crc kubenswrapper[4754]: I0126 17:51:46.058382 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_529bf77e-614d-4354-94bd-e6383f353920/cinder-api-log/0.log" Jan 26 17:51:46 crc kubenswrapper[4754]: I0126 17:51:46.061265 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_529bf77e-614d-4354-94bd-e6383f353920/cinder-api/0.log" Jan 26 17:51:46 crc kubenswrapper[4754]: I0126 17:51:46.228141 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_6147c61d-7878-41b8-8ce6-a165c9f03ede/cinder-scheduler/0.log" Jan 26 17:51:46 crc kubenswrapper[4754]: I0126 17:51:46.283326 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_6147c61d-7878-41b8-8ce6-a165c9f03ede/probe/0.log" Jan 26 17:51:46 crc kubenswrapper[4754]: I0126 17:51:46.370543 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-5d95m_51a9ce2d-f224-449d-8988-950a60783ddb/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Jan 26 17:51:46 crc kubenswrapper[4754]: I0126 17:51:46.601497 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-v9nvp_c31e5ca9-fa72-4350-bd79-b58eb9b0e7a4/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Jan 26 17:51:46 crc kubenswrapper[4754]: I0126 17:51:46.609467 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-55478c4467-c65ql_48e65463-af70-4cca-b90b-a3b4ac9a1619/init/0.log" Jan 26 17:51:46 crc kubenswrapper[4754]: I0126 17:51:46.833298 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-55478c4467-c65ql_48e65463-af70-4cca-b90b-a3b4ac9a1619/init/0.log" Jan 26 17:51:46 crc kubenswrapper[4754]: I0126 17:51:46.871375 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-55478c4467-c65ql_48e65463-af70-4cca-b90b-a3b4ac9a1619/dnsmasq-dns/0.log" Jan 26 17:51:46 crc kubenswrapper[4754]: I0126 17:51:46.926633 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-rrhnq_55b57fed-619a-44b7-af60-f6d7a43943f1/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Jan 26 17:51:47 crc kubenswrapper[4754]: I0126 17:51:47.096586 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_724e114a-1ace-4455-846f-d7ab65c593f1/glance-httpd/0.log" Jan 26 17:51:47 crc kubenswrapper[4754]: I0126 17:51:47.104526 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_724e114a-1ace-4455-846f-d7ab65c593f1/glance-log/0.log" Jan 26 17:51:47 crc kubenswrapper[4754]: I0126 17:51:47.243134 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_c0decb48-8392-4a9a-a253-82b24b4f07ef/glance-httpd/0.log" Jan 26 17:51:47 crc kubenswrapper[4754]: I0126 17:51:47.318139 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_c0decb48-8392-4a9a-a253-82b24b4f07ef/glance-log/0.log" Jan 26 17:51:47 crc kubenswrapper[4754]: I0126 17:51:47.470048 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-6dff6969b8-5bklb_0313cdfe-e309-41aa-a5ab-83d7713628f3/horizon/0.log" Jan 26 17:51:47 crc kubenswrapper[4754]: I0126 17:51:47.591015 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-qpc2q_70a69e93-9d4a-4d51-b0c9-4b986172e2ac/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Jan 26 17:51:47 crc kubenswrapper[4754]: I0126 17:51:47.807191 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-6dff6969b8-5bklb_0313cdfe-e309-41aa-a5ab-83d7713628f3/horizon-log/0.log" Jan 26 17:51:47 crc kubenswrapper[4754]: I0126 17:51:47.832343 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-r92d9_9bfc59f9-f503-4fec-9a31-240819fc3a52/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Jan 26 17:51:48 crc kubenswrapper[4754]: I0126 17:51:48.040262 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_6be68290-e783-4f39-8257-d5dc4051447a/kube-state-metrics/0.log" Jan 26 17:51:48 crc kubenswrapper[4754]: I0126 17:51:48.131463 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-6c774f65c8-h55rj_928f07ae-77be-48d4-ba56-daaa5ff400f0/keystone-api/0.log" Jan 26 17:51:48 crc kubenswrapper[4754]: I0126 17:51:48.308737 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-lpbsc_a8d7c557-f25d-4aaa-94be-cf9e8e26bc19/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Jan 26 17:51:48 crc kubenswrapper[4754]: I0126 17:51:48.720521 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-767b9f4849-cfqbz_36eccc33-604a-49ac-a1f2-c0bd41fd053d/neutron-api/0.log" Jan 26 17:51:48 crc kubenswrapper[4754]: I0126 17:51:48.745654 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-767b9f4849-cfqbz_36eccc33-604a-49ac-a1f2-c0bd41fd053d/neutron-httpd/0.log" Jan 26 17:51:49 crc kubenswrapper[4754]: I0126 17:51:49.012519 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-wljzd_18417e65-9c20-4b5e-96df-b7b79c67433a/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Jan 26 17:51:50 crc kubenswrapper[4754]: I0126 17:51:50.386580 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_7a81d671-1d63-4eac-a3fd-bf0e9bf38cc2/nova-cell1-conductor-conductor/0.log" Jan 26 17:51:50 crc kubenswrapper[4754]: I0126 17:51:50.457805 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_1d793a29-dacb-46ec-a0a8-ead07be4ce50/nova-cell0-conductor-conductor/0.log" Jan 26 17:51:50 crc kubenswrapper[4754]: I0126 17:51:50.624154 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_0bb1f389-c4a6-47dd-8445-c7125779ef38/nova-api-log/0.log" Jan 26 17:51:50 crc kubenswrapper[4754]: I0126 17:51:50.822259 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_fbd0b5ca-aa69-4a3a-a81d-5b912f5db6a1/nova-cell1-novncproxy-novncproxy/0.log" Jan 26 17:51:50 crc kubenswrapper[4754]: I0126 17:51:50.825695 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_0bb1f389-c4a6-47dd-8445-c7125779ef38/nova-api-api/0.log" Jan 26 17:51:51 crc kubenswrapper[4754]: I0126 17:51:51.518894 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-jswg2_b5408b33-3b71-46ac-87a4-413ecb9614b0/nova-edpm-deployment-openstack-edpm-ipam/0.log" Jan 26 17:51:51 crc kubenswrapper[4754]: I0126 17:51:51.752320 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_4e7eefb7-e66a-4c65-bda9-7997970991f3/nova-metadata-log/0.log" Jan 26 17:51:52 crc kubenswrapper[4754]: I0126 17:51:52.073504 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_6f952bb4-186c-4e7c-bf01-67d8a6985319/nova-scheduler-scheduler/0.log" Jan 26 17:51:52 crc kubenswrapper[4754]: I0126 17:51:52.088071 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_242703dd-5a52-4da8-af40-47c3490fd6ea/mysql-bootstrap/0.log" Jan 26 17:51:53 crc kubenswrapper[4754]: I0126 17:51:53.072370 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_4e7eefb7-e66a-4c65-bda9-7997970991f3/nova-metadata-metadata/0.log" Jan 26 17:51:53 crc kubenswrapper[4754]: I0126 17:51:53.108791 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_242703dd-5a52-4da8-af40-47c3490fd6ea/mysql-bootstrap/0.log" Jan 26 17:51:53 crc kubenswrapper[4754]: I0126 17:51:53.124228 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_242703dd-5a52-4da8-af40-47c3490fd6ea/galera/0.log" Jan 26 17:51:53 crc kubenswrapper[4754]: I0126 17:51:53.292262 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_0390a195-4349-4dae-8ae2-72b9b16af4f7/mysql-bootstrap/0.log" Jan 26 17:51:53 crc kubenswrapper[4754]: I0126 17:51:53.551540 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_0390a195-4349-4dae-8ae2-72b9b16af4f7/galera/0.log" Jan 26 17:51:53 crc kubenswrapper[4754]: I0126 17:51:53.572464 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_0390a195-4349-4dae-8ae2-72b9b16af4f7/mysql-bootstrap/0.log" Jan 26 17:51:53 crc kubenswrapper[4754]: I0126 17:51:53.605786 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_19dee3c5-51dc-46be-8454-9c10c76b3655/openstackclient/0.log" Jan 26 17:51:54 crc kubenswrapper[4754]: I0126 17:51:54.010712 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-x2tgk_28bf8197-f3d2-4ee1-9054-482fa295d92d/openstack-network-exporter/0.log" Jan 26 17:51:54 crc kubenswrapper[4754]: I0126 17:51:54.039235 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-8bgjb_6db16a49-6566-42bf-91ad-c34be46e7800/ovn-controller/0.log" Jan 26 17:51:54 crc kubenswrapper[4754]: I0126 17:51:54.255685 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-7psgg_ee1b9e37-8fd3-4280-af09-9f2f45366870/ovsdb-server-init/0.log" Jan 26 17:51:54 crc kubenswrapper[4754]: I0126 17:51:54.424928 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-7psgg_ee1b9e37-8fd3-4280-af09-9f2f45366870/ovsdb-server-init/0.log" Jan 26 17:51:54 crc kubenswrapper[4754]: I0126 17:51:54.455702 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-7psgg_ee1b9e37-8fd3-4280-af09-9f2f45366870/ovs-vswitchd/0.log" Jan 26 17:51:54 crc kubenswrapper[4754]: I0126 17:51:54.495469 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-7psgg_ee1b9e37-8fd3-4280-af09-9f2f45366870/ovsdb-server/0.log" Jan 26 17:51:54 crc kubenswrapper[4754]: I0126 17:51:54.708128 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_4af75510-1f61-40f7-a292-764facf90f1a/openstack-network-exporter/0.log" Jan 26 17:51:54 crc kubenswrapper[4754]: I0126 17:51:54.717094 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-xsq5z_9057d2fb-d9b9-4ce9-b219-9f15c5b7f051/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Jan 26 17:51:54 crc kubenswrapper[4754]: I0126 17:51:54.754378 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_4af75510-1f61-40f7-a292-764facf90f1a/ovn-northd/0.log" Jan 26 17:51:55 crc kubenswrapper[4754]: I0126 17:51:55.114058 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_85c77631-974b-42b0-a934-268213691414/openstack-network-exporter/0.log" Jan 26 17:51:55 crc kubenswrapper[4754]: I0126 17:51:55.153333 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_85c77631-974b-42b0-a934-268213691414/ovsdbserver-nb/0.log" Jan 26 17:51:55 crc kubenswrapper[4754]: I0126 17:51:55.317560 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_10afcb49-f3f2-4598-a8a8-45729720e109/ovsdbserver-sb/0.log" Jan 26 17:51:55 crc kubenswrapper[4754]: I0126 17:51:55.338410 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_10afcb49-f3f2-4598-a8a8-45729720e109/openstack-network-exporter/0.log" Jan 26 17:51:55 crc kubenswrapper[4754]: I0126 17:51:55.488162 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-577b5f64f6-qwc6n_854736cc-7859-4621-865c-69cab5f7dbeb/placement-api/0.log" Jan 26 17:51:55 crc kubenswrapper[4754]: I0126 17:51:55.635872 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_9acb97f6-82ac-4891-96dc-43a85f9c4e7c/setup-container/0.log" Jan 26 17:51:55 crc kubenswrapper[4754]: I0126 17:51:55.647698 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-577b5f64f6-qwc6n_854736cc-7859-4621-865c-69cab5f7dbeb/placement-log/0.log" Jan 26 17:51:55 crc kubenswrapper[4754]: I0126 17:51:55.767321 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_9acb97f6-82ac-4891-96dc-43a85f9c4e7c/rabbitmq/0.log" Jan 26 17:51:55 crc kubenswrapper[4754]: I0126 17:51:55.833902 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_9acb97f6-82ac-4891-96dc-43a85f9c4e7c/setup-container/0.log" Jan 26 17:51:55 crc kubenswrapper[4754]: I0126 17:51:55.917269 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_6e2da069-1c54-4801-a91c-241b80e8d17b/setup-container/0.log" Jan 26 17:51:56 crc kubenswrapper[4754]: I0126 17:51:56.153333 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_6e2da069-1c54-4801-a91c-241b80e8d17b/setup-container/0.log" Jan 26 17:51:56 crc kubenswrapper[4754]: I0126 17:51:56.401860 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-p4jpg_e4cbbee8-ed91-4ad6-983b-569cff60fa07/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Jan 26 17:51:56 crc kubenswrapper[4754]: I0126 17:51:56.411745 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_6e2da069-1c54-4801-a91c-241b80e8d17b/rabbitmq/0.log" Jan 26 17:51:56 crc kubenswrapper[4754]: I0126 17:51:56.438524 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-6w45w_09f02458-d4e3-49d4-8735-467141a57b6c/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Jan 26 17:51:56 crc kubenswrapper[4754]: I0126 17:51:56.705803 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-7rh8g_3dbed37c-3982-4625-929e-d99a8fd798e6/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Jan 26 17:51:56 crc kubenswrapper[4754]: I0126 17:51:56.707871 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-gn6jw_2d80ff89-c1fc-4331-9683-2740c69d001d/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Jan 26 17:51:56 crc kubenswrapper[4754]: I0126 17:51:56.767392 4754 scope.go:117] "RemoveContainer" containerID="7b1482a92488015f91f13a86a1aa38c65e69fa9f874f9482c4d8f0f34463d2c2" Jan 26 17:51:56 crc kubenswrapper[4754]: E0126 17:51:56.767717 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x65wv_openshift-machine-config-operator(8c3718a4-f354-4284-92e0-fdfb45a692bd)\"" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" Jan 26 17:51:56 crc kubenswrapper[4754]: I0126 17:51:56.900852 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-sqqxn_8dcb5a12-a14f-4f95-8991-5a1b9b9bb431/ssh-known-hosts-edpm-deployment/0.log" Jan 26 17:51:57 crc kubenswrapper[4754]: I0126 17:51:57.111382 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-6959bf4485-bkv7n_1622eb04-be06-4e2a-90d6-27a58ac54e60/proxy-httpd/0.log" Jan 26 17:51:57 crc kubenswrapper[4754]: I0126 17:51:57.129059 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-6959bf4485-bkv7n_1622eb04-be06-4e2a-90d6-27a58ac54e60/proxy-server/0.log" Jan 26 17:51:57 crc kubenswrapper[4754]: I0126 17:51:57.257392 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-sl9kb_cb88274a-904f-4827-9518-81b79a0e6a42/swift-ring-rebalance/0.log" Jan 26 17:51:57 crc kubenswrapper[4754]: I0126 17:51:57.376487 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_19a0730c-46d0-4029-a86b-812ba1664dcc/account-auditor/0.log" Jan 26 17:51:57 crc kubenswrapper[4754]: I0126 17:51:57.464773 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_19a0730c-46d0-4029-a86b-812ba1664dcc/account-reaper/0.log" Jan 26 17:51:57 crc kubenswrapper[4754]: I0126 17:51:57.521202 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_19a0730c-46d0-4029-a86b-812ba1664dcc/account-replicator/0.log" Jan 26 17:51:57 crc kubenswrapper[4754]: I0126 17:51:57.581127 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_19a0730c-46d0-4029-a86b-812ba1664dcc/account-server/0.log" Jan 26 17:51:57 crc kubenswrapper[4754]: I0126 17:51:57.668377 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_19a0730c-46d0-4029-a86b-812ba1664dcc/container-auditor/0.log" Jan 26 17:51:57 crc kubenswrapper[4754]: I0126 17:51:57.683278 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_19a0730c-46d0-4029-a86b-812ba1664dcc/container-replicator/0.log" Jan 26 17:51:57 crc kubenswrapper[4754]: I0126 17:51:57.759909 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_19a0730c-46d0-4029-a86b-812ba1664dcc/container-server/0.log" Jan 26 17:51:57 crc kubenswrapper[4754]: I0126 17:51:57.778077 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_19a0730c-46d0-4029-a86b-812ba1664dcc/container-updater/0.log" Jan 26 17:51:57 crc kubenswrapper[4754]: I0126 17:51:57.948566 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_19a0730c-46d0-4029-a86b-812ba1664dcc/object-expirer/0.log" Jan 26 17:51:58 crc kubenswrapper[4754]: I0126 17:51:58.002212 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_19a0730c-46d0-4029-a86b-812ba1664dcc/object-auditor/0.log" Jan 26 17:51:58 crc kubenswrapper[4754]: I0126 17:51:58.015286 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_19a0730c-46d0-4029-a86b-812ba1664dcc/object-server/0.log" Jan 26 17:51:58 crc kubenswrapper[4754]: I0126 17:51:58.048947 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_19a0730c-46d0-4029-a86b-812ba1664dcc/object-replicator/0.log" Jan 26 17:51:58 crc kubenswrapper[4754]: I0126 17:51:58.123731 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_19a0730c-46d0-4029-a86b-812ba1664dcc/object-updater/0.log" Jan 26 17:51:58 crc kubenswrapper[4754]: I0126 17:51:58.202943 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_19a0730c-46d0-4029-a86b-812ba1664dcc/rsync/0.log" Jan 26 17:51:58 crc kubenswrapper[4754]: I0126 17:51:58.289585 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_19a0730c-46d0-4029-a86b-812ba1664dcc/swift-recon-cron/0.log" Jan 26 17:51:58 crc kubenswrapper[4754]: I0126 17:51:58.430514 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-59vz4_5fade260-b289-4c35-b77e-ca9f8ec754c3/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Jan 26 17:51:58 crc kubenswrapper[4754]: I0126 17:51:58.523509 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_6f413c96-a0cb-4d11-9def-df8c3218db98/tempest-tests-tempest-tests-runner/0.log" Jan 26 17:51:58 crc kubenswrapper[4754]: I0126 17:51:58.702462 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_b323e46e-1ad6-41a9-9e7f-ad13e473f49e/test-operator-logs-container/0.log" Jan 26 17:51:58 crc kubenswrapper[4754]: I0126 17:51:58.733009 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-8vzzg_9d659728-5ede-4206-8137-8a6a62c5385c/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Jan 26 17:52:08 crc kubenswrapper[4754]: I0126 17:52:08.415472 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_6fbed1d4-2177-40ba-a3c6-03de6fc2484f/memcached/0.log" Jan 26 17:52:11 crc kubenswrapper[4754]: I0126 17:52:11.767723 4754 scope.go:117] "RemoveContainer" containerID="7b1482a92488015f91f13a86a1aa38c65e69fa9f874f9482c4d8f0f34463d2c2" Jan 26 17:52:11 crc kubenswrapper[4754]: E0126 17:52:11.768588 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x65wv_openshift-machine-config-operator(8c3718a4-f354-4284-92e0-fdfb45a692bd)\"" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" Jan 26 17:52:26 crc kubenswrapper[4754]: I0126 17:52:26.218222 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_80ea4f19923aee8b3bdba72b35be952869c3fe1246438cfe0124a675dcvvs8k_187f901c-e262-4c66-9104-7a40cb64d0e4/util/0.log" Jan 26 17:52:26 crc kubenswrapper[4754]: I0126 17:52:26.444107 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_80ea4f19923aee8b3bdba72b35be952869c3fe1246438cfe0124a675dcvvs8k_187f901c-e262-4c66-9104-7a40cb64d0e4/util/0.log" Jan 26 17:52:26 crc kubenswrapper[4754]: I0126 17:52:26.463405 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_80ea4f19923aee8b3bdba72b35be952869c3fe1246438cfe0124a675dcvvs8k_187f901c-e262-4c66-9104-7a40cb64d0e4/pull/0.log" Jan 26 17:52:26 crc kubenswrapper[4754]: I0126 17:52:26.475756 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_80ea4f19923aee8b3bdba72b35be952869c3fe1246438cfe0124a675dcvvs8k_187f901c-e262-4c66-9104-7a40cb64d0e4/pull/0.log" Jan 26 17:52:26 crc kubenswrapper[4754]: I0126 17:52:26.685400 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_80ea4f19923aee8b3bdba72b35be952869c3fe1246438cfe0124a675dcvvs8k_187f901c-e262-4c66-9104-7a40cb64d0e4/pull/0.log" Jan 26 17:52:26 crc kubenswrapper[4754]: I0126 17:52:26.720531 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_80ea4f19923aee8b3bdba72b35be952869c3fe1246438cfe0124a675dcvvs8k_187f901c-e262-4c66-9104-7a40cb64d0e4/extract/0.log" Jan 26 17:52:26 crc kubenswrapper[4754]: I0126 17:52:26.728487 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_80ea4f19923aee8b3bdba72b35be952869c3fe1246438cfe0124a675dcvvs8k_187f901c-e262-4c66-9104-7a40cb64d0e4/util/0.log" Jan 26 17:52:26 crc kubenswrapper[4754]: I0126 17:52:26.767350 4754 scope.go:117] "RemoveContainer" containerID="7b1482a92488015f91f13a86a1aa38c65e69fa9f874f9482c4d8f0f34463d2c2" Jan 26 17:52:26 crc kubenswrapper[4754]: E0126 17:52:26.767687 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x65wv_openshift-machine-config-operator(8c3718a4-f354-4284-92e0-fdfb45a692bd)\"" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" Jan 26 17:52:26 crc kubenswrapper[4754]: I0126 17:52:26.940207 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7f86f8796f-5kbpl_1a9a2d55-592f-4320-8e2e-49f65ca72dfc/manager/0.log" Jan 26 17:52:27 crc kubenswrapper[4754]: I0126 17:52:27.020030 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-7478f7dbf9-6vmhq_cb2628e4-680e-489a-8fc9-d39986c74301/manager/0.log" Jan 26 17:52:27 crc kubenswrapper[4754]: I0126 17:52:27.159936 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-b45d7bf98-bggfb_5e071de1-60dc-49d8-b965-90c2f99a6e02/manager/0.log" Jan 26 17:52:27 crc kubenswrapper[4754]: I0126 17:52:27.260762 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-78fdd796fd-c5kf9_ee98073e-2fdb-4b3e-acb8-00c71df55fa7/manager/0.log" Jan 26 17:52:27 crc kubenswrapper[4754]: I0126 17:52:27.492643 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-77d5c5b54f-8sz77_fd6b6ccf-b7e9-41fb-b663-a3392d075880/manager/0.log" Jan 26 17:52:27 crc kubenswrapper[4754]: I0126 17:52:27.512323 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-594c8c9d5d-hgnm9_f747ae1d-8181-4ef2-b332-b14db483aab6/manager/0.log" Jan 26 17:52:27 crc kubenswrapper[4754]: I0126 17:52:27.784211 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-598f7747c9-fvcgg_28161331-7731-433f-845f-2ebe1daf5fd0/manager/0.log" Jan 26 17:52:27 crc kubenswrapper[4754]: I0126 17:52:27.961022 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-694cf4f878-p66zm_5702a910-71d1-4acd-93aa-9379bc3147ce/manager/0.log" Jan 26 17:52:28 crc kubenswrapper[4754]: I0126 17:52:28.024549 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-b8b6d4659-d8db9_e3aac642-b3ee-4394-9f1d-bfac315bf162/manager/0.log" Jan 26 17:52:28 crc kubenswrapper[4754]: I0126 17:52:28.164608 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-78c6999f6f-jb5ps_16e535a9-3fc0-4385-b809-51a2bf719657/manager/0.log" Jan 26 17:52:28 crc kubenswrapper[4754]: I0126 17:52:28.275836 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-6b9fb5fdcb-fhwr4_add7fe6b-a864-439d-a3f4-80c9cc80ddc1/manager/0.log" Jan 26 17:52:28 crc kubenswrapper[4754]: I0126 17:52:28.404909 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-78d58447c5-pgmv6_2ab80530-a54c-41bf-8e4f-c895dbacb368/manager/0.log" Jan 26 17:52:28 crc kubenswrapper[4754]: I0126 17:52:28.525831 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-7bdb645866-27hg7_d23565a1-cb57-4cbe-88a2-48a01e8056f3/manager/0.log" Jan 26 17:52:28 crc kubenswrapper[4754]: I0126 17:52:28.651241 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-5f4cd88d46-fkz84_4fdc4d46-264a-4689-8ad5-0ed253f805df/manager/0.log" Jan 26 17:52:28 crc kubenswrapper[4754]: I0126 17:52:28.707016 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-6b68b8b854sclfc_f43cc0a3-41e9-4e09-837f-322a53221560/manager/0.log" Jan 26 17:52:28 crc kubenswrapper[4754]: I0126 17:52:28.972464 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-init-75cd685694-7x4dn_bfc14b7c-18c1-4442-bbb2-978889145894/operator/0.log" Jan 26 17:52:29 crc kubenswrapper[4754]: I0126 17:52:29.442889 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-zv866_39f92bf3-b95c-4950-ad79-05dedec74b02/registry-server/0.log" Jan 26 17:52:29 crc kubenswrapper[4754]: I0126 17:52:29.631041 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-6f75f45d54-kppds_8353f62d-eb53-46c3-ba4d-2d643f4f960b/manager/0.log" Jan 26 17:52:29 crc kubenswrapper[4754]: I0126 17:52:29.778960 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-79d5ccc684-gqtkb_56addd62-aeed-4139-a0db-37292a6acf8a/manager/0.log" Jan 26 17:52:29 crc kubenswrapper[4754]: I0126 17:52:29.952983 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-mqmkw_b48b5414-4034-4c0e-9d25-5053f84cf246/operator/0.log" Jan 26 17:52:30 crc kubenswrapper[4754]: I0126 17:52:30.174460 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-547cbdb99f-fmvdv_aaee86fe-018f-4055-b8c5-98e3795c53d8/manager/0.log" Jan 26 17:52:30 crc kubenswrapper[4754]: I0126 17:52:30.369958 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-85cd9769bb-dq4qw_13391684-794c-48b2-8d7f-23a122f8acc6/manager/0.log" Jan 26 17:52:30 crc kubenswrapper[4754]: I0126 17:52:30.507694 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-69797bbcbd-mr8bp_a8584d32-796d-47cd-8dd6-233374660688/manager/0.log" Jan 26 17:52:30 crc kubenswrapper[4754]: I0126 17:52:30.518237 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-59fcd4bdb5-gvt4r_93622668-0766-46f4-a216-83a7a17f36fc/manager/0.log" Jan 26 17:52:30 crc kubenswrapper[4754]: I0126 17:52:30.628168 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-564965969-k7nbl_138cc11d-6be6-43bb-994e-94ea39bb2e42/manager/0.log" Jan 26 17:52:37 crc kubenswrapper[4754]: I0126 17:52:37.046814 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-gf9bl"] Jan 26 17:52:37 crc kubenswrapper[4754]: E0126 17:52:37.047737 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2abb48e6-d767-43d1-a7e2-ea739fd9dd82" containerName="container-00" Jan 26 17:52:37 crc kubenswrapper[4754]: I0126 17:52:37.047751 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="2abb48e6-d767-43d1-a7e2-ea739fd9dd82" containerName="container-00" Jan 26 17:52:37 crc kubenswrapper[4754]: I0126 17:52:37.047950 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="2abb48e6-d767-43d1-a7e2-ea739fd9dd82" containerName="container-00" Jan 26 17:52:37 crc kubenswrapper[4754]: I0126 17:52:37.049232 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gf9bl" Jan 26 17:52:37 crc kubenswrapper[4754]: I0126 17:52:37.064619 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-gf9bl"] Jan 26 17:52:37 crc kubenswrapper[4754]: I0126 17:52:37.133012 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mdk6h\" (UniqueName: \"kubernetes.io/projected/f7c4d852-0b8e-415a-9296-5f4174f7ab22-kube-api-access-mdk6h\") pod \"certified-operators-gf9bl\" (UID: \"f7c4d852-0b8e-415a-9296-5f4174f7ab22\") " pod="openshift-marketplace/certified-operators-gf9bl" Jan 26 17:52:37 crc kubenswrapper[4754]: I0126 17:52:37.133078 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f7c4d852-0b8e-415a-9296-5f4174f7ab22-catalog-content\") pod \"certified-operators-gf9bl\" (UID: \"f7c4d852-0b8e-415a-9296-5f4174f7ab22\") " pod="openshift-marketplace/certified-operators-gf9bl" Jan 26 17:52:37 crc kubenswrapper[4754]: I0126 17:52:37.133179 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f7c4d852-0b8e-415a-9296-5f4174f7ab22-utilities\") pod \"certified-operators-gf9bl\" (UID: \"f7c4d852-0b8e-415a-9296-5f4174f7ab22\") " pod="openshift-marketplace/certified-operators-gf9bl" Jan 26 17:52:37 crc kubenswrapper[4754]: I0126 17:52:37.234814 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f7c4d852-0b8e-415a-9296-5f4174f7ab22-utilities\") pod \"certified-operators-gf9bl\" (UID: \"f7c4d852-0b8e-415a-9296-5f4174f7ab22\") " pod="openshift-marketplace/certified-operators-gf9bl" Jan 26 17:52:37 crc kubenswrapper[4754]: I0126 17:52:37.234963 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mdk6h\" (UniqueName: \"kubernetes.io/projected/f7c4d852-0b8e-415a-9296-5f4174f7ab22-kube-api-access-mdk6h\") pod \"certified-operators-gf9bl\" (UID: \"f7c4d852-0b8e-415a-9296-5f4174f7ab22\") " pod="openshift-marketplace/certified-operators-gf9bl" Jan 26 17:52:37 crc kubenswrapper[4754]: I0126 17:52:37.235005 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f7c4d852-0b8e-415a-9296-5f4174f7ab22-catalog-content\") pod \"certified-operators-gf9bl\" (UID: \"f7c4d852-0b8e-415a-9296-5f4174f7ab22\") " pod="openshift-marketplace/certified-operators-gf9bl" Jan 26 17:52:37 crc kubenswrapper[4754]: I0126 17:52:37.235438 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f7c4d852-0b8e-415a-9296-5f4174f7ab22-utilities\") pod \"certified-operators-gf9bl\" (UID: \"f7c4d852-0b8e-415a-9296-5f4174f7ab22\") " pod="openshift-marketplace/certified-operators-gf9bl" Jan 26 17:52:37 crc kubenswrapper[4754]: I0126 17:52:37.235598 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f7c4d852-0b8e-415a-9296-5f4174f7ab22-catalog-content\") pod \"certified-operators-gf9bl\" (UID: \"f7c4d852-0b8e-415a-9296-5f4174f7ab22\") " pod="openshift-marketplace/certified-operators-gf9bl" Jan 26 17:52:37 crc kubenswrapper[4754]: I0126 17:52:37.718648 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mdk6h\" (UniqueName: \"kubernetes.io/projected/f7c4d852-0b8e-415a-9296-5f4174f7ab22-kube-api-access-mdk6h\") pod \"certified-operators-gf9bl\" (UID: \"f7c4d852-0b8e-415a-9296-5f4174f7ab22\") " pod="openshift-marketplace/certified-operators-gf9bl" Jan 26 17:52:37 crc kubenswrapper[4754]: I0126 17:52:37.967545 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gf9bl" Jan 26 17:52:38 crc kubenswrapper[4754]: I0126 17:52:38.626282 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-gf9bl"] Jan 26 17:52:39 crc kubenswrapper[4754]: I0126 17:52:39.183529 4754 generic.go:334] "Generic (PLEG): container finished" podID="f7c4d852-0b8e-415a-9296-5f4174f7ab22" containerID="bf441910a9eff9376afdc147b340cffcf4e154cbd0b2c2be9e4969bebdc6a943" exitCode=0 Jan 26 17:52:39 crc kubenswrapper[4754]: I0126 17:52:39.183641 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gf9bl" event={"ID":"f7c4d852-0b8e-415a-9296-5f4174f7ab22","Type":"ContainerDied","Data":"bf441910a9eff9376afdc147b340cffcf4e154cbd0b2c2be9e4969bebdc6a943"} Jan 26 17:52:39 crc kubenswrapper[4754]: I0126 17:52:39.183900 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gf9bl" event={"ID":"f7c4d852-0b8e-415a-9296-5f4174f7ab22","Type":"ContainerStarted","Data":"a4226d4bac6dcbd872296c6f3e521673c22d493c337bc982095c5fbbeb6319b0"} Jan 26 17:52:39 crc kubenswrapper[4754]: I0126 17:52:39.243857 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-hxv65"] Jan 26 17:52:39 crc kubenswrapper[4754]: I0126 17:52:39.246530 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hxv65" Jan 26 17:52:39 crc kubenswrapper[4754]: I0126 17:52:39.267779 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-hxv65"] Jan 26 17:52:39 crc kubenswrapper[4754]: I0126 17:52:39.274831 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z9w6v\" (UniqueName: \"kubernetes.io/projected/883de6ad-32c9-4537-a029-ad0047b3b644-kube-api-access-z9w6v\") pod \"redhat-operators-hxv65\" (UID: \"883de6ad-32c9-4537-a029-ad0047b3b644\") " pod="openshift-marketplace/redhat-operators-hxv65" Jan 26 17:52:39 crc kubenswrapper[4754]: I0126 17:52:39.274893 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/883de6ad-32c9-4537-a029-ad0047b3b644-utilities\") pod \"redhat-operators-hxv65\" (UID: \"883de6ad-32c9-4537-a029-ad0047b3b644\") " pod="openshift-marketplace/redhat-operators-hxv65" Jan 26 17:52:39 crc kubenswrapper[4754]: I0126 17:52:39.274941 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/883de6ad-32c9-4537-a029-ad0047b3b644-catalog-content\") pod \"redhat-operators-hxv65\" (UID: \"883de6ad-32c9-4537-a029-ad0047b3b644\") " pod="openshift-marketplace/redhat-operators-hxv65" Jan 26 17:52:39 crc kubenswrapper[4754]: I0126 17:52:39.376610 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/883de6ad-32c9-4537-a029-ad0047b3b644-catalog-content\") pod \"redhat-operators-hxv65\" (UID: \"883de6ad-32c9-4537-a029-ad0047b3b644\") " pod="openshift-marketplace/redhat-operators-hxv65" Jan 26 17:52:39 crc kubenswrapper[4754]: I0126 17:52:39.376792 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z9w6v\" (UniqueName: \"kubernetes.io/projected/883de6ad-32c9-4537-a029-ad0047b3b644-kube-api-access-z9w6v\") pod \"redhat-operators-hxv65\" (UID: \"883de6ad-32c9-4537-a029-ad0047b3b644\") " pod="openshift-marketplace/redhat-operators-hxv65" Jan 26 17:52:39 crc kubenswrapper[4754]: I0126 17:52:39.376837 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/883de6ad-32c9-4537-a029-ad0047b3b644-utilities\") pod \"redhat-operators-hxv65\" (UID: \"883de6ad-32c9-4537-a029-ad0047b3b644\") " pod="openshift-marketplace/redhat-operators-hxv65" Jan 26 17:52:39 crc kubenswrapper[4754]: I0126 17:52:39.377327 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/883de6ad-32c9-4537-a029-ad0047b3b644-utilities\") pod \"redhat-operators-hxv65\" (UID: \"883de6ad-32c9-4537-a029-ad0047b3b644\") " pod="openshift-marketplace/redhat-operators-hxv65" Jan 26 17:52:39 crc kubenswrapper[4754]: I0126 17:52:39.377444 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/883de6ad-32c9-4537-a029-ad0047b3b644-catalog-content\") pod \"redhat-operators-hxv65\" (UID: \"883de6ad-32c9-4537-a029-ad0047b3b644\") " pod="openshift-marketplace/redhat-operators-hxv65" Jan 26 17:52:39 crc kubenswrapper[4754]: I0126 17:52:39.729185 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z9w6v\" (UniqueName: \"kubernetes.io/projected/883de6ad-32c9-4537-a029-ad0047b3b644-kube-api-access-z9w6v\") pod \"redhat-operators-hxv65\" (UID: \"883de6ad-32c9-4537-a029-ad0047b3b644\") " pod="openshift-marketplace/redhat-operators-hxv65" Jan 26 17:52:39 crc kubenswrapper[4754]: I0126 17:52:39.872027 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hxv65" Jan 26 17:52:40 crc kubenswrapper[4754]: I0126 17:52:40.213586 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gf9bl" event={"ID":"f7c4d852-0b8e-415a-9296-5f4174f7ab22","Type":"ContainerStarted","Data":"b51888acd608d13e86ff8911743337d1b22e27727a77f3ff85df7e157bf62e66"} Jan 26 17:52:40 crc kubenswrapper[4754]: I0126 17:52:40.428910 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-hxv65"] Jan 26 17:52:40 crc kubenswrapper[4754]: W0126 17:52:40.434818 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod883de6ad_32c9_4537_a029_ad0047b3b644.slice/crio-a0b582672dcf161893f6b77fc487d1c694138e26cf8d809b20fc1163f4f2474d WatchSource:0}: Error finding container a0b582672dcf161893f6b77fc487d1c694138e26cf8d809b20fc1163f4f2474d: Status 404 returned error can't find the container with id a0b582672dcf161893f6b77fc487d1c694138e26cf8d809b20fc1163f4f2474d Jan 26 17:52:41 crc kubenswrapper[4754]: I0126 17:52:41.225212 4754 generic.go:334] "Generic (PLEG): container finished" podID="883de6ad-32c9-4537-a029-ad0047b3b644" containerID="eeab07142089a538b849f37ab22e4aa0258f8afc32b2c0557d6131f029cd7888" exitCode=0 Jan 26 17:52:41 crc kubenswrapper[4754]: I0126 17:52:41.225309 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hxv65" event={"ID":"883de6ad-32c9-4537-a029-ad0047b3b644","Type":"ContainerDied","Data":"eeab07142089a538b849f37ab22e4aa0258f8afc32b2c0557d6131f029cd7888"} Jan 26 17:52:41 crc kubenswrapper[4754]: I0126 17:52:41.225343 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hxv65" event={"ID":"883de6ad-32c9-4537-a029-ad0047b3b644","Type":"ContainerStarted","Data":"a0b582672dcf161893f6b77fc487d1c694138e26cf8d809b20fc1163f4f2474d"} Jan 26 17:52:41 crc kubenswrapper[4754]: I0126 17:52:41.228514 4754 generic.go:334] "Generic (PLEG): container finished" podID="f7c4d852-0b8e-415a-9296-5f4174f7ab22" containerID="b51888acd608d13e86ff8911743337d1b22e27727a77f3ff85df7e157bf62e66" exitCode=0 Jan 26 17:52:41 crc kubenswrapper[4754]: I0126 17:52:41.228567 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gf9bl" event={"ID":"f7c4d852-0b8e-415a-9296-5f4174f7ab22","Type":"ContainerDied","Data":"b51888acd608d13e86ff8911743337d1b22e27727a77f3ff85df7e157bf62e66"} Jan 26 17:52:41 crc kubenswrapper[4754]: I0126 17:52:41.767652 4754 scope.go:117] "RemoveContainer" containerID="7b1482a92488015f91f13a86a1aa38c65e69fa9f874f9482c4d8f0f34463d2c2" Jan 26 17:52:41 crc kubenswrapper[4754]: E0126 17:52:41.768265 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x65wv_openshift-machine-config-operator(8c3718a4-f354-4284-92e0-fdfb45a692bd)\"" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" Jan 26 17:52:42 crc kubenswrapper[4754]: I0126 17:52:42.252796 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gf9bl" event={"ID":"f7c4d852-0b8e-415a-9296-5f4174f7ab22","Type":"ContainerStarted","Data":"2498c64a345ef6da10ea4e5ab09a667f0314a3751e9ceb4e0d5dd2d337f92a0d"} Jan 26 17:52:42 crc kubenswrapper[4754]: I0126 17:52:42.274256 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-gf9bl" podStartSLOduration=2.666307869 podStartE2EDuration="5.27423276s" podCreationTimestamp="2026-01-26 17:52:37 +0000 UTC" firstStartedPulling="2026-01-26 17:52:39.185806965 +0000 UTC m=+3925.709987399" lastFinishedPulling="2026-01-26 17:52:41.793731856 +0000 UTC m=+3928.317912290" observedRunningTime="2026-01-26 17:52:42.270684252 +0000 UTC m=+3928.794864716" watchObservedRunningTime="2026-01-26 17:52:42.27423276 +0000 UTC m=+3928.798413194" Jan 26 17:52:43 crc kubenswrapper[4754]: I0126 17:52:43.272060 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hxv65" event={"ID":"883de6ad-32c9-4537-a029-ad0047b3b644","Type":"ContainerStarted","Data":"9232bcb2fcf870ba117c65be7483359d10ef2f6e48d8d48006d6445753f4a10c"} Jan 26 17:52:45 crc kubenswrapper[4754]: I0126 17:52:45.288565 4754 generic.go:334] "Generic (PLEG): container finished" podID="883de6ad-32c9-4537-a029-ad0047b3b644" containerID="9232bcb2fcf870ba117c65be7483359d10ef2f6e48d8d48006d6445753f4a10c" exitCode=0 Jan 26 17:52:45 crc kubenswrapper[4754]: I0126 17:52:45.289099 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hxv65" event={"ID":"883de6ad-32c9-4537-a029-ad0047b3b644","Type":"ContainerDied","Data":"9232bcb2fcf870ba117c65be7483359d10ef2f6e48d8d48006d6445753f4a10c"} Jan 26 17:52:47 crc kubenswrapper[4754]: I0126 17:52:47.310276 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hxv65" event={"ID":"883de6ad-32c9-4537-a029-ad0047b3b644","Type":"ContainerStarted","Data":"82f56bb1c0eafd6668337afb7825307c751652945bb64815f797ca3116d0fac1"} Jan 26 17:52:47 crc kubenswrapper[4754]: I0126 17:52:47.332969 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-hxv65" podStartSLOduration=2.626687083 podStartE2EDuration="8.332951031s" podCreationTimestamp="2026-01-26 17:52:39 +0000 UTC" firstStartedPulling="2026-01-26 17:52:41.227537492 +0000 UTC m=+3927.751717926" lastFinishedPulling="2026-01-26 17:52:46.93380144 +0000 UTC m=+3933.457981874" observedRunningTime="2026-01-26 17:52:47.33145118 +0000 UTC m=+3933.855631614" watchObservedRunningTime="2026-01-26 17:52:47.332951031 +0000 UTC m=+3933.857131455" Jan 26 17:52:47 crc kubenswrapper[4754]: I0126 17:52:47.968521 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-gf9bl" Jan 26 17:52:47 crc kubenswrapper[4754]: I0126 17:52:47.968589 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-gf9bl" Jan 26 17:52:48 crc kubenswrapper[4754]: I0126 17:52:48.030480 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-gf9bl" Jan 26 17:52:48 crc kubenswrapper[4754]: I0126 17:52:48.656781 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-gf9bl" Jan 26 17:52:49 crc kubenswrapper[4754]: I0126 17:52:49.634095 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-gf9bl"] Jan 26 17:52:49 crc kubenswrapper[4754]: I0126 17:52:49.872742 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-hxv65" Jan 26 17:52:49 crc kubenswrapper[4754]: I0126 17:52:49.872824 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-hxv65" Jan 26 17:52:50 crc kubenswrapper[4754]: I0126 17:52:50.336143 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-gf9bl" podUID="f7c4d852-0b8e-415a-9296-5f4174f7ab22" containerName="registry-server" containerID="cri-o://2498c64a345ef6da10ea4e5ab09a667f0314a3751e9ceb4e0d5dd2d337f92a0d" gracePeriod=2 Jan 26 17:52:51 crc kubenswrapper[4754]: I0126 17:52:51.033914 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gf9bl" Jan 26 17:52:51 crc kubenswrapper[4754]: I0126 17:52:51.214749 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mdk6h\" (UniqueName: \"kubernetes.io/projected/f7c4d852-0b8e-415a-9296-5f4174f7ab22-kube-api-access-mdk6h\") pod \"f7c4d852-0b8e-415a-9296-5f4174f7ab22\" (UID: \"f7c4d852-0b8e-415a-9296-5f4174f7ab22\") " Jan 26 17:52:51 crc kubenswrapper[4754]: I0126 17:52:51.214930 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f7c4d852-0b8e-415a-9296-5f4174f7ab22-utilities\") pod \"f7c4d852-0b8e-415a-9296-5f4174f7ab22\" (UID: \"f7c4d852-0b8e-415a-9296-5f4174f7ab22\") " Jan 26 17:52:51 crc kubenswrapper[4754]: I0126 17:52:51.214978 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f7c4d852-0b8e-415a-9296-5f4174f7ab22-catalog-content\") pod \"f7c4d852-0b8e-415a-9296-5f4174f7ab22\" (UID: \"f7c4d852-0b8e-415a-9296-5f4174f7ab22\") " Jan 26 17:52:51 crc kubenswrapper[4754]: I0126 17:52:51.217144 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f7c4d852-0b8e-415a-9296-5f4174f7ab22-utilities" (OuterVolumeSpecName: "utilities") pod "f7c4d852-0b8e-415a-9296-5f4174f7ab22" (UID: "f7c4d852-0b8e-415a-9296-5f4174f7ab22"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:52:51 crc kubenswrapper[4754]: I0126 17:52:51.222541 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f7c4d852-0b8e-415a-9296-5f4174f7ab22-kube-api-access-mdk6h" (OuterVolumeSpecName: "kube-api-access-mdk6h") pod "f7c4d852-0b8e-415a-9296-5f4174f7ab22" (UID: "f7c4d852-0b8e-415a-9296-5f4174f7ab22"). InnerVolumeSpecName "kube-api-access-mdk6h". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:52:51 crc kubenswrapper[4754]: I0126 17:52:51.256159 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f7c4d852-0b8e-415a-9296-5f4174f7ab22-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f7c4d852-0b8e-415a-9296-5f4174f7ab22" (UID: "f7c4d852-0b8e-415a-9296-5f4174f7ab22"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:52:51 crc kubenswrapper[4754]: I0126 17:52:51.319158 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mdk6h\" (UniqueName: \"kubernetes.io/projected/f7c4d852-0b8e-415a-9296-5f4174f7ab22-kube-api-access-mdk6h\") on node \"crc\" DevicePath \"\"" Jan 26 17:52:51 crc kubenswrapper[4754]: I0126 17:52:51.319191 4754 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f7c4d852-0b8e-415a-9296-5f4174f7ab22-utilities\") on node \"crc\" DevicePath \"\"" Jan 26 17:52:51 crc kubenswrapper[4754]: I0126 17:52:51.319204 4754 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f7c4d852-0b8e-415a-9296-5f4174f7ab22-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 26 17:52:51 crc kubenswrapper[4754]: I0126 17:52:51.346423 4754 generic.go:334] "Generic (PLEG): container finished" podID="f7c4d852-0b8e-415a-9296-5f4174f7ab22" containerID="2498c64a345ef6da10ea4e5ab09a667f0314a3751e9ceb4e0d5dd2d337f92a0d" exitCode=0 Jan 26 17:52:51 crc kubenswrapper[4754]: I0126 17:52:51.346491 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gf9bl" Jan 26 17:52:51 crc kubenswrapper[4754]: I0126 17:52:51.346496 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gf9bl" event={"ID":"f7c4d852-0b8e-415a-9296-5f4174f7ab22","Type":"ContainerDied","Data":"2498c64a345ef6da10ea4e5ab09a667f0314a3751e9ceb4e0d5dd2d337f92a0d"} Jan 26 17:52:51 crc kubenswrapper[4754]: I0126 17:52:51.346660 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gf9bl" event={"ID":"f7c4d852-0b8e-415a-9296-5f4174f7ab22","Type":"ContainerDied","Data":"a4226d4bac6dcbd872296c6f3e521673c22d493c337bc982095c5fbbeb6319b0"} Jan 26 17:52:51 crc kubenswrapper[4754]: I0126 17:52:51.346706 4754 scope.go:117] "RemoveContainer" containerID="2498c64a345ef6da10ea4e5ab09a667f0314a3751e9ceb4e0d5dd2d337f92a0d" Jan 26 17:52:51 crc kubenswrapper[4754]: I0126 17:52:51.363539 4754 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-hxv65" podUID="883de6ad-32c9-4537-a029-ad0047b3b644" containerName="registry-server" probeResult="failure" output=< Jan 26 17:52:51 crc kubenswrapper[4754]: timeout: failed to connect service ":50051" within 1s Jan 26 17:52:51 crc kubenswrapper[4754]: > Jan 26 17:52:51 crc kubenswrapper[4754]: I0126 17:52:51.368243 4754 scope.go:117] "RemoveContainer" containerID="b51888acd608d13e86ff8911743337d1b22e27727a77f3ff85df7e157bf62e66" Jan 26 17:52:51 crc kubenswrapper[4754]: I0126 17:52:51.388388 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-gf9bl"] Jan 26 17:52:51 crc kubenswrapper[4754]: I0126 17:52:51.395845 4754 scope.go:117] "RemoveContainer" containerID="bf441910a9eff9376afdc147b340cffcf4e154cbd0b2c2be9e4969bebdc6a943" Jan 26 17:52:51 crc kubenswrapper[4754]: I0126 17:52:51.396972 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-gf9bl"] Jan 26 17:52:51 crc kubenswrapper[4754]: I0126 17:52:51.455215 4754 scope.go:117] "RemoveContainer" containerID="2498c64a345ef6da10ea4e5ab09a667f0314a3751e9ceb4e0d5dd2d337f92a0d" Jan 26 17:52:51 crc kubenswrapper[4754]: E0126 17:52:51.455693 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2498c64a345ef6da10ea4e5ab09a667f0314a3751e9ceb4e0d5dd2d337f92a0d\": container with ID starting with 2498c64a345ef6da10ea4e5ab09a667f0314a3751e9ceb4e0d5dd2d337f92a0d not found: ID does not exist" containerID="2498c64a345ef6da10ea4e5ab09a667f0314a3751e9ceb4e0d5dd2d337f92a0d" Jan 26 17:52:51 crc kubenswrapper[4754]: I0126 17:52:51.455732 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2498c64a345ef6da10ea4e5ab09a667f0314a3751e9ceb4e0d5dd2d337f92a0d"} err="failed to get container status \"2498c64a345ef6da10ea4e5ab09a667f0314a3751e9ceb4e0d5dd2d337f92a0d\": rpc error: code = NotFound desc = could not find container \"2498c64a345ef6da10ea4e5ab09a667f0314a3751e9ceb4e0d5dd2d337f92a0d\": container with ID starting with 2498c64a345ef6da10ea4e5ab09a667f0314a3751e9ceb4e0d5dd2d337f92a0d not found: ID does not exist" Jan 26 17:52:51 crc kubenswrapper[4754]: I0126 17:52:51.455755 4754 scope.go:117] "RemoveContainer" containerID="b51888acd608d13e86ff8911743337d1b22e27727a77f3ff85df7e157bf62e66" Jan 26 17:52:51 crc kubenswrapper[4754]: E0126 17:52:51.456038 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b51888acd608d13e86ff8911743337d1b22e27727a77f3ff85df7e157bf62e66\": container with ID starting with b51888acd608d13e86ff8911743337d1b22e27727a77f3ff85df7e157bf62e66 not found: ID does not exist" containerID="b51888acd608d13e86ff8911743337d1b22e27727a77f3ff85df7e157bf62e66" Jan 26 17:52:51 crc kubenswrapper[4754]: I0126 17:52:51.456065 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b51888acd608d13e86ff8911743337d1b22e27727a77f3ff85df7e157bf62e66"} err="failed to get container status \"b51888acd608d13e86ff8911743337d1b22e27727a77f3ff85df7e157bf62e66\": rpc error: code = NotFound desc = could not find container \"b51888acd608d13e86ff8911743337d1b22e27727a77f3ff85df7e157bf62e66\": container with ID starting with b51888acd608d13e86ff8911743337d1b22e27727a77f3ff85df7e157bf62e66 not found: ID does not exist" Jan 26 17:52:51 crc kubenswrapper[4754]: I0126 17:52:51.456087 4754 scope.go:117] "RemoveContainer" containerID="bf441910a9eff9376afdc147b340cffcf4e154cbd0b2c2be9e4969bebdc6a943" Jan 26 17:52:51 crc kubenswrapper[4754]: E0126 17:52:51.456805 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bf441910a9eff9376afdc147b340cffcf4e154cbd0b2c2be9e4969bebdc6a943\": container with ID starting with bf441910a9eff9376afdc147b340cffcf4e154cbd0b2c2be9e4969bebdc6a943 not found: ID does not exist" containerID="bf441910a9eff9376afdc147b340cffcf4e154cbd0b2c2be9e4969bebdc6a943" Jan 26 17:52:51 crc kubenswrapper[4754]: I0126 17:52:51.456843 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bf441910a9eff9376afdc147b340cffcf4e154cbd0b2c2be9e4969bebdc6a943"} err="failed to get container status \"bf441910a9eff9376afdc147b340cffcf4e154cbd0b2c2be9e4969bebdc6a943\": rpc error: code = NotFound desc = could not find container \"bf441910a9eff9376afdc147b340cffcf4e154cbd0b2c2be9e4969bebdc6a943\": container with ID starting with bf441910a9eff9376afdc147b340cffcf4e154cbd0b2c2be9e4969bebdc6a943 not found: ID does not exist" Jan 26 17:52:51 crc kubenswrapper[4754]: I0126 17:52:51.778975 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f7c4d852-0b8e-415a-9296-5f4174f7ab22" path="/var/lib/kubelet/pods/f7c4d852-0b8e-415a-9296-5f4174f7ab22/volumes" Jan 26 17:52:53 crc kubenswrapper[4754]: I0126 17:52:53.108409 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-w7zlr_a411958e-77f0-45bb-a11c-abd88d756dee/control-plane-machine-set-operator/0.log" Jan 26 17:52:53 crc kubenswrapper[4754]: I0126 17:52:53.333864 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-g64xp_5864769c-0a13-4562-a407-f67e4348e1a6/machine-api-operator/0.log" Jan 26 17:52:53 crc kubenswrapper[4754]: I0126 17:52:53.371382 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-g64xp_5864769c-0a13-4562-a407-f67e4348e1a6/kube-rbac-proxy/0.log" Jan 26 17:52:54 crc kubenswrapper[4754]: I0126 17:52:54.767595 4754 scope.go:117] "RemoveContainer" containerID="7b1482a92488015f91f13a86a1aa38c65e69fa9f874f9482c4d8f0f34463d2c2" Jan 26 17:52:54 crc kubenswrapper[4754]: E0126 17:52:54.767872 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x65wv_openshift-machine-config-operator(8c3718a4-f354-4284-92e0-fdfb45a692bd)\"" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" Jan 26 17:53:00 crc kubenswrapper[4754]: I0126 17:53:00.462294 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-hxv65" Jan 26 17:53:00 crc kubenswrapper[4754]: I0126 17:53:00.523276 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-hxv65" Jan 26 17:53:00 crc kubenswrapper[4754]: I0126 17:53:00.704214 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-hxv65"] Jan 26 17:53:02 crc kubenswrapper[4754]: I0126 17:53:02.444429 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-hxv65" podUID="883de6ad-32c9-4537-a029-ad0047b3b644" containerName="registry-server" containerID="cri-o://82f56bb1c0eafd6668337afb7825307c751652945bb64815f797ca3116d0fac1" gracePeriod=2 Jan 26 17:53:03 crc kubenswrapper[4754]: I0126 17:53:03.441579 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hxv65" Jan 26 17:53:03 crc kubenswrapper[4754]: I0126 17:53:03.454570 4754 generic.go:334] "Generic (PLEG): container finished" podID="883de6ad-32c9-4537-a029-ad0047b3b644" containerID="82f56bb1c0eafd6668337afb7825307c751652945bb64815f797ca3116d0fac1" exitCode=0 Jan 26 17:53:03 crc kubenswrapper[4754]: I0126 17:53:03.454610 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hxv65" event={"ID":"883de6ad-32c9-4537-a029-ad0047b3b644","Type":"ContainerDied","Data":"82f56bb1c0eafd6668337afb7825307c751652945bb64815f797ca3116d0fac1"} Jan 26 17:53:03 crc kubenswrapper[4754]: I0126 17:53:03.454635 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hxv65" event={"ID":"883de6ad-32c9-4537-a029-ad0047b3b644","Type":"ContainerDied","Data":"a0b582672dcf161893f6b77fc487d1c694138e26cf8d809b20fc1163f4f2474d"} Jan 26 17:53:03 crc kubenswrapper[4754]: I0126 17:53:03.454652 4754 scope.go:117] "RemoveContainer" containerID="82f56bb1c0eafd6668337afb7825307c751652945bb64815f797ca3116d0fac1" Jan 26 17:53:03 crc kubenswrapper[4754]: I0126 17:53:03.454795 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hxv65" Jan 26 17:53:03 crc kubenswrapper[4754]: I0126 17:53:03.486130 4754 scope.go:117] "RemoveContainer" containerID="9232bcb2fcf870ba117c65be7483359d10ef2f6e48d8d48006d6445753f4a10c" Jan 26 17:53:03 crc kubenswrapper[4754]: I0126 17:53:03.510791 4754 scope.go:117] "RemoveContainer" containerID="eeab07142089a538b849f37ab22e4aa0258f8afc32b2c0557d6131f029cd7888" Jan 26 17:53:03 crc kubenswrapper[4754]: I0126 17:53:03.544543 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/883de6ad-32c9-4537-a029-ad0047b3b644-catalog-content\") pod \"883de6ad-32c9-4537-a029-ad0047b3b644\" (UID: \"883de6ad-32c9-4537-a029-ad0047b3b644\") " Jan 26 17:53:03 crc kubenswrapper[4754]: I0126 17:53:03.544698 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z9w6v\" (UniqueName: \"kubernetes.io/projected/883de6ad-32c9-4537-a029-ad0047b3b644-kube-api-access-z9w6v\") pod \"883de6ad-32c9-4537-a029-ad0047b3b644\" (UID: \"883de6ad-32c9-4537-a029-ad0047b3b644\") " Jan 26 17:53:03 crc kubenswrapper[4754]: I0126 17:53:03.544767 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/883de6ad-32c9-4537-a029-ad0047b3b644-utilities\") pod \"883de6ad-32c9-4537-a029-ad0047b3b644\" (UID: \"883de6ad-32c9-4537-a029-ad0047b3b644\") " Jan 26 17:53:03 crc kubenswrapper[4754]: I0126 17:53:03.545636 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/883de6ad-32c9-4537-a029-ad0047b3b644-utilities" (OuterVolumeSpecName: "utilities") pod "883de6ad-32c9-4537-a029-ad0047b3b644" (UID: "883de6ad-32c9-4537-a029-ad0047b3b644"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:53:03 crc kubenswrapper[4754]: I0126 17:53:03.551298 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/883de6ad-32c9-4537-a029-ad0047b3b644-kube-api-access-z9w6v" (OuterVolumeSpecName: "kube-api-access-z9w6v") pod "883de6ad-32c9-4537-a029-ad0047b3b644" (UID: "883de6ad-32c9-4537-a029-ad0047b3b644"). InnerVolumeSpecName "kube-api-access-z9w6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:53:03 crc kubenswrapper[4754]: I0126 17:53:03.564637 4754 scope.go:117] "RemoveContainer" containerID="82f56bb1c0eafd6668337afb7825307c751652945bb64815f797ca3116d0fac1" Jan 26 17:53:03 crc kubenswrapper[4754]: E0126 17:53:03.565479 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"82f56bb1c0eafd6668337afb7825307c751652945bb64815f797ca3116d0fac1\": container with ID starting with 82f56bb1c0eafd6668337afb7825307c751652945bb64815f797ca3116d0fac1 not found: ID does not exist" containerID="82f56bb1c0eafd6668337afb7825307c751652945bb64815f797ca3116d0fac1" Jan 26 17:53:03 crc kubenswrapper[4754]: I0126 17:53:03.565518 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"82f56bb1c0eafd6668337afb7825307c751652945bb64815f797ca3116d0fac1"} err="failed to get container status \"82f56bb1c0eafd6668337afb7825307c751652945bb64815f797ca3116d0fac1\": rpc error: code = NotFound desc = could not find container \"82f56bb1c0eafd6668337afb7825307c751652945bb64815f797ca3116d0fac1\": container with ID starting with 82f56bb1c0eafd6668337afb7825307c751652945bb64815f797ca3116d0fac1 not found: ID does not exist" Jan 26 17:53:03 crc kubenswrapper[4754]: I0126 17:53:03.565543 4754 scope.go:117] "RemoveContainer" containerID="9232bcb2fcf870ba117c65be7483359d10ef2f6e48d8d48006d6445753f4a10c" Jan 26 17:53:03 crc kubenswrapper[4754]: E0126 17:53:03.565907 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9232bcb2fcf870ba117c65be7483359d10ef2f6e48d8d48006d6445753f4a10c\": container with ID starting with 9232bcb2fcf870ba117c65be7483359d10ef2f6e48d8d48006d6445753f4a10c not found: ID does not exist" containerID="9232bcb2fcf870ba117c65be7483359d10ef2f6e48d8d48006d6445753f4a10c" Jan 26 17:53:03 crc kubenswrapper[4754]: I0126 17:53:03.565959 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9232bcb2fcf870ba117c65be7483359d10ef2f6e48d8d48006d6445753f4a10c"} err="failed to get container status \"9232bcb2fcf870ba117c65be7483359d10ef2f6e48d8d48006d6445753f4a10c\": rpc error: code = NotFound desc = could not find container \"9232bcb2fcf870ba117c65be7483359d10ef2f6e48d8d48006d6445753f4a10c\": container with ID starting with 9232bcb2fcf870ba117c65be7483359d10ef2f6e48d8d48006d6445753f4a10c not found: ID does not exist" Jan 26 17:53:03 crc kubenswrapper[4754]: I0126 17:53:03.565985 4754 scope.go:117] "RemoveContainer" containerID="eeab07142089a538b849f37ab22e4aa0258f8afc32b2c0557d6131f029cd7888" Jan 26 17:53:03 crc kubenswrapper[4754]: E0126 17:53:03.566290 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eeab07142089a538b849f37ab22e4aa0258f8afc32b2c0557d6131f029cd7888\": container with ID starting with eeab07142089a538b849f37ab22e4aa0258f8afc32b2c0557d6131f029cd7888 not found: ID does not exist" containerID="eeab07142089a538b849f37ab22e4aa0258f8afc32b2c0557d6131f029cd7888" Jan 26 17:53:03 crc kubenswrapper[4754]: I0126 17:53:03.566312 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eeab07142089a538b849f37ab22e4aa0258f8afc32b2c0557d6131f029cd7888"} err="failed to get container status \"eeab07142089a538b849f37ab22e4aa0258f8afc32b2c0557d6131f029cd7888\": rpc error: code = NotFound desc = could not find container \"eeab07142089a538b849f37ab22e4aa0258f8afc32b2c0557d6131f029cd7888\": container with ID starting with eeab07142089a538b849f37ab22e4aa0258f8afc32b2c0557d6131f029cd7888 not found: ID does not exist" Jan 26 17:53:03 crc kubenswrapper[4754]: I0126 17:53:03.646961 4754 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/883de6ad-32c9-4537-a029-ad0047b3b644-utilities\") on node \"crc\" DevicePath \"\"" Jan 26 17:53:03 crc kubenswrapper[4754]: I0126 17:53:03.647282 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z9w6v\" (UniqueName: \"kubernetes.io/projected/883de6ad-32c9-4537-a029-ad0047b3b644-kube-api-access-z9w6v\") on node \"crc\" DevicePath \"\"" Jan 26 17:53:03 crc kubenswrapper[4754]: I0126 17:53:03.657055 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/883de6ad-32c9-4537-a029-ad0047b3b644-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "883de6ad-32c9-4537-a029-ad0047b3b644" (UID: "883de6ad-32c9-4537-a029-ad0047b3b644"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:53:03 crc kubenswrapper[4754]: I0126 17:53:03.749070 4754 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/883de6ad-32c9-4537-a029-ad0047b3b644-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 26 17:53:03 crc kubenswrapper[4754]: I0126 17:53:03.803549 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-hxv65"] Jan 26 17:53:03 crc kubenswrapper[4754]: I0126 17:53:03.817280 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-hxv65"] Jan 26 17:53:05 crc kubenswrapper[4754]: I0126 17:53:05.792268 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="883de6ad-32c9-4537-a029-ad0047b3b644" path="/var/lib/kubelet/pods/883de6ad-32c9-4537-a029-ad0047b3b644/volumes" Jan 26 17:53:06 crc kubenswrapper[4754]: I0126 17:53:06.721341 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-858654f9db-5n7pf_b2d0aba8-8dab-483d-87cd-756af280d526/cert-manager-controller/0.log" Jan 26 17:53:06 crc kubenswrapper[4754]: I0126 17:53:06.843794 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-cf98fcc89-mksh5_9bddf956-836c-41ef-9038-f889e71b6823/cert-manager-cainjector/0.log" Jan 26 17:53:06 crc kubenswrapper[4754]: I0126 17:53:06.914453 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-687f57d79b-q26n7_d46f02a6-da6f-4128-89e8-669cdb8622c6/cert-manager-webhook/0.log" Jan 26 17:53:09 crc kubenswrapper[4754]: I0126 17:53:09.768019 4754 scope.go:117] "RemoveContainer" containerID="7b1482a92488015f91f13a86a1aa38c65e69fa9f874f9482c4d8f0f34463d2c2" Jan 26 17:53:09 crc kubenswrapper[4754]: E0126 17:53:09.769875 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x65wv_openshift-machine-config-operator(8c3718a4-f354-4284-92e0-fdfb45a692bd)\"" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" Jan 26 17:53:20 crc kubenswrapper[4754]: I0126 17:53:20.069237 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-7754f76f8b-dxcqw_97017e08-dd1d-494a-b4bf-0800824588a2/nmstate-console-plugin/0.log" Jan 26 17:53:20 crc kubenswrapper[4754]: I0126 17:53:20.512957 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-54757c584b-vvxrw_6506a03e-2445-4e3c-9814-f5bb1cfbaa06/nmstate-metrics/0.log" Jan 26 17:53:20 crc kubenswrapper[4754]: I0126 17:53:20.525465 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-54757c584b-vvxrw_6506a03e-2445-4e3c-9814-f5bb1cfbaa06/kube-rbac-proxy/0.log" Jan 26 17:53:20 crc kubenswrapper[4754]: I0126 17:53:20.540904 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-qfmmc_73c39dd3-4d2e-4ee8-a218-f88eb59f36b7/nmstate-handler/0.log" Jan 26 17:53:20 crc kubenswrapper[4754]: I0126 17:53:20.896435 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-8474b5b9d8-zgmpc_fa03d76c-cc42-4ba9-ad6f-671b4b63dbab/nmstate-webhook/0.log" Jan 26 17:53:20 crc kubenswrapper[4754]: I0126 17:53:20.929706 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-646758c888-pwgfk_9bbc5233-67ae-4cca-9a95-71da7e373005/nmstate-operator/0.log" Jan 26 17:53:21 crc kubenswrapper[4754]: I0126 17:53:21.768061 4754 scope.go:117] "RemoveContainer" containerID="7b1482a92488015f91f13a86a1aa38c65e69fa9f874f9482c4d8f0f34463d2c2" Jan 26 17:53:21 crc kubenswrapper[4754]: E0126 17:53:21.768343 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x65wv_openshift-machine-config-operator(8c3718a4-f354-4284-92e0-fdfb45a692bd)\"" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" Jan 26 17:53:32 crc kubenswrapper[4754]: I0126 17:53:32.768736 4754 scope.go:117] "RemoveContainer" containerID="7b1482a92488015f91f13a86a1aa38c65e69fa9f874f9482c4d8f0f34463d2c2" Jan 26 17:53:32 crc kubenswrapper[4754]: E0126 17:53:32.769709 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x65wv_openshift-machine-config-operator(8c3718a4-f354-4284-92e0-fdfb45a692bd)\"" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" Jan 26 17:53:46 crc kubenswrapper[4754]: I0126 17:53:46.767964 4754 scope.go:117] "RemoveContainer" containerID="7b1482a92488015f91f13a86a1aa38c65e69fa9f874f9482c4d8f0f34463d2c2" Jan 26 17:53:47 crc kubenswrapper[4754]: I0126 17:53:47.572585 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6968d8fdc4-8458x_dadf6410-9227-422a-8954-a2488091b15e/kube-rbac-proxy/0.log" Jan 26 17:53:47 crc kubenswrapper[4754]: I0126 17:53:47.788423 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6968d8fdc4-8458x_dadf6410-9227-422a-8954-a2488091b15e/controller/0.log" Jan 26 17:53:47 crc kubenswrapper[4754]: I0126 17:53:47.800431 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5h7fb_9f0c3994-3669-4cb5-89b3-b9e9dbc316de/cp-frr-files/0.log" Jan 26 17:53:47 crc kubenswrapper[4754]: I0126 17:53:47.853912 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" event={"ID":"8c3718a4-f354-4284-92e0-fdfb45a692bd","Type":"ContainerStarted","Data":"25b9a1afdbf7b01e5200dd76d80daf1edca89288d2a1ec786b655b6ff7ce5a88"} Jan 26 17:53:48 crc kubenswrapper[4754]: I0126 17:53:48.031730 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5h7fb_9f0c3994-3669-4cb5-89b3-b9e9dbc316de/cp-metrics/0.log" Jan 26 17:53:48 crc kubenswrapper[4754]: I0126 17:53:48.046131 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5h7fb_9f0c3994-3669-4cb5-89b3-b9e9dbc316de/cp-reloader/0.log" Jan 26 17:53:48 crc kubenswrapper[4754]: I0126 17:53:48.076855 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5h7fb_9f0c3994-3669-4cb5-89b3-b9e9dbc316de/cp-reloader/0.log" Jan 26 17:53:48 crc kubenswrapper[4754]: I0126 17:53:48.094065 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5h7fb_9f0c3994-3669-4cb5-89b3-b9e9dbc316de/cp-frr-files/0.log" Jan 26 17:53:48 crc kubenswrapper[4754]: I0126 17:53:48.588436 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5h7fb_9f0c3994-3669-4cb5-89b3-b9e9dbc316de/cp-frr-files/0.log" Jan 26 17:53:48 crc kubenswrapper[4754]: I0126 17:53:48.625264 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5h7fb_9f0c3994-3669-4cb5-89b3-b9e9dbc316de/cp-reloader/0.log" Jan 26 17:53:48 crc kubenswrapper[4754]: I0126 17:53:48.628012 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5h7fb_9f0c3994-3669-4cb5-89b3-b9e9dbc316de/cp-metrics/0.log" Jan 26 17:53:48 crc kubenswrapper[4754]: I0126 17:53:48.641927 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5h7fb_9f0c3994-3669-4cb5-89b3-b9e9dbc316de/cp-metrics/0.log" Jan 26 17:53:48 crc kubenswrapper[4754]: I0126 17:53:48.801282 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5h7fb_9f0c3994-3669-4cb5-89b3-b9e9dbc316de/cp-frr-files/0.log" Jan 26 17:53:48 crc kubenswrapper[4754]: I0126 17:53:48.818396 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5h7fb_9f0c3994-3669-4cb5-89b3-b9e9dbc316de/cp-metrics/0.log" Jan 26 17:53:48 crc kubenswrapper[4754]: I0126 17:53:48.824267 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5h7fb_9f0c3994-3669-4cb5-89b3-b9e9dbc316de/cp-reloader/0.log" Jan 26 17:53:48 crc kubenswrapper[4754]: I0126 17:53:48.873967 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5h7fb_9f0c3994-3669-4cb5-89b3-b9e9dbc316de/controller/0.log" Jan 26 17:53:49 crc kubenswrapper[4754]: I0126 17:53:49.027909 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5h7fb_9f0c3994-3669-4cb5-89b3-b9e9dbc316de/frr-metrics/0.log" Jan 26 17:53:49 crc kubenswrapper[4754]: I0126 17:53:49.034313 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5h7fb_9f0c3994-3669-4cb5-89b3-b9e9dbc316de/kube-rbac-proxy/0.log" Jan 26 17:53:49 crc kubenswrapper[4754]: I0126 17:53:49.077223 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5h7fb_9f0c3994-3669-4cb5-89b3-b9e9dbc316de/kube-rbac-proxy-frr/0.log" Jan 26 17:53:49 crc kubenswrapper[4754]: I0126 17:53:49.241629 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5h7fb_9f0c3994-3669-4cb5-89b3-b9e9dbc316de/reloader/0.log" Jan 26 17:53:49 crc kubenswrapper[4754]: I0126 17:53:49.299171 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7df86c4f6c-zbwg9_941e5b29-8892-451b-9ba6-afdad3c1c77b/frr-k8s-webhook-server/0.log" Jan 26 17:53:49 crc kubenswrapper[4754]: I0126 17:53:49.543964 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-755bc5d786-g2p48_83c99759-10ef-42eb-a58f-23c4e6c7d089/manager/0.log" Jan 26 17:53:49 crc kubenswrapper[4754]: I0126 17:53:49.687542 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-755c486f67-b6mn4_ee80d5f4-a13e-434d-b250-1e005d84dc59/webhook-server/0.log" Jan 26 17:53:49 crc kubenswrapper[4754]: I0126 17:53:49.865074 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-drvw7_b3234c0c-afa9-41b4-88e3-70f44df6c2ce/kube-rbac-proxy/0.log" Jan 26 17:53:50 crc kubenswrapper[4754]: I0126 17:53:50.435242 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-drvw7_b3234c0c-afa9-41b4-88e3-70f44df6c2ce/speaker/0.log" Jan 26 17:53:50 crc kubenswrapper[4754]: I0126 17:53:50.555477 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5h7fb_9f0c3994-3669-4cb5-89b3-b9e9dbc316de/frr/0.log" Jan 26 17:54:02 crc kubenswrapper[4754]: I0126 17:54:02.540336 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dckzdjj_997bf4f3-9268-4058-9ee7-80c581651bcd/util/0.log" Jan 26 17:54:02 crc kubenswrapper[4754]: I0126 17:54:02.736764 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dckzdjj_997bf4f3-9268-4058-9ee7-80c581651bcd/util/0.log" Jan 26 17:54:02 crc kubenswrapper[4754]: I0126 17:54:02.747011 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dckzdjj_997bf4f3-9268-4058-9ee7-80c581651bcd/pull/0.log" Jan 26 17:54:02 crc kubenswrapper[4754]: I0126 17:54:02.776194 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dckzdjj_997bf4f3-9268-4058-9ee7-80c581651bcd/pull/0.log" Jan 26 17:54:02 crc kubenswrapper[4754]: I0126 17:54:02.892199 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dckzdjj_997bf4f3-9268-4058-9ee7-80c581651bcd/util/0.log" Jan 26 17:54:02 crc kubenswrapper[4754]: I0126 17:54:02.944360 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dckzdjj_997bf4f3-9268-4058-9ee7-80c581651bcd/extract/0.log" Jan 26 17:54:02 crc kubenswrapper[4754]: I0126 17:54:02.952499 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dckzdjj_997bf4f3-9268-4058-9ee7-80c581651bcd/pull/0.log" Jan 26 17:54:03 crc kubenswrapper[4754]: I0126 17:54:03.081093 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713v2nvl_4212a345-63c7-4b01-b13d-12f9a3fc297a/util/0.log" Jan 26 17:54:03 crc kubenswrapper[4754]: I0126 17:54:03.405567 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713v2nvl_4212a345-63c7-4b01-b13d-12f9a3fc297a/pull/0.log" Jan 26 17:54:03 crc kubenswrapper[4754]: I0126 17:54:03.410016 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713v2nvl_4212a345-63c7-4b01-b13d-12f9a3fc297a/util/0.log" Jan 26 17:54:03 crc kubenswrapper[4754]: I0126 17:54:03.416636 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713v2nvl_4212a345-63c7-4b01-b13d-12f9a3fc297a/pull/0.log" Jan 26 17:54:03 crc kubenswrapper[4754]: I0126 17:54:03.566654 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713v2nvl_4212a345-63c7-4b01-b13d-12f9a3fc297a/extract/0.log" Jan 26 17:54:03 crc kubenswrapper[4754]: I0126 17:54:03.567813 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713v2nvl_4212a345-63c7-4b01-b13d-12f9a3fc297a/pull/0.log" Jan 26 17:54:03 crc kubenswrapper[4754]: I0126 17:54:03.570105 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713v2nvl_4212a345-63c7-4b01-b13d-12f9a3fc297a/util/0.log" Jan 26 17:54:03 crc kubenswrapper[4754]: I0126 17:54:03.752408 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-jvfdk_ceb4005f-7a24-4c06-8afd-5cad6c195c57/extract-utilities/0.log" Jan 26 17:54:03 crc kubenswrapper[4754]: I0126 17:54:03.922383 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-jvfdk_ceb4005f-7a24-4c06-8afd-5cad6c195c57/extract-content/0.log" Jan 26 17:54:03 crc kubenswrapper[4754]: I0126 17:54:03.923813 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-jvfdk_ceb4005f-7a24-4c06-8afd-5cad6c195c57/extract-utilities/0.log" Jan 26 17:54:03 crc kubenswrapper[4754]: I0126 17:54:03.925800 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-jvfdk_ceb4005f-7a24-4c06-8afd-5cad6c195c57/extract-content/0.log" Jan 26 17:54:04 crc kubenswrapper[4754]: I0126 17:54:04.082603 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-jvfdk_ceb4005f-7a24-4c06-8afd-5cad6c195c57/extract-utilities/0.log" Jan 26 17:54:04 crc kubenswrapper[4754]: I0126 17:54:04.116881 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-jvfdk_ceb4005f-7a24-4c06-8afd-5cad6c195c57/extract-content/0.log" Jan 26 17:54:04 crc kubenswrapper[4754]: I0126 17:54:04.327790 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-c44vn_02fabee4-adea-4bca-ba0a-e6b98f6e68dd/extract-utilities/0.log" Jan 26 17:54:04 crc kubenswrapper[4754]: I0126 17:54:04.408969 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-jvfdk_ceb4005f-7a24-4c06-8afd-5cad6c195c57/registry-server/0.log" Jan 26 17:54:04 crc kubenswrapper[4754]: I0126 17:54:04.491099 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-c44vn_02fabee4-adea-4bca-ba0a-e6b98f6e68dd/extract-utilities/0.log" Jan 26 17:54:04 crc kubenswrapper[4754]: I0126 17:54:04.535976 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-c44vn_02fabee4-adea-4bca-ba0a-e6b98f6e68dd/extract-content/0.log" Jan 26 17:54:04 crc kubenswrapper[4754]: I0126 17:54:04.555956 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-c44vn_02fabee4-adea-4bca-ba0a-e6b98f6e68dd/extract-content/0.log" Jan 26 17:54:04 crc kubenswrapper[4754]: I0126 17:54:04.740875 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-c44vn_02fabee4-adea-4bca-ba0a-e6b98f6e68dd/extract-utilities/0.log" Jan 26 17:54:04 crc kubenswrapper[4754]: I0126 17:54:04.741007 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-c44vn_02fabee4-adea-4bca-ba0a-e6b98f6e68dd/extract-content/0.log" Jan 26 17:54:04 crc kubenswrapper[4754]: I0126 17:54:04.960171 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-psxqw_41e45928-06b9-4854-9061-16e4053b0fb1/extract-utilities/0.log" Jan 26 17:54:05 crc kubenswrapper[4754]: I0126 17:54:05.027860 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-mp4qz_eedf4e24-8d2e-4ec6-9caa-f5af47592b89/marketplace-operator/0.log" Jan 26 17:54:05 crc kubenswrapper[4754]: I0126 17:54:05.311710 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-psxqw_41e45928-06b9-4854-9061-16e4053b0fb1/extract-utilities/0.log" Jan 26 17:54:05 crc kubenswrapper[4754]: I0126 17:54:05.325357 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-psxqw_41e45928-06b9-4854-9061-16e4053b0fb1/extract-content/0.log" Jan 26 17:54:05 crc kubenswrapper[4754]: I0126 17:54:05.343309 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-psxqw_41e45928-06b9-4854-9061-16e4053b0fb1/extract-content/0.log" Jan 26 17:54:05 crc kubenswrapper[4754]: I0126 17:54:05.477557 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-c44vn_02fabee4-adea-4bca-ba0a-e6b98f6e68dd/registry-server/0.log" Jan 26 17:54:05 crc kubenswrapper[4754]: I0126 17:54:05.525844 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-psxqw_41e45928-06b9-4854-9061-16e4053b0fb1/extract-content/0.log" Jan 26 17:54:05 crc kubenswrapper[4754]: I0126 17:54:05.530622 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-psxqw_41e45928-06b9-4854-9061-16e4053b0fb1/extract-utilities/0.log" Jan 26 17:54:05 crc kubenswrapper[4754]: I0126 17:54:05.673560 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-psxqw_41e45928-06b9-4854-9061-16e4053b0fb1/registry-server/0.log" Jan 26 17:54:05 crc kubenswrapper[4754]: I0126 17:54:05.743418 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-dn77h_2dfa533f-5161-4ef0-bc9b-397abae75b23/extract-utilities/0.log" Jan 26 17:54:05 crc kubenswrapper[4754]: I0126 17:54:05.913431 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-dn77h_2dfa533f-5161-4ef0-bc9b-397abae75b23/extract-content/0.log" Jan 26 17:54:05 crc kubenswrapper[4754]: I0126 17:54:05.935720 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-dn77h_2dfa533f-5161-4ef0-bc9b-397abae75b23/extract-content/0.log" Jan 26 17:54:05 crc kubenswrapper[4754]: I0126 17:54:05.966995 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-dn77h_2dfa533f-5161-4ef0-bc9b-397abae75b23/extract-utilities/0.log" Jan 26 17:54:06 crc kubenswrapper[4754]: I0126 17:54:06.123233 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-dn77h_2dfa533f-5161-4ef0-bc9b-397abae75b23/extract-content/0.log" Jan 26 17:54:06 crc kubenswrapper[4754]: I0126 17:54:06.125679 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-dn77h_2dfa533f-5161-4ef0-bc9b-397abae75b23/extract-utilities/0.log" Jan 26 17:54:06 crc kubenswrapper[4754]: I0126 17:54:06.729419 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-dn77h_2dfa533f-5161-4ef0-bc9b-397abae75b23/registry-server/0.log" Jan 26 17:55:52 crc kubenswrapper[4754]: I0126 17:55:52.990207 4754 generic.go:334] "Generic (PLEG): container finished" podID="8fd36f5c-ba61-4968-80fd-d82a19020b87" containerID="16ff16dad3135c7cfd616f48c6518a11f6b888cdc8cb783551fbefa71546d0da" exitCode=0 Jan 26 17:55:52 crc kubenswrapper[4754]: I0126 17:55:52.990306 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-ldztm/must-gather-rxq7n" event={"ID":"8fd36f5c-ba61-4968-80fd-d82a19020b87","Type":"ContainerDied","Data":"16ff16dad3135c7cfd616f48c6518a11f6b888cdc8cb783551fbefa71546d0da"} Jan 26 17:55:52 crc kubenswrapper[4754]: I0126 17:55:52.991565 4754 scope.go:117] "RemoveContainer" containerID="16ff16dad3135c7cfd616f48c6518a11f6b888cdc8cb783551fbefa71546d0da" Jan 26 17:55:53 crc kubenswrapper[4754]: I0126 17:55:53.753733 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-ldztm_must-gather-rxq7n_8fd36f5c-ba61-4968-80fd-d82a19020b87/gather/0.log" Jan 26 17:56:01 crc kubenswrapper[4754]: I0126 17:56:01.921169 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-ldztm/must-gather-rxq7n"] Jan 26 17:56:01 crc kubenswrapper[4754]: I0126 17:56:01.922305 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-ldztm/must-gather-rxq7n" podUID="8fd36f5c-ba61-4968-80fd-d82a19020b87" containerName="copy" containerID="cri-o://cfa489b848ca1551df79275c29bbd3d9ff8fa27da4b862e0bc1d45b7822fdaee" gracePeriod=2 Jan 26 17:56:01 crc kubenswrapper[4754]: I0126 17:56:01.928814 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-ldztm/must-gather-rxq7n"] Jan 26 17:56:02 crc kubenswrapper[4754]: I0126 17:56:02.082570 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-ldztm_must-gather-rxq7n_8fd36f5c-ba61-4968-80fd-d82a19020b87/copy/0.log" Jan 26 17:56:02 crc kubenswrapper[4754]: I0126 17:56:02.083284 4754 generic.go:334] "Generic (PLEG): container finished" podID="8fd36f5c-ba61-4968-80fd-d82a19020b87" containerID="cfa489b848ca1551df79275c29bbd3d9ff8fa27da4b862e0bc1d45b7822fdaee" exitCode=143 Jan 26 17:56:02 crc kubenswrapper[4754]: I0126 17:56:02.365620 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-ldztm_must-gather-rxq7n_8fd36f5c-ba61-4968-80fd-d82a19020b87/copy/0.log" Jan 26 17:56:02 crc kubenswrapper[4754]: I0126 17:56:02.370016 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ldztm/must-gather-rxq7n" Jan 26 17:56:02 crc kubenswrapper[4754]: I0126 17:56:02.498312 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/8fd36f5c-ba61-4968-80fd-d82a19020b87-must-gather-output\") pod \"8fd36f5c-ba61-4968-80fd-d82a19020b87\" (UID: \"8fd36f5c-ba61-4968-80fd-d82a19020b87\") " Jan 26 17:56:02 crc kubenswrapper[4754]: I0126 17:56:02.498456 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5d4rs\" (UniqueName: \"kubernetes.io/projected/8fd36f5c-ba61-4968-80fd-d82a19020b87-kube-api-access-5d4rs\") pod \"8fd36f5c-ba61-4968-80fd-d82a19020b87\" (UID: \"8fd36f5c-ba61-4968-80fd-d82a19020b87\") " Jan 26 17:56:02 crc kubenswrapper[4754]: I0126 17:56:02.505530 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8fd36f5c-ba61-4968-80fd-d82a19020b87-kube-api-access-5d4rs" (OuterVolumeSpecName: "kube-api-access-5d4rs") pod "8fd36f5c-ba61-4968-80fd-d82a19020b87" (UID: "8fd36f5c-ba61-4968-80fd-d82a19020b87"). InnerVolumeSpecName "kube-api-access-5d4rs". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:56:02 crc kubenswrapper[4754]: I0126 17:56:02.601343 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5d4rs\" (UniqueName: \"kubernetes.io/projected/8fd36f5c-ba61-4968-80fd-d82a19020b87-kube-api-access-5d4rs\") on node \"crc\" DevicePath \"\"" Jan 26 17:56:02 crc kubenswrapper[4754]: I0126 17:56:02.658215 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8fd36f5c-ba61-4968-80fd-d82a19020b87-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "8fd36f5c-ba61-4968-80fd-d82a19020b87" (UID: "8fd36f5c-ba61-4968-80fd-d82a19020b87"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:56:02 crc kubenswrapper[4754]: I0126 17:56:02.713946 4754 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/8fd36f5c-ba61-4968-80fd-d82a19020b87-must-gather-output\") on node \"crc\" DevicePath \"\"" Jan 26 17:56:03 crc kubenswrapper[4754]: I0126 17:56:03.101336 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-ldztm_must-gather-rxq7n_8fd36f5c-ba61-4968-80fd-d82a19020b87/copy/0.log" Jan 26 17:56:03 crc kubenswrapper[4754]: I0126 17:56:03.101911 4754 scope.go:117] "RemoveContainer" containerID="cfa489b848ca1551df79275c29bbd3d9ff8fa27da4b862e0bc1d45b7822fdaee" Jan 26 17:56:03 crc kubenswrapper[4754]: I0126 17:56:03.101992 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ldztm/must-gather-rxq7n" Jan 26 17:56:03 crc kubenswrapper[4754]: I0126 17:56:03.124366 4754 scope.go:117] "RemoveContainer" containerID="16ff16dad3135c7cfd616f48c6518a11f6b888cdc8cb783551fbefa71546d0da" Jan 26 17:56:03 crc kubenswrapper[4754]: I0126 17:56:03.786557 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8fd36f5c-ba61-4968-80fd-d82a19020b87" path="/var/lib/kubelet/pods/8fd36f5c-ba61-4968-80fd-d82a19020b87/volumes" Jan 26 17:56:07 crc kubenswrapper[4754]: I0126 17:56:07.129513 4754 patch_prober.go:28] interesting pod/machine-config-daemon-x65wv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 26 17:56:07 crc kubenswrapper[4754]: I0126 17:56:07.129873 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 26 17:56:37 crc kubenswrapper[4754]: I0126 17:56:37.129570 4754 patch_prober.go:28] interesting pod/machine-config-daemon-x65wv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 26 17:56:37 crc kubenswrapper[4754]: I0126 17:56:37.130031 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 26 17:57:07 crc kubenswrapper[4754]: I0126 17:57:07.128896 4754 patch_prober.go:28] interesting pod/machine-config-daemon-x65wv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 26 17:57:07 crc kubenswrapper[4754]: I0126 17:57:07.129790 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 26 17:57:07 crc kubenswrapper[4754]: I0126 17:57:07.129875 4754 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" Jan 26 17:57:07 crc kubenswrapper[4754]: I0126 17:57:07.131288 4754 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"25b9a1afdbf7b01e5200dd76d80daf1edca89288d2a1ec786b655b6ff7ce5a88"} pod="openshift-machine-config-operator/machine-config-daemon-x65wv" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 26 17:57:07 crc kubenswrapper[4754]: I0126 17:57:07.131421 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" containerName="machine-config-daemon" containerID="cri-o://25b9a1afdbf7b01e5200dd76d80daf1edca89288d2a1ec786b655b6ff7ce5a88" gracePeriod=600 Jan 26 17:57:07 crc kubenswrapper[4754]: I0126 17:57:07.671544 4754 generic.go:334] "Generic (PLEG): container finished" podID="8c3718a4-f354-4284-92e0-fdfb45a692bd" containerID="25b9a1afdbf7b01e5200dd76d80daf1edca89288d2a1ec786b655b6ff7ce5a88" exitCode=0 Jan 26 17:57:07 crc kubenswrapper[4754]: I0126 17:57:07.671625 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" event={"ID":"8c3718a4-f354-4284-92e0-fdfb45a692bd","Type":"ContainerDied","Data":"25b9a1afdbf7b01e5200dd76d80daf1edca89288d2a1ec786b655b6ff7ce5a88"} Jan 26 17:57:07 crc kubenswrapper[4754]: I0126 17:57:07.671922 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" event={"ID":"8c3718a4-f354-4284-92e0-fdfb45a692bd","Type":"ContainerStarted","Data":"b36649b0f03300e411a658fea6f425b2be30bb8d773b1d9ef5bb8f521458f9a6"} Jan 26 17:57:07 crc kubenswrapper[4754]: I0126 17:57:07.671944 4754 scope.go:117] "RemoveContainer" containerID="7b1482a92488015f91f13a86a1aa38c65e69fa9f874f9482c4d8f0f34463d2c2" Jan 26 17:57:28 crc kubenswrapper[4754]: I0126 17:57:28.124510 4754 scope.go:117] "RemoveContainer" containerID="9763fe4f8cd2d2d229ae61d0290c1fe2bfc0e075232bde8c244f8cbbf16b7b72" Jan 26 17:58:59 crc kubenswrapper[4754]: I0126 17:58:59.606501 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-zb6c2/must-gather-xhxjk"] Jan 26 17:58:59 crc kubenswrapper[4754]: E0126 17:58:59.607689 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="883de6ad-32c9-4537-a029-ad0047b3b644" containerName="extract-content" Jan 26 17:58:59 crc kubenswrapper[4754]: I0126 17:58:59.607708 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="883de6ad-32c9-4537-a029-ad0047b3b644" containerName="extract-content" Jan 26 17:58:59 crc kubenswrapper[4754]: E0126 17:58:59.607726 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8fd36f5c-ba61-4968-80fd-d82a19020b87" containerName="gather" Jan 26 17:58:59 crc kubenswrapper[4754]: I0126 17:58:59.607734 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="8fd36f5c-ba61-4968-80fd-d82a19020b87" containerName="gather" Jan 26 17:58:59 crc kubenswrapper[4754]: E0126 17:58:59.607747 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8fd36f5c-ba61-4968-80fd-d82a19020b87" containerName="copy" Jan 26 17:58:59 crc kubenswrapper[4754]: I0126 17:58:59.607759 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="8fd36f5c-ba61-4968-80fd-d82a19020b87" containerName="copy" Jan 26 17:58:59 crc kubenswrapper[4754]: E0126 17:58:59.607767 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="883de6ad-32c9-4537-a029-ad0047b3b644" containerName="registry-server" Jan 26 17:58:59 crc kubenswrapper[4754]: I0126 17:58:59.607774 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="883de6ad-32c9-4537-a029-ad0047b3b644" containerName="registry-server" Jan 26 17:58:59 crc kubenswrapper[4754]: E0126 17:58:59.607802 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="883de6ad-32c9-4537-a029-ad0047b3b644" containerName="extract-utilities" Jan 26 17:58:59 crc kubenswrapper[4754]: I0126 17:58:59.607811 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="883de6ad-32c9-4537-a029-ad0047b3b644" containerName="extract-utilities" Jan 26 17:58:59 crc kubenswrapper[4754]: E0126 17:58:59.607837 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7c4d852-0b8e-415a-9296-5f4174f7ab22" containerName="extract-utilities" Jan 26 17:58:59 crc kubenswrapper[4754]: I0126 17:58:59.607846 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7c4d852-0b8e-415a-9296-5f4174f7ab22" containerName="extract-utilities" Jan 26 17:58:59 crc kubenswrapper[4754]: E0126 17:58:59.607862 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7c4d852-0b8e-415a-9296-5f4174f7ab22" containerName="registry-server" Jan 26 17:58:59 crc kubenswrapper[4754]: I0126 17:58:59.607869 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7c4d852-0b8e-415a-9296-5f4174f7ab22" containerName="registry-server" Jan 26 17:58:59 crc kubenswrapper[4754]: E0126 17:58:59.607878 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7c4d852-0b8e-415a-9296-5f4174f7ab22" containerName="extract-content" Jan 26 17:58:59 crc kubenswrapper[4754]: I0126 17:58:59.607888 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7c4d852-0b8e-415a-9296-5f4174f7ab22" containerName="extract-content" Jan 26 17:58:59 crc kubenswrapper[4754]: I0126 17:58:59.608095 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="883de6ad-32c9-4537-a029-ad0047b3b644" containerName="registry-server" Jan 26 17:58:59 crc kubenswrapper[4754]: I0126 17:58:59.608110 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="f7c4d852-0b8e-415a-9296-5f4174f7ab22" containerName="registry-server" Jan 26 17:58:59 crc kubenswrapper[4754]: I0126 17:58:59.608125 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="8fd36f5c-ba61-4968-80fd-d82a19020b87" containerName="copy" Jan 26 17:58:59 crc kubenswrapper[4754]: I0126 17:58:59.608146 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="8fd36f5c-ba61-4968-80fd-d82a19020b87" containerName="gather" Jan 26 17:58:59 crc kubenswrapper[4754]: I0126 17:58:59.609355 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-zb6c2/must-gather-xhxjk" Jan 26 17:58:59 crc kubenswrapper[4754]: I0126 17:58:59.613589 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-zb6c2"/"default-dockercfg-zpwhk" Jan 26 17:58:59 crc kubenswrapper[4754]: I0126 17:58:59.613688 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-zb6c2"/"kube-root-ca.crt" Jan 26 17:58:59 crc kubenswrapper[4754]: I0126 17:58:59.619159 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-zb6c2"/"openshift-service-ca.crt" Jan 26 17:58:59 crc kubenswrapper[4754]: I0126 17:58:59.629989 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/bea1bcb2-12bb-4034-b289-9d7eda6175b2-must-gather-output\") pod \"must-gather-xhxjk\" (UID: \"bea1bcb2-12bb-4034-b289-9d7eda6175b2\") " pod="openshift-must-gather-zb6c2/must-gather-xhxjk" Jan 26 17:58:59 crc kubenswrapper[4754]: I0126 17:58:59.630064 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fdfn4\" (UniqueName: \"kubernetes.io/projected/bea1bcb2-12bb-4034-b289-9d7eda6175b2-kube-api-access-fdfn4\") pod \"must-gather-xhxjk\" (UID: \"bea1bcb2-12bb-4034-b289-9d7eda6175b2\") " pod="openshift-must-gather-zb6c2/must-gather-xhxjk" Jan 26 17:58:59 crc kubenswrapper[4754]: I0126 17:58:59.633644 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-zb6c2/must-gather-xhxjk"] Jan 26 17:58:59 crc kubenswrapper[4754]: I0126 17:58:59.731459 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fdfn4\" (UniqueName: \"kubernetes.io/projected/bea1bcb2-12bb-4034-b289-9d7eda6175b2-kube-api-access-fdfn4\") pod \"must-gather-xhxjk\" (UID: \"bea1bcb2-12bb-4034-b289-9d7eda6175b2\") " pod="openshift-must-gather-zb6c2/must-gather-xhxjk" Jan 26 17:58:59 crc kubenswrapper[4754]: I0126 17:58:59.731689 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/bea1bcb2-12bb-4034-b289-9d7eda6175b2-must-gather-output\") pod \"must-gather-xhxjk\" (UID: \"bea1bcb2-12bb-4034-b289-9d7eda6175b2\") " pod="openshift-must-gather-zb6c2/must-gather-xhxjk" Jan 26 17:58:59 crc kubenswrapper[4754]: I0126 17:58:59.732287 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/bea1bcb2-12bb-4034-b289-9d7eda6175b2-must-gather-output\") pod \"must-gather-xhxjk\" (UID: \"bea1bcb2-12bb-4034-b289-9d7eda6175b2\") " pod="openshift-must-gather-zb6c2/must-gather-xhxjk" Jan 26 17:58:59 crc kubenswrapper[4754]: I0126 17:58:59.756787 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fdfn4\" (UniqueName: \"kubernetes.io/projected/bea1bcb2-12bb-4034-b289-9d7eda6175b2-kube-api-access-fdfn4\") pod \"must-gather-xhxjk\" (UID: \"bea1bcb2-12bb-4034-b289-9d7eda6175b2\") " pod="openshift-must-gather-zb6c2/must-gather-xhxjk" Jan 26 17:58:59 crc kubenswrapper[4754]: I0126 17:58:59.929907 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-zb6c2/must-gather-xhxjk" Jan 26 17:59:00 crc kubenswrapper[4754]: I0126 17:59:00.388300 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-zb6c2/must-gather-xhxjk"] Jan 26 17:59:00 crc kubenswrapper[4754]: I0126 17:59:00.721239 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-zb6c2/must-gather-xhxjk" event={"ID":"bea1bcb2-12bb-4034-b289-9d7eda6175b2","Type":"ContainerStarted","Data":"d432469e4534c9613a668b7e303bbc4ee01cac5fb65975e5c21882834d3a8d11"} Jan 26 17:59:00 crc kubenswrapper[4754]: I0126 17:59:00.721293 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-zb6c2/must-gather-xhxjk" event={"ID":"bea1bcb2-12bb-4034-b289-9d7eda6175b2","Type":"ContainerStarted","Data":"5dafdf3f8db64caee5cfdb7a131511647283aed2714847d8469e15b2e2eb9c56"} Jan 26 17:59:01 crc kubenswrapper[4754]: I0126 17:59:01.737078 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-zb6c2/must-gather-xhxjk" event={"ID":"bea1bcb2-12bb-4034-b289-9d7eda6175b2","Type":"ContainerStarted","Data":"479d15afd09a697338c933f25fd2f03484df7b08f42579a078a42e06f51bbcfc"} Jan 26 17:59:01 crc kubenswrapper[4754]: I0126 17:59:01.759393 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-zb6c2/must-gather-xhxjk" podStartSLOduration=2.759370244 podStartE2EDuration="2.759370244s" podCreationTimestamp="2026-01-26 17:58:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 17:59:01.755478708 +0000 UTC m=+4308.279659162" watchObservedRunningTime="2026-01-26 17:59:01.759370244 +0000 UTC m=+4308.283550688" Jan 26 17:59:03 crc kubenswrapper[4754]: I0126 17:59:03.867365 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-zb6c2/crc-debug-sk748"] Jan 26 17:59:03 crc kubenswrapper[4754]: I0126 17:59:03.871037 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-zb6c2/crc-debug-sk748" Jan 26 17:59:04 crc kubenswrapper[4754]: I0126 17:59:04.011746 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/8203e10d-062b-4b97-9573-bcfd343b428e-host\") pod \"crc-debug-sk748\" (UID: \"8203e10d-062b-4b97-9573-bcfd343b428e\") " pod="openshift-must-gather-zb6c2/crc-debug-sk748" Jan 26 17:59:04 crc kubenswrapper[4754]: I0126 17:59:04.011882 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2n5hp\" (UniqueName: \"kubernetes.io/projected/8203e10d-062b-4b97-9573-bcfd343b428e-kube-api-access-2n5hp\") pod \"crc-debug-sk748\" (UID: \"8203e10d-062b-4b97-9573-bcfd343b428e\") " pod="openshift-must-gather-zb6c2/crc-debug-sk748" Jan 26 17:59:04 crc kubenswrapper[4754]: I0126 17:59:04.114093 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/8203e10d-062b-4b97-9573-bcfd343b428e-host\") pod \"crc-debug-sk748\" (UID: \"8203e10d-062b-4b97-9573-bcfd343b428e\") " pod="openshift-must-gather-zb6c2/crc-debug-sk748" Jan 26 17:59:04 crc kubenswrapper[4754]: I0126 17:59:04.114164 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2n5hp\" (UniqueName: \"kubernetes.io/projected/8203e10d-062b-4b97-9573-bcfd343b428e-kube-api-access-2n5hp\") pod \"crc-debug-sk748\" (UID: \"8203e10d-062b-4b97-9573-bcfd343b428e\") " pod="openshift-must-gather-zb6c2/crc-debug-sk748" Jan 26 17:59:04 crc kubenswrapper[4754]: I0126 17:59:04.114592 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/8203e10d-062b-4b97-9573-bcfd343b428e-host\") pod \"crc-debug-sk748\" (UID: \"8203e10d-062b-4b97-9573-bcfd343b428e\") " pod="openshift-must-gather-zb6c2/crc-debug-sk748" Jan 26 17:59:04 crc kubenswrapper[4754]: I0126 17:59:04.133043 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2n5hp\" (UniqueName: \"kubernetes.io/projected/8203e10d-062b-4b97-9573-bcfd343b428e-kube-api-access-2n5hp\") pod \"crc-debug-sk748\" (UID: \"8203e10d-062b-4b97-9573-bcfd343b428e\") " pod="openshift-must-gather-zb6c2/crc-debug-sk748" Jan 26 17:59:04 crc kubenswrapper[4754]: I0126 17:59:04.195898 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-zb6c2/crc-debug-sk748" Jan 26 17:59:04 crc kubenswrapper[4754]: W0126 17:59:04.220687 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8203e10d_062b_4b97_9573_bcfd343b428e.slice/crio-26f8362a5df1fa6f19fc1401fbbc073047e4ea1d2fc518bd2828ea2eca6d29ef WatchSource:0}: Error finding container 26f8362a5df1fa6f19fc1401fbbc073047e4ea1d2fc518bd2828ea2eca6d29ef: Status 404 returned error can't find the container with id 26f8362a5df1fa6f19fc1401fbbc073047e4ea1d2fc518bd2828ea2eca6d29ef Jan 26 17:59:04 crc kubenswrapper[4754]: I0126 17:59:04.764926 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-zb6c2/crc-debug-sk748" event={"ID":"8203e10d-062b-4b97-9573-bcfd343b428e","Type":"ContainerStarted","Data":"c7923bfbe51ba7551e08e70ea0f69993fab9e8005f2bbb3eb514c63a240c7f43"} Jan 26 17:59:04 crc kubenswrapper[4754]: I0126 17:59:04.765486 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-zb6c2/crc-debug-sk748" event={"ID":"8203e10d-062b-4b97-9573-bcfd343b428e","Type":"ContainerStarted","Data":"26f8362a5df1fa6f19fc1401fbbc073047e4ea1d2fc518bd2828ea2eca6d29ef"} Jan 26 17:59:04 crc kubenswrapper[4754]: I0126 17:59:04.790596 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-zb6c2/crc-debug-sk748" podStartSLOduration=1.790577008 podStartE2EDuration="1.790577008s" podCreationTimestamp="2026-01-26 17:59:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 17:59:04.777394097 +0000 UTC m=+4311.301574531" watchObservedRunningTime="2026-01-26 17:59:04.790577008 +0000 UTC m=+4311.314757462" Jan 26 17:59:07 crc kubenswrapper[4754]: I0126 17:59:07.129532 4754 patch_prober.go:28] interesting pod/machine-config-daemon-x65wv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 26 17:59:07 crc kubenswrapper[4754]: I0126 17:59:07.130732 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 26 17:59:36 crc kubenswrapper[4754]: E0126 17:59:36.518555 4754 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8203e10d_062b_4b97_9573_bcfd343b428e.slice/crio-conmon-c7923bfbe51ba7551e08e70ea0f69993fab9e8005f2bbb3eb514c63a240c7f43.scope\": RecentStats: unable to find data in memory cache]" Jan 26 17:59:37 crc kubenswrapper[4754]: I0126 17:59:37.077900 4754 generic.go:334] "Generic (PLEG): container finished" podID="8203e10d-062b-4b97-9573-bcfd343b428e" containerID="c7923bfbe51ba7551e08e70ea0f69993fab9e8005f2bbb3eb514c63a240c7f43" exitCode=0 Jan 26 17:59:37 crc kubenswrapper[4754]: I0126 17:59:37.077998 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-zb6c2/crc-debug-sk748" event={"ID":"8203e10d-062b-4b97-9573-bcfd343b428e","Type":"ContainerDied","Data":"c7923bfbe51ba7551e08e70ea0f69993fab9e8005f2bbb3eb514c63a240c7f43"} Jan 26 17:59:37 crc kubenswrapper[4754]: I0126 17:59:37.129619 4754 patch_prober.go:28] interesting pod/machine-config-daemon-x65wv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 26 17:59:37 crc kubenswrapper[4754]: I0126 17:59:37.129758 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 26 17:59:38 crc kubenswrapper[4754]: I0126 17:59:38.198214 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-zb6c2/crc-debug-sk748" Jan 26 17:59:38 crc kubenswrapper[4754]: I0126 17:59:38.230302 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-zb6c2/crc-debug-sk748"] Jan 26 17:59:38 crc kubenswrapper[4754]: I0126 17:59:38.238048 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-zb6c2/crc-debug-sk748"] Jan 26 17:59:38 crc kubenswrapper[4754]: I0126 17:59:38.254631 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/8203e10d-062b-4b97-9573-bcfd343b428e-host\") pod \"8203e10d-062b-4b97-9573-bcfd343b428e\" (UID: \"8203e10d-062b-4b97-9573-bcfd343b428e\") " Jan 26 17:59:38 crc kubenswrapper[4754]: I0126 17:59:38.254724 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2n5hp\" (UniqueName: \"kubernetes.io/projected/8203e10d-062b-4b97-9573-bcfd343b428e-kube-api-access-2n5hp\") pod \"8203e10d-062b-4b97-9573-bcfd343b428e\" (UID: \"8203e10d-062b-4b97-9573-bcfd343b428e\") " Jan 26 17:59:38 crc kubenswrapper[4754]: I0126 17:59:38.254922 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8203e10d-062b-4b97-9573-bcfd343b428e-host" (OuterVolumeSpecName: "host") pod "8203e10d-062b-4b97-9573-bcfd343b428e" (UID: "8203e10d-062b-4b97-9573-bcfd343b428e"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 17:59:38 crc kubenswrapper[4754]: I0126 17:59:38.255196 4754 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/8203e10d-062b-4b97-9573-bcfd343b428e-host\") on node \"crc\" DevicePath \"\"" Jan 26 17:59:38 crc kubenswrapper[4754]: I0126 17:59:38.261205 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8203e10d-062b-4b97-9573-bcfd343b428e-kube-api-access-2n5hp" (OuterVolumeSpecName: "kube-api-access-2n5hp") pod "8203e10d-062b-4b97-9573-bcfd343b428e" (UID: "8203e10d-062b-4b97-9573-bcfd343b428e"). InnerVolumeSpecName "kube-api-access-2n5hp". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:59:38 crc kubenswrapper[4754]: I0126 17:59:38.357146 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2n5hp\" (UniqueName: \"kubernetes.io/projected/8203e10d-062b-4b97-9573-bcfd343b428e-kube-api-access-2n5hp\") on node \"crc\" DevicePath \"\"" Jan 26 17:59:39 crc kubenswrapper[4754]: I0126 17:59:39.102863 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="26f8362a5df1fa6f19fc1401fbbc073047e4ea1d2fc518bd2828ea2eca6d29ef" Jan 26 17:59:39 crc kubenswrapper[4754]: I0126 17:59:39.102931 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-zb6c2/crc-debug-sk748" Jan 26 17:59:39 crc kubenswrapper[4754]: I0126 17:59:39.440481 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-zb6c2/crc-debug-pbm5p"] Jan 26 17:59:39 crc kubenswrapper[4754]: E0126 17:59:39.441161 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8203e10d-062b-4b97-9573-bcfd343b428e" containerName="container-00" Jan 26 17:59:39 crc kubenswrapper[4754]: I0126 17:59:39.441174 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="8203e10d-062b-4b97-9573-bcfd343b428e" containerName="container-00" Jan 26 17:59:39 crc kubenswrapper[4754]: I0126 17:59:39.441370 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="8203e10d-062b-4b97-9573-bcfd343b428e" containerName="container-00" Jan 26 17:59:39 crc kubenswrapper[4754]: I0126 17:59:39.442006 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-zb6c2/crc-debug-pbm5p" Jan 26 17:59:39 crc kubenswrapper[4754]: I0126 17:59:39.581026 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dfm47\" (UniqueName: \"kubernetes.io/projected/7320f5e8-c79d-4707-bd0f-8a66298dbd28-kube-api-access-dfm47\") pod \"crc-debug-pbm5p\" (UID: \"7320f5e8-c79d-4707-bd0f-8a66298dbd28\") " pod="openshift-must-gather-zb6c2/crc-debug-pbm5p" Jan 26 17:59:39 crc kubenswrapper[4754]: I0126 17:59:39.581144 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7320f5e8-c79d-4707-bd0f-8a66298dbd28-host\") pod \"crc-debug-pbm5p\" (UID: \"7320f5e8-c79d-4707-bd0f-8a66298dbd28\") " pod="openshift-must-gather-zb6c2/crc-debug-pbm5p" Jan 26 17:59:39 crc kubenswrapper[4754]: I0126 17:59:39.682371 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7320f5e8-c79d-4707-bd0f-8a66298dbd28-host\") pod \"crc-debug-pbm5p\" (UID: \"7320f5e8-c79d-4707-bd0f-8a66298dbd28\") " pod="openshift-must-gather-zb6c2/crc-debug-pbm5p" Jan 26 17:59:39 crc kubenswrapper[4754]: I0126 17:59:39.682495 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7320f5e8-c79d-4707-bd0f-8a66298dbd28-host\") pod \"crc-debug-pbm5p\" (UID: \"7320f5e8-c79d-4707-bd0f-8a66298dbd28\") " pod="openshift-must-gather-zb6c2/crc-debug-pbm5p" Jan 26 17:59:39 crc kubenswrapper[4754]: I0126 17:59:39.682546 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dfm47\" (UniqueName: \"kubernetes.io/projected/7320f5e8-c79d-4707-bd0f-8a66298dbd28-kube-api-access-dfm47\") pod \"crc-debug-pbm5p\" (UID: \"7320f5e8-c79d-4707-bd0f-8a66298dbd28\") " pod="openshift-must-gather-zb6c2/crc-debug-pbm5p" Jan 26 17:59:39 crc kubenswrapper[4754]: I0126 17:59:39.714463 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dfm47\" (UniqueName: \"kubernetes.io/projected/7320f5e8-c79d-4707-bd0f-8a66298dbd28-kube-api-access-dfm47\") pod \"crc-debug-pbm5p\" (UID: \"7320f5e8-c79d-4707-bd0f-8a66298dbd28\") " pod="openshift-must-gather-zb6c2/crc-debug-pbm5p" Jan 26 17:59:39 crc kubenswrapper[4754]: I0126 17:59:39.762453 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-zb6c2/crc-debug-pbm5p" Jan 26 17:59:39 crc kubenswrapper[4754]: I0126 17:59:39.778429 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8203e10d-062b-4b97-9573-bcfd343b428e" path="/var/lib/kubelet/pods/8203e10d-062b-4b97-9573-bcfd343b428e/volumes" Jan 26 17:59:40 crc kubenswrapper[4754]: I0126 17:59:40.121305 4754 generic.go:334] "Generic (PLEG): container finished" podID="7320f5e8-c79d-4707-bd0f-8a66298dbd28" containerID="4fd664dac47dbc4bdab1902f5ac0867dfc6646ba19bdedb5fcc34599d67ba8aa" exitCode=0 Jan 26 17:59:40 crc kubenswrapper[4754]: I0126 17:59:40.121396 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-zb6c2/crc-debug-pbm5p" event={"ID":"7320f5e8-c79d-4707-bd0f-8a66298dbd28","Type":"ContainerDied","Data":"4fd664dac47dbc4bdab1902f5ac0867dfc6646ba19bdedb5fcc34599d67ba8aa"} Jan 26 17:59:40 crc kubenswrapper[4754]: I0126 17:59:40.121594 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-zb6c2/crc-debug-pbm5p" event={"ID":"7320f5e8-c79d-4707-bd0f-8a66298dbd28","Type":"ContainerStarted","Data":"23b9ac15a1479d2d334ffc619a5de0d9342d059b84fe7b3c47e2672502b4274f"} Jan 26 17:59:40 crc kubenswrapper[4754]: I0126 17:59:40.485841 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-zb6c2/crc-debug-pbm5p"] Jan 26 17:59:40 crc kubenswrapper[4754]: I0126 17:59:40.491757 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-zb6c2/crc-debug-pbm5p"] Jan 26 17:59:40 crc kubenswrapper[4754]: I0126 17:59:40.873953 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-9pk77"] Jan 26 17:59:40 crc kubenswrapper[4754]: E0126 17:59:40.874523 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7320f5e8-c79d-4707-bd0f-8a66298dbd28" containerName="container-00" Jan 26 17:59:40 crc kubenswrapper[4754]: I0126 17:59:40.874613 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="7320f5e8-c79d-4707-bd0f-8a66298dbd28" containerName="container-00" Jan 26 17:59:40 crc kubenswrapper[4754]: I0126 17:59:40.874953 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="7320f5e8-c79d-4707-bd0f-8a66298dbd28" containerName="container-00" Jan 26 17:59:40 crc kubenswrapper[4754]: I0126 17:59:40.877545 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9pk77" Jan 26 17:59:40 crc kubenswrapper[4754]: I0126 17:59:40.888345 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-9pk77"] Jan 26 17:59:41 crc kubenswrapper[4754]: I0126 17:59:41.042286 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4ht45\" (UniqueName: \"kubernetes.io/projected/2810e86d-447c-4a21-b18d-c479684410d4-kube-api-access-4ht45\") pod \"redhat-marketplace-9pk77\" (UID: \"2810e86d-447c-4a21-b18d-c479684410d4\") " pod="openshift-marketplace/redhat-marketplace-9pk77" Jan 26 17:59:41 crc kubenswrapper[4754]: I0126 17:59:41.042355 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2810e86d-447c-4a21-b18d-c479684410d4-utilities\") pod \"redhat-marketplace-9pk77\" (UID: \"2810e86d-447c-4a21-b18d-c479684410d4\") " pod="openshift-marketplace/redhat-marketplace-9pk77" Jan 26 17:59:41 crc kubenswrapper[4754]: I0126 17:59:41.042403 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2810e86d-447c-4a21-b18d-c479684410d4-catalog-content\") pod \"redhat-marketplace-9pk77\" (UID: \"2810e86d-447c-4a21-b18d-c479684410d4\") " pod="openshift-marketplace/redhat-marketplace-9pk77" Jan 26 17:59:41 crc kubenswrapper[4754]: I0126 17:59:41.144885 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2810e86d-447c-4a21-b18d-c479684410d4-utilities\") pod \"redhat-marketplace-9pk77\" (UID: \"2810e86d-447c-4a21-b18d-c479684410d4\") " pod="openshift-marketplace/redhat-marketplace-9pk77" Jan 26 17:59:41 crc kubenswrapper[4754]: I0126 17:59:41.144982 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2810e86d-447c-4a21-b18d-c479684410d4-catalog-content\") pod \"redhat-marketplace-9pk77\" (UID: \"2810e86d-447c-4a21-b18d-c479684410d4\") " pod="openshift-marketplace/redhat-marketplace-9pk77" Jan 26 17:59:41 crc kubenswrapper[4754]: I0126 17:59:41.145144 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4ht45\" (UniqueName: \"kubernetes.io/projected/2810e86d-447c-4a21-b18d-c479684410d4-kube-api-access-4ht45\") pod \"redhat-marketplace-9pk77\" (UID: \"2810e86d-447c-4a21-b18d-c479684410d4\") " pod="openshift-marketplace/redhat-marketplace-9pk77" Jan 26 17:59:41 crc kubenswrapper[4754]: I0126 17:59:41.147318 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2810e86d-447c-4a21-b18d-c479684410d4-utilities\") pod \"redhat-marketplace-9pk77\" (UID: \"2810e86d-447c-4a21-b18d-c479684410d4\") " pod="openshift-marketplace/redhat-marketplace-9pk77" Jan 26 17:59:41 crc kubenswrapper[4754]: I0126 17:59:41.147383 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2810e86d-447c-4a21-b18d-c479684410d4-catalog-content\") pod \"redhat-marketplace-9pk77\" (UID: \"2810e86d-447c-4a21-b18d-c479684410d4\") " pod="openshift-marketplace/redhat-marketplace-9pk77" Jan 26 17:59:41 crc kubenswrapper[4754]: I0126 17:59:41.166505 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4ht45\" (UniqueName: \"kubernetes.io/projected/2810e86d-447c-4a21-b18d-c479684410d4-kube-api-access-4ht45\") pod \"redhat-marketplace-9pk77\" (UID: \"2810e86d-447c-4a21-b18d-c479684410d4\") " pod="openshift-marketplace/redhat-marketplace-9pk77" Jan 26 17:59:41 crc kubenswrapper[4754]: I0126 17:59:41.214628 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9pk77" Jan 26 17:59:41 crc kubenswrapper[4754]: I0126 17:59:41.314962 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-zb6c2/crc-debug-pbm5p" Jan 26 17:59:41 crc kubenswrapper[4754]: I0126 17:59:41.450399 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7320f5e8-c79d-4707-bd0f-8a66298dbd28-host\") pod \"7320f5e8-c79d-4707-bd0f-8a66298dbd28\" (UID: \"7320f5e8-c79d-4707-bd0f-8a66298dbd28\") " Jan 26 17:59:41 crc kubenswrapper[4754]: I0126 17:59:41.450525 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dfm47\" (UniqueName: \"kubernetes.io/projected/7320f5e8-c79d-4707-bd0f-8a66298dbd28-kube-api-access-dfm47\") pod \"7320f5e8-c79d-4707-bd0f-8a66298dbd28\" (UID: \"7320f5e8-c79d-4707-bd0f-8a66298dbd28\") " Jan 26 17:59:41 crc kubenswrapper[4754]: I0126 17:59:41.451558 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7320f5e8-c79d-4707-bd0f-8a66298dbd28-host" (OuterVolumeSpecName: "host") pod "7320f5e8-c79d-4707-bd0f-8a66298dbd28" (UID: "7320f5e8-c79d-4707-bd0f-8a66298dbd28"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 17:59:41 crc kubenswrapper[4754]: I0126 17:59:41.460006 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7320f5e8-c79d-4707-bd0f-8a66298dbd28-kube-api-access-dfm47" (OuterVolumeSpecName: "kube-api-access-dfm47") pod "7320f5e8-c79d-4707-bd0f-8a66298dbd28" (UID: "7320f5e8-c79d-4707-bd0f-8a66298dbd28"). InnerVolumeSpecName "kube-api-access-dfm47". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:59:41 crc kubenswrapper[4754]: I0126 17:59:41.554444 4754 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7320f5e8-c79d-4707-bd0f-8a66298dbd28-host\") on node \"crc\" DevicePath \"\"" Jan 26 17:59:41 crc kubenswrapper[4754]: I0126 17:59:41.554499 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dfm47\" (UniqueName: \"kubernetes.io/projected/7320f5e8-c79d-4707-bd0f-8a66298dbd28-kube-api-access-dfm47\") on node \"crc\" DevicePath \"\"" Jan 26 17:59:41 crc kubenswrapper[4754]: I0126 17:59:41.783216 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7320f5e8-c79d-4707-bd0f-8a66298dbd28" path="/var/lib/kubelet/pods/7320f5e8-c79d-4707-bd0f-8a66298dbd28/volumes" Jan 26 17:59:41 crc kubenswrapper[4754]: I0126 17:59:41.784542 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-9pk77"] Jan 26 17:59:41 crc kubenswrapper[4754]: I0126 17:59:41.804657 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-zb6c2/crc-debug-gbx5r"] Jan 26 17:59:41 crc kubenswrapper[4754]: I0126 17:59:41.806021 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-zb6c2/crc-debug-gbx5r" Jan 26 17:59:41 crc kubenswrapper[4754]: I0126 17:59:41.962393 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-68rb9\" (UniqueName: \"kubernetes.io/projected/7767b350-fc8f-4ce7-806f-cdd69586ee24-kube-api-access-68rb9\") pod \"crc-debug-gbx5r\" (UID: \"7767b350-fc8f-4ce7-806f-cdd69586ee24\") " pod="openshift-must-gather-zb6c2/crc-debug-gbx5r" Jan 26 17:59:41 crc kubenswrapper[4754]: I0126 17:59:41.962530 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7767b350-fc8f-4ce7-806f-cdd69586ee24-host\") pod \"crc-debug-gbx5r\" (UID: \"7767b350-fc8f-4ce7-806f-cdd69586ee24\") " pod="openshift-must-gather-zb6c2/crc-debug-gbx5r" Jan 26 17:59:42 crc kubenswrapper[4754]: I0126 17:59:42.063954 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7767b350-fc8f-4ce7-806f-cdd69586ee24-host\") pod \"crc-debug-gbx5r\" (UID: \"7767b350-fc8f-4ce7-806f-cdd69586ee24\") " pod="openshift-must-gather-zb6c2/crc-debug-gbx5r" Jan 26 17:59:42 crc kubenswrapper[4754]: I0126 17:59:42.064095 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7767b350-fc8f-4ce7-806f-cdd69586ee24-host\") pod \"crc-debug-gbx5r\" (UID: \"7767b350-fc8f-4ce7-806f-cdd69586ee24\") " pod="openshift-must-gather-zb6c2/crc-debug-gbx5r" Jan 26 17:59:42 crc kubenswrapper[4754]: I0126 17:59:42.064111 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-68rb9\" (UniqueName: \"kubernetes.io/projected/7767b350-fc8f-4ce7-806f-cdd69586ee24-kube-api-access-68rb9\") pod \"crc-debug-gbx5r\" (UID: \"7767b350-fc8f-4ce7-806f-cdd69586ee24\") " pod="openshift-must-gather-zb6c2/crc-debug-gbx5r" Jan 26 17:59:42 crc kubenswrapper[4754]: I0126 17:59:42.084585 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-68rb9\" (UniqueName: \"kubernetes.io/projected/7767b350-fc8f-4ce7-806f-cdd69586ee24-kube-api-access-68rb9\") pod \"crc-debug-gbx5r\" (UID: \"7767b350-fc8f-4ce7-806f-cdd69586ee24\") " pod="openshift-must-gather-zb6c2/crc-debug-gbx5r" Jan 26 17:59:42 crc kubenswrapper[4754]: I0126 17:59:42.125653 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-zb6c2/crc-debug-gbx5r" Jan 26 17:59:42 crc kubenswrapper[4754]: I0126 17:59:42.136171 4754 scope.go:117] "RemoveContainer" containerID="4fd664dac47dbc4bdab1902f5ac0867dfc6646ba19bdedb5fcc34599d67ba8aa" Jan 26 17:59:42 crc kubenswrapper[4754]: I0126 17:59:42.136176 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-zb6c2/crc-debug-pbm5p" Jan 26 17:59:42 crc kubenswrapper[4754]: I0126 17:59:42.137735 4754 generic.go:334] "Generic (PLEG): container finished" podID="2810e86d-447c-4a21-b18d-c479684410d4" containerID="6f545d64c01a1f82949a50994a589e8749c0b206d176d33d3daa34a7a65e3d14" exitCode=0 Jan 26 17:59:42 crc kubenswrapper[4754]: I0126 17:59:42.137775 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9pk77" event={"ID":"2810e86d-447c-4a21-b18d-c479684410d4","Type":"ContainerDied","Data":"6f545d64c01a1f82949a50994a589e8749c0b206d176d33d3daa34a7a65e3d14"} Jan 26 17:59:42 crc kubenswrapper[4754]: I0126 17:59:42.137818 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9pk77" event={"ID":"2810e86d-447c-4a21-b18d-c479684410d4","Type":"ContainerStarted","Data":"3fedc7852ccfd77e69747a558d28f528321166cfac1410ff3b1e60fbe4a97e71"} Jan 26 17:59:42 crc kubenswrapper[4754]: I0126 17:59:42.139510 4754 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 26 17:59:43 crc kubenswrapper[4754]: I0126 17:59:43.152698 4754 generic.go:334] "Generic (PLEG): container finished" podID="7767b350-fc8f-4ce7-806f-cdd69586ee24" containerID="70efe48eb247001d09f6324111cd286b2254d33b782dd9e5aa0b8fc46aa7fd56" exitCode=0 Jan 26 17:59:43 crc kubenswrapper[4754]: I0126 17:59:43.152782 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-zb6c2/crc-debug-gbx5r" event={"ID":"7767b350-fc8f-4ce7-806f-cdd69586ee24","Type":"ContainerDied","Data":"70efe48eb247001d09f6324111cd286b2254d33b782dd9e5aa0b8fc46aa7fd56"} Jan 26 17:59:43 crc kubenswrapper[4754]: I0126 17:59:43.152816 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-zb6c2/crc-debug-gbx5r" event={"ID":"7767b350-fc8f-4ce7-806f-cdd69586ee24","Type":"ContainerStarted","Data":"3630450cd6ba50486e258834f09b177abb5741c02f139ffe2956226a453b10d5"} Jan 26 17:59:43 crc kubenswrapper[4754]: I0126 17:59:43.195105 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-zb6c2/crc-debug-gbx5r"] Jan 26 17:59:43 crc kubenswrapper[4754]: I0126 17:59:43.202548 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-zb6c2/crc-debug-gbx5r"] Jan 26 17:59:44 crc kubenswrapper[4754]: I0126 17:59:44.164749 4754 generic.go:334] "Generic (PLEG): container finished" podID="2810e86d-447c-4a21-b18d-c479684410d4" containerID="09ddf75ec82189a0582f26cdd6f83feee5ce5fc67a2b50ace84709076f807005" exitCode=0 Jan 26 17:59:44 crc kubenswrapper[4754]: I0126 17:59:44.164942 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9pk77" event={"ID":"2810e86d-447c-4a21-b18d-c479684410d4","Type":"ContainerDied","Data":"09ddf75ec82189a0582f26cdd6f83feee5ce5fc67a2b50ace84709076f807005"} Jan 26 17:59:44 crc kubenswrapper[4754]: I0126 17:59:44.273847 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-zb6c2/crc-debug-gbx5r" Jan 26 17:59:44 crc kubenswrapper[4754]: I0126 17:59:44.410901 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-68rb9\" (UniqueName: \"kubernetes.io/projected/7767b350-fc8f-4ce7-806f-cdd69586ee24-kube-api-access-68rb9\") pod \"7767b350-fc8f-4ce7-806f-cdd69586ee24\" (UID: \"7767b350-fc8f-4ce7-806f-cdd69586ee24\") " Jan 26 17:59:44 crc kubenswrapper[4754]: I0126 17:59:44.411000 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7767b350-fc8f-4ce7-806f-cdd69586ee24-host\") pod \"7767b350-fc8f-4ce7-806f-cdd69586ee24\" (UID: \"7767b350-fc8f-4ce7-806f-cdd69586ee24\") " Jan 26 17:59:44 crc kubenswrapper[4754]: I0126 17:59:44.411094 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7767b350-fc8f-4ce7-806f-cdd69586ee24-host" (OuterVolumeSpecName: "host") pod "7767b350-fc8f-4ce7-806f-cdd69586ee24" (UID: "7767b350-fc8f-4ce7-806f-cdd69586ee24"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 17:59:44 crc kubenswrapper[4754]: I0126 17:59:44.411457 4754 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7767b350-fc8f-4ce7-806f-cdd69586ee24-host\") on node \"crc\" DevicePath \"\"" Jan 26 17:59:44 crc kubenswrapper[4754]: I0126 17:59:44.417507 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7767b350-fc8f-4ce7-806f-cdd69586ee24-kube-api-access-68rb9" (OuterVolumeSpecName: "kube-api-access-68rb9") pod "7767b350-fc8f-4ce7-806f-cdd69586ee24" (UID: "7767b350-fc8f-4ce7-806f-cdd69586ee24"). InnerVolumeSpecName "kube-api-access-68rb9". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:59:44 crc kubenswrapper[4754]: I0126 17:59:44.513260 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-68rb9\" (UniqueName: \"kubernetes.io/projected/7767b350-fc8f-4ce7-806f-cdd69586ee24-kube-api-access-68rb9\") on node \"crc\" DevicePath \"\"" Jan 26 17:59:45 crc kubenswrapper[4754]: I0126 17:59:45.173256 4754 scope.go:117] "RemoveContainer" containerID="70efe48eb247001d09f6324111cd286b2254d33b782dd9e5aa0b8fc46aa7fd56" Jan 26 17:59:45 crc kubenswrapper[4754]: I0126 17:59:45.173302 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-zb6c2/crc-debug-gbx5r" Jan 26 17:59:45 crc kubenswrapper[4754]: I0126 17:59:45.777163 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7767b350-fc8f-4ce7-806f-cdd69586ee24" path="/var/lib/kubelet/pods/7767b350-fc8f-4ce7-806f-cdd69586ee24/volumes" Jan 26 17:59:46 crc kubenswrapper[4754]: I0126 17:59:46.183701 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9pk77" event={"ID":"2810e86d-447c-4a21-b18d-c479684410d4","Type":"ContainerStarted","Data":"654b515ff7636fd0b5603a93425be5d0e4e7f245666f2f1d87b3b294f435143d"} Jan 26 17:59:46 crc kubenswrapper[4754]: I0126 17:59:46.208686 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-9pk77" podStartSLOduration=3.299330663 podStartE2EDuration="6.208641952s" podCreationTimestamp="2026-01-26 17:59:40 +0000 UTC" firstStartedPulling="2026-01-26 17:59:42.139305674 +0000 UTC m=+4348.663486108" lastFinishedPulling="2026-01-26 17:59:45.048616963 +0000 UTC m=+4351.572797397" observedRunningTime="2026-01-26 17:59:46.199694697 +0000 UTC m=+4352.723875141" watchObservedRunningTime="2026-01-26 17:59:46.208641952 +0000 UTC m=+4352.732822396" Jan 26 17:59:51 crc kubenswrapper[4754]: I0126 17:59:51.215556 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-9pk77" Jan 26 17:59:51 crc kubenswrapper[4754]: I0126 17:59:51.215949 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-9pk77" Jan 26 17:59:51 crc kubenswrapper[4754]: I0126 17:59:51.284583 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-9pk77" Jan 26 17:59:51 crc kubenswrapper[4754]: I0126 17:59:51.331838 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-9pk77" Jan 26 17:59:51 crc kubenswrapper[4754]: I0126 17:59:51.540308 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-9pk77"] Jan 26 17:59:53 crc kubenswrapper[4754]: I0126 17:59:53.255899 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-9pk77" podUID="2810e86d-447c-4a21-b18d-c479684410d4" containerName="registry-server" containerID="cri-o://654b515ff7636fd0b5603a93425be5d0e4e7f245666f2f1d87b3b294f435143d" gracePeriod=2 Jan 26 17:59:53 crc kubenswrapper[4754]: I0126 17:59:53.743854 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9pk77" Jan 26 17:59:53 crc kubenswrapper[4754]: I0126 17:59:53.895640 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2810e86d-447c-4a21-b18d-c479684410d4-catalog-content\") pod \"2810e86d-447c-4a21-b18d-c479684410d4\" (UID: \"2810e86d-447c-4a21-b18d-c479684410d4\") " Jan 26 17:59:53 crc kubenswrapper[4754]: I0126 17:59:53.895789 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2810e86d-447c-4a21-b18d-c479684410d4-utilities\") pod \"2810e86d-447c-4a21-b18d-c479684410d4\" (UID: \"2810e86d-447c-4a21-b18d-c479684410d4\") " Jan 26 17:59:53 crc kubenswrapper[4754]: I0126 17:59:53.895872 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4ht45\" (UniqueName: \"kubernetes.io/projected/2810e86d-447c-4a21-b18d-c479684410d4-kube-api-access-4ht45\") pod \"2810e86d-447c-4a21-b18d-c479684410d4\" (UID: \"2810e86d-447c-4a21-b18d-c479684410d4\") " Jan 26 17:59:53 crc kubenswrapper[4754]: I0126 17:59:53.896860 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2810e86d-447c-4a21-b18d-c479684410d4-utilities" (OuterVolumeSpecName: "utilities") pod "2810e86d-447c-4a21-b18d-c479684410d4" (UID: "2810e86d-447c-4a21-b18d-c479684410d4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:59:53 crc kubenswrapper[4754]: I0126 17:59:53.903688 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2810e86d-447c-4a21-b18d-c479684410d4-kube-api-access-4ht45" (OuterVolumeSpecName: "kube-api-access-4ht45") pod "2810e86d-447c-4a21-b18d-c479684410d4" (UID: "2810e86d-447c-4a21-b18d-c479684410d4"). InnerVolumeSpecName "kube-api-access-4ht45". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:59:53 crc kubenswrapper[4754]: I0126 17:59:53.935421 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2810e86d-447c-4a21-b18d-c479684410d4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2810e86d-447c-4a21-b18d-c479684410d4" (UID: "2810e86d-447c-4a21-b18d-c479684410d4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:59:53 crc kubenswrapper[4754]: I0126 17:59:53.998558 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4ht45\" (UniqueName: \"kubernetes.io/projected/2810e86d-447c-4a21-b18d-c479684410d4-kube-api-access-4ht45\") on node \"crc\" DevicePath \"\"" Jan 26 17:59:53 crc kubenswrapper[4754]: I0126 17:59:53.998600 4754 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2810e86d-447c-4a21-b18d-c479684410d4-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 26 17:59:53 crc kubenswrapper[4754]: I0126 17:59:53.998608 4754 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2810e86d-447c-4a21-b18d-c479684410d4-utilities\") on node \"crc\" DevicePath \"\"" Jan 26 17:59:54 crc kubenswrapper[4754]: I0126 17:59:54.265717 4754 generic.go:334] "Generic (PLEG): container finished" podID="2810e86d-447c-4a21-b18d-c479684410d4" containerID="654b515ff7636fd0b5603a93425be5d0e4e7f245666f2f1d87b3b294f435143d" exitCode=0 Jan 26 17:59:54 crc kubenswrapper[4754]: I0126 17:59:54.265769 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9pk77" event={"ID":"2810e86d-447c-4a21-b18d-c479684410d4","Type":"ContainerDied","Data":"654b515ff7636fd0b5603a93425be5d0e4e7f245666f2f1d87b3b294f435143d"} Jan 26 17:59:54 crc kubenswrapper[4754]: I0126 17:59:54.265817 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9pk77" Jan 26 17:59:54 crc kubenswrapper[4754]: I0126 17:59:54.265835 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9pk77" event={"ID":"2810e86d-447c-4a21-b18d-c479684410d4","Type":"ContainerDied","Data":"3fedc7852ccfd77e69747a558d28f528321166cfac1410ff3b1e60fbe4a97e71"} Jan 26 17:59:54 crc kubenswrapper[4754]: I0126 17:59:54.265863 4754 scope.go:117] "RemoveContainer" containerID="654b515ff7636fd0b5603a93425be5d0e4e7f245666f2f1d87b3b294f435143d" Jan 26 17:59:54 crc kubenswrapper[4754]: I0126 17:59:54.287517 4754 scope.go:117] "RemoveContainer" containerID="09ddf75ec82189a0582f26cdd6f83feee5ce5fc67a2b50ace84709076f807005" Jan 26 17:59:54 crc kubenswrapper[4754]: I0126 17:59:54.310019 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-9pk77"] Jan 26 17:59:54 crc kubenswrapper[4754]: I0126 17:59:54.321465 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-9pk77"] Jan 26 17:59:54 crc kubenswrapper[4754]: I0126 17:59:54.328123 4754 scope.go:117] "RemoveContainer" containerID="6f545d64c01a1f82949a50994a589e8749c0b206d176d33d3daa34a7a65e3d14" Jan 26 17:59:54 crc kubenswrapper[4754]: I0126 17:59:54.374379 4754 scope.go:117] "RemoveContainer" containerID="654b515ff7636fd0b5603a93425be5d0e4e7f245666f2f1d87b3b294f435143d" Jan 26 17:59:54 crc kubenswrapper[4754]: E0126 17:59:54.374940 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"654b515ff7636fd0b5603a93425be5d0e4e7f245666f2f1d87b3b294f435143d\": container with ID starting with 654b515ff7636fd0b5603a93425be5d0e4e7f245666f2f1d87b3b294f435143d not found: ID does not exist" containerID="654b515ff7636fd0b5603a93425be5d0e4e7f245666f2f1d87b3b294f435143d" Jan 26 17:59:54 crc kubenswrapper[4754]: I0126 17:59:54.374977 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"654b515ff7636fd0b5603a93425be5d0e4e7f245666f2f1d87b3b294f435143d"} err="failed to get container status \"654b515ff7636fd0b5603a93425be5d0e4e7f245666f2f1d87b3b294f435143d\": rpc error: code = NotFound desc = could not find container \"654b515ff7636fd0b5603a93425be5d0e4e7f245666f2f1d87b3b294f435143d\": container with ID starting with 654b515ff7636fd0b5603a93425be5d0e4e7f245666f2f1d87b3b294f435143d not found: ID does not exist" Jan 26 17:59:54 crc kubenswrapper[4754]: I0126 17:59:54.375001 4754 scope.go:117] "RemoveContainer" containerID="09ddf75ec82189a0582f26cdd6f83feee5ce5fc67a2b50ace84709076f807005" Jan 26 17:59:54 crc kubenswrapper[4754]: E0126 17:59:54.378756 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"09ddf75ec82189a0582f26cdd6f83feee5ce5fc67a2b50ace84709076f807005\": container with ID starting with 09ddf75ec82189a0582f26cdd6f83feee5ce5fc67a2b50ace84709076f807005 not found: ID does not exist" containerID="09ddf75ec82189a0582f26cdd6f83feee5ce5fc67a2b50ace84709076f807005" Jan 26 17:59:54 crc kubenswrapper[4754]: I0126 17:59:54.378787 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"09ddf75ec82189a0582f26cdd6f83feee5ce5fc67a2b50ace84709076f807005"} err="failed to get container status \"09ddf75ec82189a0582f26cdd6f83feee5ce5fc67a2b50ace84709076f807005\": rpc error: code = NotFound desc = could not find container \"09ddf75ec82189a0582f26cdd6f83feee5ce5fc67a2b50ace84709076f807005\": container with ID starting with 09ddf75ec82189a0582f26cdd6f83feee5ce5fc67a2b50ace84709076f807005 not found: ID does not exist" Jan 26 17:59:54 crc kubenswrapper[4754]: I0126 17:59:54.378807 4754 scope.go:117] "RemoveContainer" containerID="6f545d64c01a1f82949a50994a589e8749c0b206d176d33d3daa34a7a65e3d14" Jan 26 17:59:54 crc kubenswrapper[4754]: E0126 17:59:54.379313 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6f545d64c01a1f82949a50994a589e8749c0b206d176d33d3daa34a7a65e3d14\": container with ID starting with 6f545d64c01a1f82949a50994a589e8749c0b206d176d33d3daa34a7a65e3d14 not found: ID does not exist" containerID="6f545d64c01a1f82949a50994a589e8749c0b206d176d33d3daa34a7a65e3d14" Jan 26 17:59:54 crc kubenswrapper[4754]: I0126 17:59:54.379616 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6f545d64c01a1f82949a50994a589e8749c0b206d176d33d3daa34a7a65e3d14"} err="failed to get container status \"6f545d64c01a1f82949a50994a589e8749c0b206d176d33d3daa34a7a65e3d14\": rpc error: code = NotFound desc = could not find container \"6f545d64c01a1f82949a50994a589e8749c0b206d176d33d3daa34a7a65e3d14\": container with ID starting with 6f545d64c01a1f82949a50994a589e8749c0b206d176d33d3daa34a7a65e3d14 not found: ID does not exist" Jan 26 17:59:55 crc kubenswrapper[4754]: I0126 17:59:55.778318 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2810e86d-447c-4a21-b18d-c479684410d4" path="/var/lib/kubelet/pods/2810e86d-447c-4a21-b18d-c479684410d4/volumes" Jan 26 18:00:00 crc kubenswrapper[4754]: I0126 18:00:00.196039 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29490840-jx5lq"] Jan 26 18:00:00 crc kubenswrapper[4754]: E0126 18:00:00.196815 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7767b350-fc8f-4ce7-806f-cdd69586ee24" containerName="container-00" Jan 26 18:00:00 crc kubenswrapper[4754]: I0126 18:00:00.196832 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="7767b350-fc8f-4ce7-806f-cdd69586ee24" containerName="container-00" Jan 26 18:00:00 crc kubenswrapper[4754]: E0126 18:00:00.196843 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2810e86d-447c-4a21-b18d-c479684410d4" containerName="extract-utilities" Jan 26 18:00:00 crc kubenswrapper[4754]: I0126 18:00:00.196851 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="2810e86d-447c-4a21-b18d-c479684410d4" containerName="extract-utilities" Jan 26 18:00:00 crc kubenswrapper[4754]: E0126 18:00:00.196875 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2810e86d-447c-4a21-b18d-c479684410d4" containerName="registry-server" Jan 26 18:00:00 crc kubenswrapper[4754]: I0126 18:00:00.196883 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="2810e86d-447c-4a21-b18d-c479684410d4" containerName="registry-server" Jan 26 18:00:00 crc kubenswrapper[4754]: E0126 18:00:00.196899 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2810e86d-447c-4a21-b18d-c479684410d4" containerName="extract-content" Jan 26 18:00:00 crc kubenswrapper[4754]: I0126 18:00:00.196905 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="2810e86d-447c-4a21-b18d-c479684410d4" containerName="extract-content" Jan 26 18:00:00 crc kubenswrapper[4754]: I0126 18:00:00.197128 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="2810e86d-447c-4a21-b18d-c479684410d4" containerName="registry-server" Jan 26 18:00:00 crc kubenswrapper[4754]: I0126 18:00:00.197157 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="7767b350-fc8f-4ce7-806f-cdd69586ee24" containerName="container-00" Jan 26 18:00:00 crc kubenswrapper[4754]: I0126 18:00:00.197955 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29490840-jx5lq" Jan 26 18:00:00 crc kubenswrapper[4754]: I0126 18:00:00.201051 4754 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Jan 26 18:00:00 crc kubenswrapper[4754]: I0126 18:00:00.201276 4754 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Jan 26 18:00:00 crc kubenswrapper[4754]: I0126 18:00:00.208620 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29490840-jx5lq"] Jan 26 18:00:00 crc kubenswrapper[4754]: I0126 18:00:00.217865 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ffa2cc83-da4f-4449-9110-515629c7c128-config-volume\") pod \"collect-profiles-29490840-jx5lq\" (UID: \"ffa2cc83-da4f-4449-9110-515629c7c128\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29490840-jx5lq" Jan 26 18:00:00 crc kubenswrapper[4754]: I0126 18:00:00.217924 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ddbtv\" (UniqueName: \"kubernetes.io/projected/ffa2cc83-da4f-4449-9110-515629c7c128-kube-api-access-ddbtv\") pod \"collect-profiles-29490840-jx5lq\" (UID: \"ffa2cc83-da4f-4449-9110-515629c7c128\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29490840-jx5lq" Jan 26 18:00:00 crc kubenswrapper[4754]: I0126 18:00:00.217958 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ffa2cc83-da4f-4449-9110-515629c7c128-secret-volume\") pod \"collect-profiles-29490840-jx5lq\" (UID: \"ffa2cc83-da4f-4449-9110-515629c7c128\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29490840-jx5lq" Jan 26 18:00:00 crc kubenswrapper[4754]: I0126 18:00:00.319920 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ffa2cc83-da4f-4449-9110-515629c7c128-config-volume\") pod \"collect-profiles-29490840-jx5lq\" (UID: \"ffa2cc83-da4f-4449-9110-515629c7c128\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29490840-jx5lq" Jan 26 18:00:00 crc kubenswrapper[4754]: I0126 18:00:00.319957 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ddbtv\" (UniqueName: \"kubernetes.io/projected/ffa2cc83-da4f-4449-9110-515629c7c128-kube-api-access-ddbtv\") pod \"collect-profiles-29490840-jx5lq\" (UID: \"ffa2cc83-da4f-4449-9110-515629c7c128\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29490840-jx5lq" Jan 26 18:00:00 crc kubenswrapper[4754]: I0126 18:00:00.319981 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ffa2cc83-da4f-4449-9110-515629c7c128-secret-volume\") pod \"collect-profiles-29490840-jx5lq\" (UID: \"ffa2cc83-da4f-4449-9110-515629c7c128\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29490840-jx5lq" Jan 26 18:00:00 crc kubenswrapper[4754]: I0126 18:00:00.321048 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ffa2cc83-da4f-4449-9110-515629c7c128-config-volume\") pod \"collect-profiles-29490840-jx5lq\" (UID: \"ffa2cc83-da4f-4449-9110-515629c7c128\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29490840-jx5lq" Jan 26 18:00:00 crc kubenswrapper[4754]: I0126 18:00:00.340759 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ffa2cc83-da4f-4449-9110-515629c7c128-secret-volume\") pod \"collect-profiles-29490840-jx5lq\" (UID: \"ffa2cc83-da4f-4449-9110-515629c7c128\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29490840-jx5lq" Jan 26 18:00:00 crc kubenswrapper[4754]: I0126 18:00:00.341161 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ddbtv\" (UniqueName: \"kubernetes.io/projected/ffa2cc83-da4f-4449-9110-515629c7c128-kube-api-access-ddbtv\") pod \"collect-profiles-29490840-jx5lq\" (UID: \"ffa2cc83-da4f-4449-9110-515629c7c128\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29490840-jx5lq" Jan 26 18:00:00 crc kubenswrapper[4754]: I0126 18:00:00.530535 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29490840-jx5lq" Jan 26 18:00:00 crc kubenswrapper[4754]: I0126 18:00:00.960759 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29490840-jx5lq"] Jan 26 18:00:01 crc kubenswrapper[4754]: I0126 18:00:01.333096 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29490840-jx5lq" event={"ID":"ffa2cc83-da4f-4449-9110-515629c7c128","Type":"ContainerStarted","Data":"a64e2281c460f045ebbefaf4aa96ec2fe406bcb577ee0d51de32586ad3329163"} Jan 26 18:00:01 crc kubenswrapper[4754]: I0126 18:00:01.333150 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29490840-jx5lq" event={"ID":"ffa2cc83-da4f-4449-9110-515629c7c128","Type":"ContainerStarted","Data":"fbf45662c5236593aeec6a23a2d4b5e2f33405877eee99ffab78c01a8823e8eb"} Jan 26 18:00:01 crc kubenswrapper[4754]: I0126 18:00:01.354610 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29490840-jx5lq" podStartSLOduration=1.354590707 podStartE2EDuration="1.354590707s" podCreationTimestamp="2026-01-26 18:00:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 18:00:01.347312858 +0000 UTC m=+4367.871493292" watchObservedRunningTime="2026-01-26 18:00:01.354590707 +0000 UTC m=+4367.878771141" Jan 26 18:00:02 crc kubenswrapper[4754]: I0126 18:00:02.341480 4754 generic.go:334] "Generic (PLEG): container finished" podID="ffa2cc83-da4f-4449-9110-515629c7c128" containerID="a64e2281c460f045ebbefaf4aa96ec2fe406bcb577ee0d51de32586ad3329163" exitCode=0 Jan 26 18:00:02 crc kubenswrapper[4754]: I0126 18:00:02.341530 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29490840-jx5lq" event={"ID":"ffa2cc83-da4f-4449-9110-515629c7c128","Type":"ContainerDied","Data":"a64e2281c460f045ebbefaf4aa96ec2fe406bcb577ee0d51de32586ad3329163"} Jan 26 18:00:03 crc kubenswrapper[4754]: I0126 18:00:03.709950 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29490840-jx5lq" Jan 26 18:00:03 crc kubenswrapper[4754]: I0126 18:00:03.882332 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ddbtv\" (UniqueName: \"kubernetes.io/projected/ffa2cc83-da4f-4449-9110-515629c7c128-kube-api-access-ddbtv\") pod \"ffa2cc83-da4f-4449-9110-515629c7c128\" (UID: \"ffa2cc83-da4f-4449-9110-515629c7c128\") " Jan 26 18:00:03 crc kubenswrapper[4754]: I0126 18:00:03.882474 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ffa2cc83-da4f-4449-9110-515629c7c128-config-volume\") pod \"ffa2cc83-da4f-4449-9110-515629c7c128\" (UID: \"ffa2cc83-da4f-4449-9110-515629c7c128\") " Jan 26 18:00:03 crc kubenswrapper[4754]: I0126 18:00:03.882618 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ffa2cc83-da4f-4449-9110-515629c7c128-secret-volume\") pod \"ffa2cc83-da4f-4449-9110-515629c7c128\" (UID: \"ffa2cc83-da4f-4449-9110-515629c7c128\") " Jan 26 18:00:03 crc kubenswrapper[4754]: I0126 18:00:03.883286 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ffa2cc83-da4f-4449-9110-515629c7c128-config-volume" (OuterVolumeSpecName: "config-volume") pod "ffa2cc83-da4f-4449-9110-515629c7c128" (UID: "ffa2cc83-da4f-4449-9110-515629c7c128"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 18:00:03 crc kubenswrapper[4754]: I0126 18:00:03.885846 4754 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ffa2cc83-da4f-4449-9110-515629c7c128-config-volume\") on node \"crc\" DevicePath \"\"" Jan 26 18:00:03 crc kubenswrapper[4754]: I0126 18:00:03.899023 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ffa2cc83-da4f-4449-9110-515629c7c128-kube-api-access-ddbtv" (OuterVolumeSpecName: "kube-api-access-ddbtv") pod "ffa2cc83-da4f-4449-9110-515629c7c128" (UID: "ffa2cc83-da4f-4449-9110-515629c7c128"). InnerVolumeSpecName "kube-api-access-ddbtv". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 18:00:03 crc kubenswrapper[4754]: I0126 18:00:03.902918 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ffa2cc83-da4f-4449-9110-515629c7c128-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "ffa2cc83-da4f-4449-9110-515629c7c128" (UID: "ffa2cc83-da4f-4449-9110-515629c7c128"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 18:00:03 crc kubenswrapper[4754]: I0126 18:00:03.987272 4754 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ffa2cc83-da4f-4449-9110-515629c7c128-secret-volume\") on node \"crc\" DevicePath \"\"" Jan 26 18:00:03 crc kubenswrapper[4754]: I0126 18:00:03.987313 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ddbtv\" (UniqueName: \"kubernetes.io/projected/ffa2cc83-da4f-4449-9110-515629c7c128-kube-api-access-ddbtv\") on node \"crc\" DevicePath \"\"" Jan 26 18:00:04 crc kubenswrapper[4754]: I0126 18:00:04.359224 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29490840-jx5lq" event={"ID":"ffa2cc83-da4f-4449-9110-515629c7c128","Type":"ContainerDied","Data":"fbf45662c5236593aeec6a23a2d4b5e2f33405877eee99ffab78c01a8823e8eb"} Jan 26 18:00:04 crc kubenswrapper[4754]: I0126 18:00:04.359263 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29490840-jx5lq" Jan 26 18:00:04 crc kubenswrapper[4754]: I0126 18:00:04.359266 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fbf45662c5236593aeec6a23a2d4b5e2f33405877eee99ffab78c01a8823e8eb" Jan 26 18:00:04 crc kubenswrapper[4754]: I0126 18:00:04.437232 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29490795-zg9wx"] Jan 26 18:00:04 crc kubenswrapper[4754]: I0126 18:00:04.444972 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29490795-zg9wx"] Jan 26 18:00:05 crc kubenswrapper[4754]: I0126 18:00:05.777485 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="abd6ded6-5cbf-43d4-99d1-5e03586f1ee6" path="/var/lib/kubelet/pods/abd6ded6-5cbf-43d4-99d1-5e03586f1ee6/volumes" Jan 26 18:00:07 crc kubenswrapper[4754]: I0126 18:00:07.129201 4754 patch_prober.go:28] interesting pod/machine-config-daemon-x65wv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 26 18:00:07 crc kubenswrapper[4754]: I0126 18:00:07.129556 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 26 18:00:07 crc kubenswrapper[4754]: I0126 18:00:07.129612 4754 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" Jan 26 18:00:07 crc kubenswrapper[4754]: I0126 18:00:07.130490 4754 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"b36649b0f03300e411a658fea6f425b2be30bb8d773b1d9ef5bb8f521458f9a6"} pod="openshift-machine-config-operator/machine-config-daemon-x65wv" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 26 18:00:07 crc kubenswrapper[4754]: I0126 18:00:07.130572 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" containerName="machine-config-daemon" containerID="cri-o://b36649b0f03300e411a658fea6f425b2be30bb8d773b1d9ef5bb8f521458f9a6" gracePeriod=600 Jan 26 18:00:07 crc kubenswrapper[4754]: E0126 18:00:07.257078 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x65wv_openshift-machine-config-operator(8c3718a4-f354-4284-92e0-fdfb45a692bd)\"" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" Jan 26 18:00:07 crc kubenswrapper[4754]: E0126 18:00:07.316141 4754 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8c3718a4_f354_4284_92e0_fdfb45a692bd.slice/crio-b36649b0f03300e411a658fea6f425b2be30bb8d773b1d9ef5bb8f521458f9a6.scope\": RecentStats: unable to find data in memory cache]" Jan 26 18:00:07 crc kubenswrapper[4754]: I0126 18:00:07.390702 4754 generic.go:334] "Generic (PLEG): container finished" podID="8c3718a4-f354-4284-92e0-fdfb45a692bd" containerID="b36649b0f03300e411a658fea6f425b2be30bb8d773b1d9ef5bb8f521458f9a6" exitCode=0 Jan 26 18:00:07 crc kubenswrapper[4754]: I0126 18:00:07.390750 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" event={"ID":"8c3718a4-f354-4284-92e0-fdfb45a692bd","Type":"ContainerDied","Data":"b36649b0f03300e411a658fea6f425b2be30bb8d773b1d9ef5bb8f521458f9a6"} Jan 26 18:00:07 crc kubenswrapper[4754]: I0126 18:00:07.390783 4754 scope.go:117] "RemoveContainer" containerID="25b9a1afdbf7b01e5200dd76d80daf1edca89288d2a1ec786b655b6ff7ce5a88" Jan 26 18:00:07 crc kubenswrapper[4754]: I0126 18:00:07.394819 4754 scope.go:117] "RemoveContainer" containerID="b36649b0f03300e411a658fea6f425b2be30bb8d773b1d9ef5bb8f521458f9a6" Jan 26 18:00:07 crc kubenswrapper[4754]: E0126 18:00:07.395283 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x65wv_openshift-machine-config-operator(8c3718a4-f354-4284-92e0-fdfb45a692bd)\"" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" Jan 26 18:00:15 crc kubenswrapper[4754]: I0126 18:00:15.247304 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-5d4bd978-rj87r_8c7859b2-be6e-49c3-8392-47ed649a1f68/barbican-api/0.log" Jan 26 18:00:15 crc kubenswrapper[4754]: I0126 18:00:15.414577 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-5d4bd978-rj87r_8c7859b2-be6e-49c3-8392-47ed649a1f68/barbican-api-log/0.log" Jan 26 18:00:15 crc kubenswrapper[4754]: I0126 18:00:15.464476 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-99649c64-sldzz_0b004ba0-01f6-49f9-8d8d-56033dcd3533/barbican-keystone-listener/0.log" Jan 26 18:00:15 crc kubenswrapper[4754]: I0126 18:00:15.486921 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-99649c64-sldzz_0b004ba0-01f6-49f9-8d8d-56033dcd3533/barbican-keystone-listener-log/0.log" Jan 26 18:00:15 crc kubenswrapper[4754]: I0126 18:00:15.648053 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-5677764859-tswcr_d64625ff-6586-4c80-a720-b2febd49a966/barbican-worker-log/0.log" Jan 26 18:00:15 crc kubenswrapper[4754]: I0126 18:00:15.674317 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-5677764859-tswcr_d64625ff-6586-4c80-a720-b2febd49a966/barbican-worker/0.log" Jan 26 18:00:15 crc kubenswrapper[4754]: I0126 18:00:15.831466 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-nng92_fc4da904-699d-44a4-995d-d5ac9b05695b/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Jan 26 18:00:15 crc kubenswrapper[4754]: I0126 18:00:15.874985 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_b41f5e0e-1829-41f9-926c-ce15178e16c8/ceilometer-central-agent/0.log" Jan 26 18:00:15 crc kubenswrapper[4754]: I0126 18:00:15.937309 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_b41f5e0e-1829-41f9-926c-ce15178e16c8/ceilometer-notification-agent/0.log" Jan 26 18:00:16 crc kubenswrapper[4754]: I0126 18:00:16.054222 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_b41f5e0e-1829-41f9-926c-ce15178e16c8/proxy-httpd/0.log" Jan 26 18:00:16 crc kubenswrapper[4754]: I0126 18:00:16.084646 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_b41f5e0e-1829-41f9-926c-ce15178e16c8/sg-core/0.log" Jan 26 18:00:16 crc kubenswrapper[4754]: I0126 18:00:16.186333 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_529bf77e-614d-4354-94bd-e6383f353920/cinder-api/0.log" Jan 26 18:00:16 crc kubenswrapper[4754]: I0126 18:00:16.276657 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_529bf77e-614d-4354-94bd-e6383f353920/cinder-api-log/0.log" Jan 26 18:00:16 crc kubenswrapper[4754]: I0126 18:00:16.441124 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_6147c61d-7878-41b8-8ce6-a165c9f03ede/cinder-scheduler/0.log" Jan 26 18:00:16 crc kubenswrapper[4754]: I0126 18:00:16.486492 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_6147c61d-7878-41b8-8ce6-a165c9f03ede/probe/0.log" Jan 26 18:00:16 crc kubenswrapper[4754]: I0126 18:00:16.489390 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-5d95m_51a9ce2d-f224-449d-8988-950a60783ddb/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Jan 26 18:00:16 crc kubenswrapper[4754]: I0126 18:00:16.716011 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-v9nvp_c31e5ca9-fa72-4350-bd79-b58eb9b0e7a4/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Jan 26 18:00:16 crc kubenswrapper[4754]: I0126 18:00:16.751801 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-55478c4467-c65ql_48e65463-af70-4cca-b90b-a3b4ac9a1619/init/0.log" Jan 26 18:00:16 crc kubenswrapper[4754]: I0126 18:00:16.915838 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-55478c4467-c65ql_48e65463-af70-4cca-b90b-a3b4ac9a1619/init/0.log" Jan 26 18:00:16 crc kubenswrapper[4754]: I0126 18:00:16.962450 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-55478c4467-c65ql_48e65463-af70-4cca-b90b-a3b4ac9a1619/dnsmasq-dns/0.log" Jan 26 18:00:16 crc kubenswrapper[4754]: I0126 18:00:16.979801 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-rrhnq_55b57fed-619a-44b7-af60-f6d7a43943f1/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Jan 26 18:00:17 crc kubenswrapper[4754]: I0126 18:00:17.137540 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_724e114a-1ace-4455-846f-d7ab65c593f1/glance-httpd/0.log" Jan 26 18:00:17 crc kubenswrapper[4754]: I0126 18:00:17.171369 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_724e114a-1ace-4455-846f-d7ab65c593f1/glance-log/0.log" Jan 26 18:00:17 crc kubenswrapper[4754]: I0126 18:00:17.324031 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_c0decb48-8392-4a9a-a253-82b24b4f07ef/glance-log/0.log" Jan 26 18:00:17 crc kubenswrapper[4754]: I0126 18:00:17.359434 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_c0decb48-8392-4a9a-a253-82b24b4f07ef/glance-httpd/0.log" Jan 26 18:00:17 crc kubenswrapper[4754]: I0126 18:00:17.548653 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-6dff6969b8-5bklb_0313cdfe-e309-41aa-a5ab-83d7713628f3/horizon/0.log" Jan 26 18:00:17 crc kubenswrapper[4754]: I0126 18:00:17.642712 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-qpc2q_70a69e93-9d4a-4d51-b0c9-4b986172e2ac/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Jan 26 18:00:17 crc kubenswrapper[4754]: I0126 18:00:17.785620 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-r92d9_9bfc59f9-f503-4fec-9a31-240819fc3a52/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Jan 26 18:00:18 crc kubenswrapper[4754]: I0126 18:00:18.046510 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-6dff6969b8-5bklb_0313cdfe-e309-41aa-a5ab-83d7713628f3/horizon-log/0.log" Jan 26 18:00:18 crc kubenswrapper[4754]: I0126 18:00:18.165742 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-6c774f65c8-h55rj_928f07ae-77be-48d4-ba56-daaa5ff400f0/keystone-api/0.log" Jan 26 18:00:18 crc kubenswrapper[4754]: I0126 18:00:18.212891 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_6be68290-e783-4f39-8257-d5dc4051447a/kube-state-metrics/0.log" Jan 26 18:00:18 crc kubenswrapper[4754]: I0126 18:00:18.384528 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-lpbsc_a8d7c557-f25d-4aaa-94be-cf9e8e26bc19/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Jan 26 18:00:18 crc kubenswrapper[4754]: I0126 18:00:18.707055 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-767b9f4849-cfqbz_36eccc33-604a-49ac-a1f2-c0bd41fd053d/neutron-api/0.log" Jan 26 18:00:19 crc kubenswrapper[4754]: I0126 18:00:19.135345 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-767b9f4849-cfqbz_36eccc33-604a-49ac-a1f2-c0bd41fd053d/neutron-httpd/0.log" Jan 26 18:00:19 crc kubenswrapper[4754]: I0126 18:00:19.201194 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-wljzd_18417e65-9c20-4b5e-96df-b7b79c67433a/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Jan 26 18:00:19 crc kubenswrapper[4754]: I0126 18:00:19.767828 4754 scope.go:117] "RemoveContainer" containerID="b36649b0f03300e411a658fea6f425b2be30bb8d773b1d9ef5bb8f521458f9a6" Jan 26 18:00:19 crc kubenswrapper[4754]: E0126 18:00:19.768066 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x65wv_openshift-machine-config-operator(8c3718a4-f354-4284-92e0-fdfb45a692bd)\"" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" Jan 26 18:00:19 crc kubenswrapper[4754]: I0126 18:00:19.828310 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_0bb1f389-c4a6-47dd-8445-c7125779ef38/nova-api-log/0.log" Jan 26 18:00:19 crc kubenswrapper[4754]: I0126 18:00:19.913303 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_1d793a29-dacb-46ec-a0a8-ead07be4ce50/nova-cell0-conductor-conductor/0.log" Jan 26 18:00:20 crc kubenswrapper[4754]: I0126 18:00:20.195844 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_7a81d671-1d63-4eac-a3fd-bf0e9bf38cc2/nova-cell1-conductor-conductor/0.log" Jan 26 18:00:20 crc kubenswrapper[4754]: I0126 18:00:20.292034 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_fbd0b5ca-aa69-4a3a-a81d-5b912f5db6a1/nova-cell1-novncproxy-novncproxy/0.log" Jan 26 18:00:20 crc kubenswrapper[4754]: I0126 18:00:20.358526 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_0bb1f389-c4a6-47dd-8445-c7125779ef38/nova-api-api/0.log" Jan 26 18:00:20 crc kubenswrapper[4754]: I0126 18:00:20.414407 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-jswg2_b5408b33-3b71-46ac-87a4-413ecb9614b0/nova-edpm-deployment-openstack-edpm-ipam/0.log" Jan 26 18:00:21 crc kubenswrapper[4754]: I0126 18:00:21.065835 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_4e7eefb7-e66a-4c65-bda9-7997970991f3/nova-metadata-log/0.log" Jan 26 18:00:21 crc kubenswrapper[4754]: I0126 18:00:21.258059 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_242703dd-5a52-4da8-af40-47c3490fd6ea/mysql-bootstrap/0.log" Jan 26 18:00:21 crc kubenswrapper[4754]: I0126 18:00:21.401057 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_6f952bb4-186c-4e7c-bf01-67d8a6985319/nova-scheduler-scheduler/0.log" Jan 26 18:00:21 crc kubenswrapper[4754]: I0126 18:00:21.431040 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_242703dd-5a52-4da8-af40-47c3490fd6ea/mysql-bootstrap/0.log" Jan 26 18:00:21 crc kubenswrapper[4754]: I0126 18:00:21.537257 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_242703dd-5a52-4da8-af40-47c3490fd6ea/galera/0.log" Jan 26 18:00:21 crc kubenswrapper[4754]: I0126 18:00:21.754528 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_0390a195-4349-4dae-8ae2-72b9b16af4f7/mysql-bootstrap/0.log" Jan 26 18:00:21 crc kubenswrapper[4754]: I0126 18:00:21.824713 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_0390a195-4349-4dae-8ae2-72b9b16af4f7/mysql-bootstrap/0.log" Jan 26 18:00:21 crc kubenswrapper[4754]: I0126 18:00:21.884483 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_0390a195-4349-4dae-8ae2-72b9b16af4f7/galera/0.log" Jan 26 18:00:22 crc kubenswrapper[4754]: I0126 18:00:22.040455 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_19dee3c5-51dc-46be-8454-9c10c76b3655/openstackclient/0.log" Jan 26 18:00:22 crc kubenswrapper[4754]: I0126 18:00:22.149759 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-8bgjb_6db16a49-6566-42bf-91ad-c34be46e7800/ovn-controller/0.log" Jan 26 18:00:22 crc kubenswrapper[4754]: I0126 18:00:22.371264 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-x2tgk_28bf8197-f3d2-4ee1-9054-482fa295d92d/openstack-network-exporter/0.log" Jan 26 18:00:22 crc kubenswrapper[4754]: I0126 18:00:22.482765 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-7psgg_ee1b9e37-8fd3-4280-af09-9f2f45366870/ovsdb-server-init/0.log" Jan 26 18:00:22 crc kubenswrapper[4754]: I0126 18:00:22.615817 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_4e7eefb7-e66a-4c65-bda9-7997970991f3/nova-metadata-metadata/0.log" Jan 26 18:00:22 crc kubenswrapper[4754]: I0126 18:00:22.638005 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-7psgg_ee1b9e37-8fd3-4280-af09-9f2f45366870/ovs-vswitchd/0.log" Jan 26 18:00:22 crc kubenswrapper[4754]: I0126 18:00:22.654457 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-7psgg_ee1b9e37-8fd3-4280-af09-9f2f45366870/ovsdb-server-init/0.log" Jan 26 18:00:22 crc kubenswrapper[4754]: I0126 18:00:22.728713 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-7psgg_ee1b9e37-8fd3-4280-af09-9f2f45366870/ovsdb-server/0.log" Jan 26 18:00:22 crc kubenswrapper[4754]: I0126 18:00:22.884816 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-xsq5z_9057d2fb-d9b9-4ce9-b219-9f15c5b7f051/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Jan 26 18:00:22 crc kubenswrapper[4754]: I0126 18:00:22.903180 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_4af75510-1f61-40f7-a292-764facf90f1a/openstack-network-exporter/0.log" Jan 26 18:00:23 crc kubenswrapper[4754]: I0126 18:00:23.117523 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_4af75510-1f61-40f7-a292-764facf90f1a/ovn-northd/0.log" Jan 26 18:00:23 crc kubenswrapper[4754]: I0126 18:00:23.140473 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_85c77631-974b-42b0-a934-268213691414/openstack-network-exporter/0.log" Jan 26 18:00:23 crc kubenswrapper[4754]: I0126 18:00:23.171607 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_85c77631-974b-42b0-a934-268213691414/ovsdbserver-nb/0.log" Jan 26 18:00:23 crc kubenswrapper[4754]: I0126 18:00:23.324236 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_10afcb49-f3f2-4598-a8a8-45729720e109/ovsdbserver-sb/0.log" Jan 26 18:00:23 crc kubenswrapper[4754]: I0126 18:00:23.340357 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_10afcb49-f3f2-4598-a8a8-45729720e109/openstack-network-exporter/0.log" Jan 26 18:00:23 crc kubenswrapper[4754]: I0126 18:00:23.594382 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-577b5f64f6-qwc6n_854736cc-7859-4621-865c-69cab5f7dbeb/placement-api/0.log" Jan 26 18:00:23 crc kubenswrapper[4754]: I0126 18:00:23.650811 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_9acb97f6-82ac-4891-96dc-43a85f9c4e7c/setup-container/0.log" Jan 26 18:00:23 crc kubenswrapper[4754]: I0126 18:00:23.664541 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-577b5f64f6-qwc6n_854736cc-7859-4621-865c-69cab5f7dbeb/placement-log/0.log" Jan 26 18:00:23 crc kubenswrapper[4754]: I0126 18:00:23.978224 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_9acb97f6-82ac-4891-96dc-43a85f9c4e7c/rabbitmq/0.log" Jan 26 18:00:23 crc kubenswrapper[4754]: I0126 18:00:23.991381 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_9acb97f6-82ac-4891-96dc-43a85f9c4e7c/setup-container/0.log" Jan 26 18:00:24 crc kubenswrapper[4754]: I0126 18:00:24.004535 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_6e2da069-1c54-4801-a91c-241b80e8d17b/setup-container/0.log" Jan 26 18:00:24 crc kubenswrapper[4754]: I0126 18:00:24.217154 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_6e2da069-1c54-4801-a91c-241b80e8d17b/setup-container/0.log" Jan 26 18:00:24 crc kubenswrapper[4754]: I0126 18:00:24.282578 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_6e2da069-1c54-4801-a91c-241b80e8d17b/rabbitmq/0.log" Jan 26 18:00:24 crc kubenswrapper[4754]: I0126 18:00:24.289180 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-p4jpg_e4cbbee8-ed91-4ad6-983b-569cff60fa07/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Jan 26 18:00:24 crc kubenswrapper[4754]: I0126 18:00:24.452404 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-6w45w_09f02458-d4e3-49d4-8735-467141a57b6c/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Jan 26 18:00:24 crc kubenswrapper[4754]: I0126 18:00:24.530651 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-gn6jw_2d80ff89-c1fc-4331-9683-2740c69d001d/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Jan 26 18:00:24 crc kubenswrapper[4754]: I0126 18:00:24.667371 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-7rh8g_3dbed37c-3982-4625-929e-d99a8fd798e6/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Jan 26 18:00:24 crc kubenswrapper[4754]: I0126 18:00:24.750851 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-sqqxn_8dcb5a12-a14f-4f95-8991-5a1b9b9bb431/ssh-known-hosts-edpm-deployment/0.log" Jan 26 18:00:24 crc kubenswrapper[4754]: I0126 18:00:24.960399 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-6959bf4485-bkv7n_1622eb04-be06-4e2a-90d6-27a58ac54e60/proxy-server/0.log" Jan 26 18:00:25 crc kubenswrapper[4754]: I0126 18:00:25.064575 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-6959bf4485-bkv7n_1622eb04-be06-4e2a-90d6-27a58ac54e60/proxy-httpd/0.log" Jan 26 18:00:25 crc kubenswrapper[4754]: I0126 18:00:25.150059 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-sl9kb_cb88274a-904f-4827-9518-81b79a0e6a42/swift-ring-rebalance/0.log" Jan 26 18:00:25 crc kubenswrapper[4754]: I0126 18:00:25.285230 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_19a0730c-46d0-4029-a86b-812ba1664dcc/account-auditor/0.log" Jan 26 18:00:25 crc kubenswrapper[4754]: I0126 18:00:25.300993 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_19a0730c-46d0-4029-a86b-812ba1664dcc/account-reaper/0.log" Jan 26 18:00:25 crc kubenswrapper[4754]: I0126 18:00:25.383446 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_19a0730c-46d0-4029-a86b-812ba1664dcc/account-replicator/0.log" Jan 26 18:00:25 crc kubenswrapper[4754]: I0126 18:00:25.478398 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_19a0730c-46d0-4029-a86b-812ba1664dcc/account-server/0.log" Jan 26 18:00:25 crc kubenswrapper[4754]: I0126 18:00:25.479097 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_19a0730c-46d0-4029-a86b-812ba1664dcc/container-auditor/0.log" Jan 26 18:00:25 crc kubenswrapper[4754]: I0126 18:00:25.515201 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_19a0730c-46d0-4029-a86b-812ba1664dcc/container-replicator/0.log" Jan 26 18:00:25 crc kubenswrapper[4754]: I0126 18:00:25.621727 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_19a0730c-46d0-4029-a86b-812ba1664dcc/container-server/0.log" Jan 26 18:00:25 crc kubenswrapper[4754]: I0126 18:00:25.700860 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_19a0730c-46d0-4029-a86b-812ba1664dcc/container-updater/0.log" Jan 26 18:00:25 crc kubenswrapper[4754]: I0126 18:00:25.742386 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_19a0730c-46d0-4029-a86b-812ba1664dcc/object-expirer/0.log" Jan 26 18:00:25 crc kubenswrapper[4754]: I0126 18:00:25.760207 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_19a0730c-46d0-4029-a86b-812ba1664dcc/object-auditor/0.log" Jan 26 18:00:25 crc kubenswrapper[4754]: I0126 18:00:25.894483 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_19a0730c-46d0-4029-a86b-812ba1664dcc/object-replicator/0.log" Jan 26 18:00:25 crc kubenswrapper[4754]: I0126 18:00:25.928899 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_19a0730c-46d0-4029-a86b-812ba1664dcc/object-server/0.log" Jan 26 18:00:25 crc kubenswrapper[4754]: I0126 18:00:25.970361 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_19a0730c-46d0-4029-a86b-812ba1664dcc/object-updater/0.log" Jan 26 18:00:26 crc kubenswrapper[4754]: I0126 18:00:26.002403 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_19a0730c-46d0-4029-a86b-812ba1664dcc/rsync/0.log" Jan 26 18:00:26 crc kubenswrapper[4754]: I0126 18:00:26.111805 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_19a0730c-46d0-4029-a86b-812ba1664dcc/swift-recon-cron/0.log" Jan 26 18:00:26 crc kubenswrapper[4754]: I0126 18:00:26.265159 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-59vz4_5fade260-b289-4c35-b77e-ca9f8ec754c3/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Jan 26 18:00:26 crc kubenswrapper[4754]: I0126 18:00:26.322404 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_6f413c96-a0cb-4d11-9def-df8c3218db98/tempest-tests-tempest-tests-runner/0.log" Jan 26 18:00:26 crc kubenswrapper[4754]: I0126 18:00:26.493530 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_b323e46e-1ad6-41a9-9e7f-ad13e473f49e/test-operator-logs-container/0.log" Jan 26 18:00:26 crc kubenswrapper[4754]: I0126 18:00:26.540497 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-8vzzg_9d659728-5ede-4206-8137-8a6a62c5385c/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Jan 26 18:00:28 crc kubenswrapper[4754]: I0126 18:00:28.242345 4754 scope.go:117] "RemoveContainer" containerID="ab2be4d7b3c6674c982dfb9a730a8d5f3b38c70d808dcf1f5a3a8ebaf1dd7013" Jan 26 18:00:34 crc kubenswrapper[4754]: I0126 18:00:34.768196 4754 scope.go:117] "RemoveContainer" containerID="b36649b0f03300e411a658fea6f425b2be30bb8d773b1d9ef5bb8f521458f9a6" Jan 26 18:00:34 crc kubenswrapper[4754]: E0126 18:00:34.768780 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x65wv_openshift-machine-config-operator(8c3718a4-f354-4284-92e0-fdfb45a692bd)\"" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" Jan 26 18:00:39 crc kubenswrapper[4754]: I0126 18:00:39.711997 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_6fbed1d4-2177-40ba-a3c6-03de6fc2484f/memcached/0.log" Jan 26 18:00:49 crc kubenswrapper[4754]: I0126 18:00:49.768075 4754 scope.go:117] "RemoveContainer" containerID="b36649b0f03300e411a658fea6f425b2be30bb8d773b1d9ef5bb8f521458f9a6" Jan 26 18:00:49 crc kubenswrapper[4754]: E0126 18:00:49.768919 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x65wv_openshift-machine-config-operator(8c3718a4-f354-4284-92e0-fdfb45a692bd)\"" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" Jan 26 18:00:54 crc kubenswrapper[4754]: I0126 18:00:54.286017 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_80ea4f19923aee8b3bdba72b35be952869c3fe1246438cfe0124a675dcvvs8k_187f901c-e262-4c66-9104-7a40cb64d0e4/util/0.log" Jan 26 18:00:54 crc kubenswrapper[4754]: I0126 18:00:54.438774 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_80ea4f19923aee8b3bdba72b35be952869c3fe1246438cfe0124a675dcvvs8k_187f901c-e262-4c66-9104-7a40cb64d0e4/util/0.log" Jan 26 18:00:54 crc kubenswrapper[4754]: I0126 18:00:54.455041 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_80ea4f19923aee8b3bdba72b35be952869c3fe1246438cfe0124a675dcvvs8k_187f901c-e262-4c66-9104-7a40cb64d0e4/pull/0.log" Jan 26 18:00:54 crc kubenswrapper[4754]: I0126 18:00:54.455180 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_80ea4f19923aee8b3bdba72b35be952869c3fe1246438cfe0124a675dcvvs8k_187f901c-e262-4c66-9104-7a40cb64d0e4/pull/0.log" Jan 26 18:00:54 crc kubenswrapper[4754]: I0126 18:00:54.626072 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_80ea4f19923aee8b3bdba72b35be952869c3fe1246438cfe0124a675dcvvs8k_187f901c-e262-4c66-9104-7a40cb64d0e4/util/0.log" Jan 26 18:00:54 crc kubenswrapper[4754]: I0126 18:00:54.650210 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_80ea4f19923aee8b3bdba72b35be952869c3fe1246438cfe0124a675dcvvs8k_187f901c-e262-4c66-9104-7a40cb64d0e4/extract/0.log" Jan 26 18:00:54 crc kubenswrapper[4754]: I0126 18:00:54.653313 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_80ea4f19923aee8b3bdba72b35be952869c3fe1246438cfe0124a675dcvvs8k_187f901c-e262-4c66-9104-7a40cb64d0e4/pull/0.log" Jan 26 18:00:54 crc kubenswrapper[4754]: I0126 18:00:54.860393 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7f86f8796f-5kbpl_1a9a2d55-592f-4320-8e2e-49f65ca72dfc/manager/0.log" Jan 26 18:00:54 crc kubenswrapper[4754]: I0126 18:00:54.868961 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-7478f7dbf9-6vmhq_cb2628e4-680e-489a-8fc9-d39986c74301/manager/0.log" Jan 26 18:00:55 crc kubenswrapper[4754]: I0126 18:00:55.041337 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-b45d7bf98-bggfb_5e071de1-60dc-49d8-b965-90c2f99a6e02/manager/0.log" Jan 26 18:00:55 crc kubenswrapper[4754]: I0126 18:00:55.127555 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-78fdd796fd-c5kf9_ee98073e-2fdb-4b3e-acb8-00c71df55fa7/manager/0.log" Jan 26 18:00:55 crc kubenswrapper[4754]: I0126 18:00:55.199876 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-594c8c9d5d-hgnm9_f747ae1d-8181-4ef2-b332-b14db483aab6/manager/0.log" Jan 26 18:00:55 crc kubenswrapper[4754]: I0126 18:00:55.302192 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-77d5c5b54f-8sz77_fd6b6ccf-b7e9-41fb-b663-a3392d075880/manager/0.log" Jan 26 18:00:55 crc kubenswrapper[4754]: I0126 18:00:55.544530 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-598f7747c9-fvcgg_28161331-7731-433f-845f-2ebe1daf5fd0/manager/0.log" Jan 26 18:00:55 crc kubenswrapper[4754]: I0126 18:00:55.712545 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-694cf4f878-p66zm_5702a910-71d1-4acd-93aa-9379bc3147ce/manager/0.log" Jan 26 18:00:55 crc kubenswrapper[4754]: I0126 18:00:55.762685 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-b8b6d4659-d8db9_e3aac642-b3ee-4394-9f1d-bfac315bf162/manager/0.log" Jan 26 18:00:55 crc kubenswrapper[4754]: I0126 18:00:55.823634 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-78c6999f6f-jb5ps_16e535a9-3fc0-4385-b809-51a2bf719657/manager/0.log" Jan 26 18:00:55 crc kubenswrapper[4754]: I0126 18:00:55.979633 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-6b9fb5fdcb-fhwr4_add7fe6b-a864-439d-a3f4-80c9cc80ddc1/manager/0.log" Jan 26 18:00:56 crc kubenswrapper[4754]: I0126 18:00:56.097166 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-78d58447c5-pgmv6_2ab80530-a54c-41bf-8e4f-c895dbacb368/manager/0.log" Jan 26 18:00:56 crc kubenswrapper[4754]: I0126 18:00:56.237391 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-7bdb645866-27hg7_d23565a1-cb57-4cbe-88a2-48a01e8056f3/manager/0.log" Jan 26 18:00:56 crc kubenswrapper[4754]: I0126 18:00:56.293106 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-5f4cd88d46-fkz84_4fdc4d46-264a-4689-8ad5-0ed253f805df/manager/0.log" Jan 26 18:00:56 crc kubenswrapper[4754]: I0126 18:00:56.386640 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-6b68b8b854sclfc_f43cc0a3-41e9-4e09-837f-322a53221560/manager/0.log" Jan 26 18:00:56 crc kubenswrapper[4754]: I0126 18:00:56.576464 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-init-75cd685694-7x4dn_bfc14b7c-18c1-4442-bbb2-978889145894/operator/0.log" Jan 26 18:00:56 crc kubenswrapper[4754]: I0126 18:00:56.782267 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-zv866_39f92bf3-b95c-4950-ad79-05dedec74b02/registry-server/0.log" Jan 26 18:00:57 crc kubenswrapper[4754]: I0126 18:00:57.014149 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-6f75f45d54-kppds_8353f62d-eb53-46c3-ba4d-2d643f4f960b/manager/0.log" Jan 26 18:00:57 crc kubenswrapper[4754]: I0126 18:00:57.092754 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-79d5ccc684-gqtkb_56addd62-aeed-4139-a0db-37292a6acf8a/manager/0.log" Jan 26 18:00:57 crc kubenswrapper[4754]: I0126 18:00:57.358435 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-mqmkw_b48b5414-4034-4c0e-9d25-5053f84cf246/operator/0.log" Jan 26 18:00:57 crc kubenswrapper[4754]: I0126 18:00:57.584013 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-547cbdb99f-fmvdv_aaee86fe-018f-4055-b8c5-98e3795c53d8/manager/0.log" Jan 26 18:00:57 crc kubenswrapper[4754]: I0126 18:00:57.659628 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-85cd9769bb-dq4qw_13391684-794c-48b2-8d7f-23a122f8acc6/manager/0.log" Jan 26 18:00:57 crc kubenswrapper[4754]: I0126 18:00:57.854085 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-69797bbcbd-mr8bp_a8584d32-796d-47cd-8dd6-233374660688/manager/0.log" Jan 26 18:00:57 crc kubenswrapper[4754]: I0126 18:00:57.893237 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-59fcd4bdb5-gvt4r_93622668-0766-46f4-a216-83a7a17f36fc/manager/0.log" Jan 26 18:00:58 crc kubenswrapper[4754]: I0126 18:00:58.819548 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-564965969-k7nbl_138cc11d-6be6-43bb-994e-94ea39bb2e42/manager/0.log" Jan 26 18:01:00 crc kubenswrapper[4754]: I0126 18:01:00.160762 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29490841-jnsnh"] Jan 26 18:01:00 crc kubenswrapper[4754]: E0126 18:01:00.161385 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ffa2cc83-da4f-4449-9110-515629c7c128" containerName="collect-profiles" Jan 26 18:01:00 crc kubenswrapper[4754]: I0126 18:01:00.161398 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="ffa2cc83-da4f-4449-9110-515629c7c128" containerName="collect-profiles" Jan 26 18:01:00 crc kubenswrapper[4754]: I0126 18:01:00.161609 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="ffa2cc83-da4f-4449-9110-515629c7c128" containerName="collect-profiles" Jan 26 18:01:00 crc kubenswrapper[4754]: I0126 18:01:00.162182 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29490841-jnsnh" Jan 26 18:01:00 crc kubenswrapper[4754]: I0126 18:01:00.172376 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29490841-jnsnh"] Jan 26 18:01:00 crc kubenswrapper[4754]: I0126 18:01:00.270388 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58e32fd7-763d-484f-bc95-0c175ab37147-combined-ca-bundle\") pod \"keystone-cron-29490841-jnsnh\" (UID: \"58e32fd7-763d-484f-bc95-0c175ab37147\") " pod="openstack/keystone-cron-29490841-jnsnh" Jan 26 18:01:00 crc kubenswrapper[4754]: I0126 18:01:00.270608 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/58e32fd7-763d-484f-bc95-0c175ab37147-fernet-keys\") pod \"keystone-cron-29490841-jnsnh\" (UID: \"58e32fd7-763d-484f-bc95-0c175ab37147\") " pod="openstack/keystone-cron-29490841-jnsnh" Jan 26 18:01:00 crc kubenswrapper[4754]: I0126 18:01:00.270963 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58e32fd7-763d-484f-bc95-0c175ab37147-config-data\") pod \"keystone-cron-29490841-jnsnh\" (UID: \"58e32fd7-763d-484f-bc95-0c175ab37147\") " pod="openstack/keystone-cron-29490841-jnsnh" Jan 26 18:01:00 crc kubenswrapper[4754]: I0126 18:01:00.271011 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9lh5d\" (UniqueName: \"kubernetes.io/projected/58e32fd7-763d-484f-bc95-0c175ab37147-kube-api-access-9lh5d\") pod \"keystone-cron-29490841-jnsnh\" (UID: \"58e32fd7-763d-484f-bc95-0c175ab37147\") " pod="openstack/keystone-cron-29490841-jnsnh" Jan 26 18:01:00 crc kubenswrapper[4754]: I0126 18:01:00.372877 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58e32fd7-763d-484f-bc95-0c175ab37147-config-data\") pod \"keystone-cron-29490841-jnsnh\" (UID: \"58e32fd7-763d-484f-bc95-0c175ab37147\") " pod="openstack/keystone-cron-29490841-jnsnh" Jan 26 18:01:00 crc kubenswrapper[4754]: I0126 18:01:00.372946 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9lh5d\" (UniqueName: \"kubernetes.io/projected/58e32fd7-763d-484f-bc95-0c175ab37147-kube-api-access-9lh5d\") pod \"keystone-cron-29490841-jnsnh\" (UID: \"58e32fd7-763d-484f-bc95-0c175ab37147\") " pod="openstack/keystone-cron-29490841-jnsnh" Jan 26 18:01:00 crc kubenswrapper[4754]: I0126 18:01:00.373017 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58e32fd7-763d-484f-bc95-0c175ab37147-combined-ca-bundle\") pod \"keystone-cron-29490841-jnsnh\" (UID: \"58e32fd7-763d-484f-bc95-0c175ab37147\") " pod="openstack/keystone-cron-29490841-jnsnh" Jan 26 18:01:00 crc kubenswrapper[4754]: I0126 18:01:00.373066 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/58e32fd7-763d-484f-bc95-0c175ab37147-fernet-keys\") pod \"keystone-cron-29490841-jnsnh\" (UID: \"58e32fd7-763d-484f-bc95-0c175ab37147\") " pod="openstack/keystone-cron-29490841-jnsnh" Jan 26 18:01:00 crc kubenswrapper[4754]: I0126 18:01:00.379852 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58e32fd7-763d-484f-bc95-0c175ab37147-combined-ca-bundle\") pod \"keystone-cron-29490841-jnsnh\" (UID: \"58e32fd7-763d-484f-bc95-0c175ab37147\") " pod="openstack/keystone-cron-29490841-jnsnh" Jan 26 18:01:00 crc kubenswrapper[4754]: I0126 18:01:00.382782 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/58e32fd7-763d-484f-bc95-0c175ab37147-fernet-keys\") pod \"keystone-cron-29490841-jnsnh\" (UID: \"58e32fd7-763d-484f-bc95-0c175ab37147\") " pod="openstack/keystone-cron-29490841-jnsnh" Jan 26 18:01:00 crc kubenswrapper[4754]: I0126 18:01:00.392974 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58e32fd7-763d-484f-bc95-0c175ab37147-config-data\") pod \"keystone-cron-29490841-jnsnh\" (UID: \"58e32fd7-763d-484f-bc95-0c175ab37147\") " pod="openstack/keystone-cron-29490841-jnsnh" Jan 26 18:01:00 crc kubenswrapper[4754]: I0126 18:01:00.394268 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9lh5d\" (UniqueName: \"kubernetes.io/projected/58e32fd7-763d-484f-bc95-0c175ab37147-kube-api-access-9lh5d\") pod \"keystone-cron-29490841-jnsnh\" (UID: \"58e32fd7-763d-484f-bc95-0c175ab37147\") " pod="openstack/keystone-cron-29490841-jnsnh" Jan 26 18:01:00 crc kubenswrapper[4754]: I0126 18:01:00.521077 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29490841-jnsnh" Jan 26 18:01:00 crc kubenswrapper[4754]: I0126 18:01:00.962735 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29490841-jnsnh"] Jan 26 18:01:01 crc kubenswrapper[4754]: I0126 18:01:01.767141 4754 scope.go:117] "RemoveContainer" containerID="b36649b0f03300e411a658fea6f425b2be30bb8d773b1d9ef5bb8f521458f9a6" Jan 26 18:01:01 crc kubenswrapper[4754]: E0126 18:01:01.767621 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x65wv_openshift-machine-config-operator(8c3718a4-f354-4284-92e0-fdfb45a692bd)\"" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" Jan 26 18:01:01 crc kubenswrapper[4754]: I0126 18:01:01.928580 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29490841-jnsnh" event={"ID":"58e32fd7-763d-484f-bc95-0c175ab37147","Type":"ContainerStarted","Data":"94afaac334d5fe9f3dfd3d9540d73506bb3c9c04c6d4d4f553768b00000c8183"} Jan 26 18:01:01 crc kubenswrapper[4754]: I0126 18:01:01.928806 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29490841-jnsnh" event={"ID":"58e32fd7-763d-484f-bc95-0c175ab37147","Type":"ContainerStarted","Data":"b4c81dc02f5b7433cf8c132ead99bc94108b0a3a35675795c616fbe381f9b501"} Jan 26 18:01:01 crc kubenswrapper[4754]: I0126 18:01:01.943977 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29490841-jnsnh" podStartSLOduration=1.943956689 podStartE2EDuration="1.943956689s" podCreationTimestamp="2026-01-26 18:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 18:01:01.941023448 +0000 UTC m=+4428.465203882" watchObservedRunningTime="2026-01-26 18:01:01.943956689 +0000 UTC m=+4428.468137123" Jan 26 18:01:03 crc kubenswrapper[4754]: I0126 18:01:03.948346 4754 generic.go:334] "Generic (PLEG): container finished" podID="58e32fd7-763d-484f-bc95-0c175ab37147" containerID="94afaac334d5fe9f3dfd3d9540d73506bb3c9c04c6d4d4f553768b00000c8183" exitCode=0 Jan 26 18:01:03 crc kubenswrapper[4754]: I0126 18:01:03.948435 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29490841-jnsnh" event={"ID":"58e32fd7-763d-484f-bc95-0c175ab37147","Type":"ContainerDied","Data":"94afaac334d5fe9f3dfd3d9540d73506bb3c9c04c6d4d4f553768b00000c8183"} Jan 26 18:01:05 crc kubenswrapper[4754]: I0126 18:01:05.323806 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29490841-jnsnh" Jan 26 18:01:05 crc kubenswrapper[4754]: I0126 18:01:05.376519 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9lh5d\" (UniqueName: \"kubernetes.io/projected/58e32fd7-763d-484f-bc95-0c175ab37147-kube-api-access-9lh5d\") pod \"58e32fd7-763d-484f-bc95-0c175ab37147\" (UID: \"58e32fd7-763d-484f-bc95-0c175ab37147\") " Jan 26 18:01:05 crc kubenswrapper[4754]: I0126 18:01:05.376957 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58e32fd7-763d-484f-bc95-0c175ab37147-config-data\") pod \"58e32fd7-763d-484f-bc95-0c175ab37147\" (UID: \"58e32fd7-763d-484f-bc95-0c175ab37147\") " Jan 26 18:01:05 crc kubenswrapper[4754]: I0126 18:01:05.377070 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58e32fd7-763d-484f-bc95-0c175ab37147-combined-ca-bundle\") pod \"58e32fd7-763d-484f-bc95-0c175ab37147\" (UID: \"58e32fd7-763d-484f-bc95-0c175ab37147\") " Jan 26 18:01:05 crc kubenswrapper[4754]: I0126 18:01:05.377126 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/58e32fd7-763d-484f-bc95-0c175ab37147-fernet-keys\") pod \"58e32fd7-763d-484f-bc95-0c175ab37147\" (UID: \"58e32fd7-763d-484f-bc95-0c175ab37147\") " Jan 26 18:01:05 crc kubenswrapper[4754]: I0126 18:01:05.384644 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/58e32fd7-763d-484f-bc95-0c175ab37147-kube-api-access-9lh5d" (OuterVolumeSpecName: "kube-api-access-9lh5d") pod "58e32fd7-763d-484f-bc95-0c175ab37147" (UID: "58e32fd7-763d-484f-bc95-0c175ab37147"). InnerVolumeSpecName "kube-api-access-9lh5d". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 18:01:05 crc kubenswrapper[4754]: I0126 18:01:05.404652 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/58e32fd7-763d-484f-bc95-0c175ab37147-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "58e32fd7-763d-484f-bc95-0c175ab37147" (UID: "58e32fd7-763d-484f-bc95-0c175ab37147"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 18:01:05 crc kubenswrapper[4754]: I0126 18:01:05.446086 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/58e32fd7-763d-484f-bc95-0c175ab37147-config-data" (OuterVolumeSpecName: "config-data") pod "58e32fd7-763d-484f-bc95-0c175ab37147" (UID: "58e32fd7-763d-484f-bc95-0c175ab37147"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 18:01:05 crc kubenswrapper[4754]: I0126 18:01:05.467475 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/58e32fd7-763d-484f-bc95-0c175ab37147-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "58e32fd7-763d-484f-bc95-0c175ab37147" (UID: "58e32fd7-763d-484f-bc95-0c175ab37147"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 18:01:05 crc kubenswrapper[4754]: I0126 18:01:05.480049 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9lh5d\" (UniqueName: \"kubernetes.io/projected/58e32fd7-763d-484f-bc95-0c175ab37147-kube-api-access-9lh5d\") on node \"crc\" DevicePath \"\"" Jan 26 18:01:05 crc kubenswrapper[4754]: I0126 18:01:05.480095 4754 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58e32fd7-763d-484f-bc95-0c175ab37147-config-data\") on node \"crc\" DevicePath \"\"" Jan 26 18:01:05 crc kubenswrapper[4754]: I0126 18:01:05.480106 4754 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58e32fd7-763d-484f-bc95-0c175ab37147-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 26 18:01:05 crc kubenswrapper[4754]: I0126 18:01:05.480116 4754 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/58e32fd7-763d-484f-bc95-0c175ab37147-fernet-keys\") on node \"crc\" DevicePath \"\"" Jan 26 18:01:05 crc kubenswrapper[4754]: I0126 18:01:05.967918 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29490841-jnsnh" event={"ID":"58e32fd7-763d-484f-bc95-0c175ab37147","Type":"ContainerDied","Data":"b4c81dc02f5b7433cf8c132ead99bc94108b0a3a35675795c616fbe381f9b501"} Jan 26 18:01:05 crc kubenswrapper[4754]: I0126 18:01:05.967962 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b4c81dc02f5b7433cf8c132ead99bc94108b0a3a35675795c616fbe381f9b501" Jan 26 18:01:05 crc kubenswrapper[4754]: I0126 18:01:05.968037 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29490841-jnsnh" Jan 26 18:01:12 crc kubenswrapper[4754]: I0126 18:01:12.767632 4754 scope.go:117] "RemoveContainer" containerID="b36649b0f03300e411a658fea6f425b2be30bb8d773b1d9ef5bb8f521458f9a6" Jan 26 18:01:12 crc kubenswrapper[4754]: E0126 18:01:12.768437 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x65wv_openshift-machine-config-operator(8c3718a4-f354-4284-92e0-fdfb45a692bd)\"" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" Jan 26 18:01:18 crc kubenswrapper[4754]: I0126 18:01:18.072623 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-w7zlr_a411958e-77f0-45bb-a11c-abd88d756dee/control-plane-machine-set-operator/0.log" Jan 26 18:01:18 crc kubenswrapper[4754]: I0126 18:01:18.239241 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-g64xp_5864769c-0a13-4562-a407-f67e4348e1a6/machine-api-operator/0.log" Jan 26 18:01:18 crc kubenswrapper[4754]: I0126 18:01:18.262678 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-g64xp_5864769c-0a13-4562-a407-f67e4348e1a6/kube-rbac-proxy/0.log" Jan 26 18:01:26 crc kubenswrapper[4754]: I0126 18:01:26.767826 4754 scope.go:117] "RemoveContainer" containerID="b36649b0f03300e411a658fea6f425b2be30bb8d773b1d9ef5bb8f521458f9a6" Jan 26 18:01:26 crc kubenswrapper[4754]: E0126 18:01:26.768485 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x65wv_openshift-machine-config-operator(8c3718a4-f354-4284-92e0-fdfb45a692bd)\"" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" Jan 26 18:01:32 crc kubenswrapper[4754]: I0126 18:01:32.764824 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-858654f9db-5n7pf_b2d0aba8-8dab-483d-87cd-756af280d526/cert-manager-controller/0.log" Jan 26 18:01:32 crc kubenswrapper[4754]: I0126 18:01:32.823564 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-cf98fcc89-mksh5_9bddf956-836c-41ef-9038-f889e71b6823/cert-manager-cainjector/0.log" Jan 26 18:01:32 crc kubenswrapper[4754]: I0126 18:01:32.924696 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-687f57d79b-q26n7_d46f02a6-da6f-4128-89e8-669cdb8622c6/cert-manager-webhook/0.log" Jan 26 18:01:40 crc kubenswrapper[4754]: I0126 18:01:40.768062 4754 scope.go:117] "RemoveContainer" containerID="b36649b0f03300e411a658fea6f425b2be30bb8d773b1d9ef5bb8f521458f9a6" Jan 26 18:01:40 crc kubenswrapper[4754]: E0126 18:01:40.770325 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x65wv_openshift-machine-config-operator(8c3718a4-f354-4284-92e0-fdfb45a692bd)\"" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" Jan 26 18:01:46 crc kubenswrapper[4754]: I0126 18:01:46.362071 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-7754f76f8b-dxcqw_97017e08-dd1d-494a-b4bf-0800824588a2/nmstate-console-plugin/0.log" Jan 26 18:01:46 crc kubenswrapper[4754]: I0126 18:01:46.481081 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-qfmmc_73c39dd3-4d2e-4ee8-a218-f88eb59f36b7/nmstate-handler/0.log" Jan 26 18:01:46 crc kubenswrapper[4754]: I0126 18:01:46.541228 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-54757c584b-vvxrw_6506a03e-2445-4e3c-9814-f5bb1cfbaa06/kube-rbac-proxy/0.log" Jan 26 18:01:46 crc kubenswrapper[4754]: I0126 18:01:46.649301 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-54757c584b-vvxrw_6506a03e-2445-4e3c-9814-f5bb1cfbaa06/nmstate-metrics/0.log" Jan 26 18:01:46 crc kubenswrapper[4754]: I0126 18:01:46.723519 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-646758c888-pwgfk_9bbc5233-67ae-4cca-9a95-71da7e373005/nmstate-operator/0.log" Jan 26 18:01:46 crc kubenswrapper[4754]: I0126 18:01:46.816633 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-8474b5b9d8-zgmpc_fa03d76c-cc42-4ba9-ad6f-671b4b63dbab/nmstate-webhook/0.log" Jan 26 18:01:53 crc kubenswrapper[4754]: I0126 18:01:53.776055 4754 scope.go:117] "RemoveContainer" containerID="b36649b0f03300e411a658fea6f425b2be30bb8d773b1d9ef5bb8f521458f9a6" Jan 26 18:01:53 crc kubenswrapper[4754]: E0126 18:01:53.778474 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x65wv_openshift-machine-config-operator(8c3718a4-f354-4284-92e0-fdfb45a692bd)\"" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" Jan 26 18:02:07 crc kubenswrapper[4754]: I0126 18:02:07.767466 4754 scope.go:117] "RemoveContainer" containerID="b36649b0f03300e411a658fea6f425b2be30bb8d773b1d9ef5bb8f521458f9a6" Jan 26 18:02:07 crc kubenswrapper[4754]: E0126 18:02:07.768220 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x65wv_openshift-machine-config-operator(8c3718a4-f354-4284-92e0-fdfb45a692bd)\"" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" Jan 26 18:02:15 crc kubenswrapper[4754]: I0126 18:02:15.216252 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6968d8fdc4-8458x_dadf6410-9227-422a-8954-a2488091b15e/kube-rbac-proxy/0.log" Jan 26 18:02:15 crc kubenswrapper[4754]: I0126 18:02:15.312280 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6968d8fdc4-8458x_dadf6410-9227-422a-8954-a2488091b15e/controller/0.log" Jan 26 18:02:15 crc kubenswrapper[4754]: I0126 18:02:15.471916 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5h7fb_9f0c3994-3669-4cb5-89b3-b9e9dbc316de/cp-frr-files/0.log" Jan 26 18:02:15 crc kubenswrapper[4754]: I0126 18:02:15.636318 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5h7fb_9f0c3994-3669-4cb5-89b3-b9e9dbc316de/cp-frr-files/0.log" Jan 26 18:02:15 crc kubenswrapper[4754]: I0126 18:02:15.647859 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5h7fb_9f0c3994-3669-4cb5-89b3-b9e9dbc316de/cp-metrics/0.log" Jan 26 18:02:15 crc kubenswrapper[4754]: I0126 18:02:15.649160 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5h7fb_9f0c3994-3669-4cb5-89b3-b9e9dbc316de/cp-reloader/0.log" Jan 26 18:02:15 crc kubenswrapper[4754]: I0126 18:02:15.670651 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5h7fb_9f0c3994-3669-4cb5-89b3-b9e9dbc316de/cp-reloader/0.log" Jan 26 18:02:15 crc kubenswrapper[4754]: I0126 18:02:15.832477 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5h7fb_9f0c3994-3669-4cb5-89b3-b9e9dbc316de/cp-frr-files/0.log" Jan 26 18:02:15 crc kubenswrapper[4754]: I0126 18:02:15.902581 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5h7fb_9f0c3994-3669-4cb5-89b3-b9e9dbc316de/cp-metrics/0.log" Jan 26 18:02:15 crc kubenswrapper[4754]: I0126 18:02:15.902887 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5h7fb_9f0c3994-3669-4cb5-89b3-b9e9dbc316de/cp-metrics/0.log" Jan 26 18:02:15 crc kubenswrapper[4754]: I0126 18:02:15.946605 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5h7fb_9f0c3994-3669-4cb5-89b3-b9e9dbc316de/cp-reloader/0.log" Jan 26 18:02:16 crc kubenswrapper[4754]: I0126 18:02:16.172617 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5h7fb_9f0c3994-3669-4cb5-89b3-b9e9dbc316de/controller/0.log" Jan 26 18:02:16 crc kubenswrapper[4754]: I0126 18:02:16.180915 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5h7fb_9f0c3994-3669-4cb5-89b3-b9e9dbc316de/cp-frr-files/0.log" Jan 26 18:02:16 crc kubenswrapper[4754]: I0126 18:02:16.223934 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5h7fb_9f0c3994-3669-4cb5-89b3-b9e9dbc316de/cp-metrics/0.log" Jan 26 18:02:16 crc kubenswrapper[4754]: I0126 18:02:16.228293 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5h7fb_9f0c3994-3669-4cb5-89b3-b9e9dbc316de/cp-reloader/0.log" Jan 26 18:02:16 crc kubenswrapper[4754]: I0126 18:02:16.401241 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5h7fb_9f0c3994-3669-4cb5-89b3-b9e9dbc316de/frr-metrics/0.log" Jan 26 18:02:16 crc kubenswrapper[4754]: I0126 18:02:16.426622 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5h7fb_9f0c3994-3669-4cb5-89b3-b9e9dbc316de/kube-rbac-proxy-frr/0.log" Jan 26 18:02:16 crc kubenswrapper[4754]: I0126 18:02:16.485529 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5h7fb_9f0c3994-3669-4cb5-89b3-b9e9dbc316de/kube-rbac-proxy/0.log" Jan 26 18:02:16 crc kubenswrapper[4754]: I0126 18:02:16.677532 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5h7fb_9f0c3994-3669-4cb5-89b3-b9e9dbc316de/reloader/0.log" Jan 26 18:02:16 crc kubenswrapper[4754]: I0126 18:02:16.732761 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7df86c4f6c-zbwg9_941e5b29-8892-451b-9ba6-afdad3c1c77b/frr-k8s-webhook-server/0.log" Jan 26 18:02:17 crc kubenswrapper[4754]: I0126 18:02:17.038476 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-755bc5d786-g2p48_83c99759-10ef-42eb-a58f-23c4e6c7d089/manager/0.log" Jan 26 18:02:17 crc kubenswrapper[4754]: I0126 18:02:17.294055 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-755c486f67-b6mn4_ee80d5f4-a13e-434d-b250-1e005d84dc59/webhook-server/0.log" Jan 26 18:02:17 crc kubenswrapper[4754]: I0126 18:02:17.455611 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-drvw7_b3234c0c-afa9-41b4-88e3-70f44df6c2ce/kube-rbac-proxy/0.log" Jan 26 18:02:17 crc kubenswrapper[4754]: I0126 18:02:17.554578 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5h7fb_9f0c3994-3669-4cb5-89b3-b9e9dbc316de/frr/0.log" Jan 26 18:02:17 crc kubenswrapper[4754]: I0126 18:02:17.853991 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-drvw7_b3234c0c-afa9-41b4-88e3-70f44df6c2ce/speaker/0.log" Jan 26 18:02:19 crc kubenswrapper[4754]: I0126 18:02:19.766951 4754 scope.go:117] "RemoveContainer" containerID="b36649b0f03300e411a658fea6f425b2be30bb8d773b1d9ef5bb8f521458f9a6" Jan 26 18:02:19 crc kubenswrapper[4754]: E0126 18:02:19.767479 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x65wv_openshift-machine-config-operator(8c3718a4-f354-4284-92e0-fdfb45a692bd)\"" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" Jan 26 18:02:32 crc kubenswrapper[4754]: I0126 18:02:32.913930 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dckzdjj_997bf4f3-9268-4058-9ee7-80c581651bcd/util/0.log" Jan 26 18:02:33 crc kubenswrapper[4754]: I0126 18:02:33.139656 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dckzdjj_997bf4f3-9268-4058-9ee7-80c581651bcd/pull/0.log" Jan 26 18:02:33 crc kubenswrapper[4754]: I0126 18:02:33.142002 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dckzdjj_997bf4f3-9268-4058-9ee7-80c581651bcd/util/0.log" Jan 26 18:02:33 crc kubenswrapper[4754]: I0126 18:02:33.173829 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dckzdjj_997bf4f3-9268-4058-9ee7-80c581651bcd/pull/0.log" Jan 26 18:02:33 crc kubenswrapper[4754]: I0126 18:02:33.373945 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dckzdjj_997bf4f3-9268-4058-9ee7-80c581651bcd/util/0.log" Jan 26 18:02:33 crc kubenswrapper[4754]: I0126 18:02:33.378059 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dckzdjj_997bf4f3-9268-4058-9ee7-80c581651bcd/pull/0.log" Jan 26 18:02:33 crc kubenswrapper[4754]: I0126 18:02:33.395512 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dckzdjj_997bf4f3-9268-4058-9ee7-80c581651bcd/extract/0.log" Jan 26 18:02:34 crc kubenswrapper[4754]: I0126 18:02:34.143593 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713v2nvl_4212a345-63c7-4b01-b13d-12f9a3fc297a/util/0.log" Jan 26 18:02:34 crc kubenswrapper[4754]: I0126 18:02:34.345492 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713v2nvl_4212a345-63c7-4b01-b13d-12f9a3fc297a/pull/0.log" Jan 26 18:02:34 crc kubenswrapper[4754]: I0126 18:02:34.371752 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713v2nvl_4212a345-63c7-4b01-b13d-12f9a3fc297a/pull/0.log" Jan 26 18:02:34 crc kubenswrapper[4754]: I0126 18:02:34.382511 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713v2nvl_4212a345-63c7-4b01-b13d-12f9a3fc297a/util/0.log" Jan 26 18:02:34 crc kubenswrapper[4754]: I0126 18:02:34.587085 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713v2nvl_4212a345-63c7-4b01-b13d-12f9a3fc297a/util/0.log" Jan 26 18:02:34 crc kubenswrapper[4754]: I0126 18:02:34.620988 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713v2nvl_4212a345-63c7-4b01-b13d-12f9a3fc297a/pull/0.log" Jan 26 18:02:34 crc kubenswrapper[4754]: I0126 18:02:34.645555 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713v2nvl_4212a345-63c7-4b01-b13d-12f9a3fc297a/extract/0.log" Jan 26 18:02:34 crc kubenswrapper[4754]: I0126 18:02:34.761374 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-jvfdk_ceb4005f-7a24-4c06-8afd-5cad6c195c57/extract-utilities/0.log" Jan 26 18:02:34 crc kubenswrapper[4754]: I0126 18:02:34.767278 4754 scope.go:117] "RemoveContainer" containerID="b36649b0f03300e411a658fea6f425b2be30bb8d773b1d9ef5bb8f521458f9a6" Jan 26 18:02:34 crc kubenswrapper[4754]: E0126 18:02:34.767655 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x65wv_openshift-machine-config-operator(8c3718a4-f354-4284-92e0-fdfb45a692bd)\"" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" Jan 26 18:02:34 crc kubenswrapper[4754]: I0126 18:02:34.943496 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-jvfdk_ceb4005f-7a24-4c06-8afd-5cad6c195c57/extract-utilities/0.log" Jan 26 18:02:35 crc kubenswrapper[4754]: I0126 18:02:35.013674 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-jvfdk_ceb4005f-7a24-4c06-8afd-5cad6c195c57/extract-content/0.log" Jan 26 18:02:35 crc kubenswrapper[4754]: I0126 18:02:35.051917 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-jvfdk_ceb4005f-7a24-4c06-8afd-5cad6c195c57/extract-content/0.log" Jan 26 18:02:35 crc kubenswrapper[4754]: I0126 18:02:35.260522 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-jvfdk_ceb4005f-7a24-4c06-8afd-5cad6c195c57/extract-utilities/0.log" Jan 26 18:02:35 crc kubenswrapper[4754]: I0126 18:02:35.270241 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-jvfdk_ceb4005f-7a24-4c06-8afd-5cad6c195c57/extract-content/0.log" Jan 26 18:02:35 crc kubenswrapper[4754]: I0126 18:02:35.498742 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-c44vn_02fabee4-adea-4bca-ba0a-e6b98f6e68dd/extract-utilities/0.log" Jan 26 18:02:35 crc kubenswrapper[4754]: I0126 18:02:35.632345 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-jvfdk_ceb4005f-7a24-4c06-8afd-5cad6c195c57/registry-server/0.log" Jan 26 18:02:35 crc kubenswrapper[4754]: I0126 18:02:35.731714 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-c44vn_02fabee4-adea-4bca-ba0a-e6b98f6e68dd/extract-utilities/0.log" Jan 26 18:02:35 crc kubenswrapper[4754]: I0126 18:02:35.756253 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-c44vn_02fabee4-adea-4bca-ba0a-e6b98f6e68dd/extract-content/0.log" Jan 26 18:02:35 crc kubenswrapper[4754]: I0126 18:02:35.816084 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-c44vn_02fabee4-adea-4bca-ba0a-e6b98f6e68dd/extract-content/0.log" Jan 26 18:02:35 crc kubenswrapper[4754]: I0126 18:02:35.913638 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-c44vn_02fabee4-adea-4bca-ba0a-e6b98f6e68dd/extract-utilities/0.log" Jan 26 18:02:35 crc kubenswrapper[4754]: I0126 18:02:35.979253 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-c44vn_02fabee4-adea-4bca-ba0a-e6b98f6e68dd/extract-content/0.log" Jan 26 18:02:36 crc kubenswrapper[4754]: I0126 18:02:36.231154 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-mp4qz_eedf4e24-8d2e-4ec6-9caa-f5af47592b89/marketplace-operator/0.log" Jan 26 18:02:36 crc kubenswrapper[4754]: I0126 18:02:36.274094 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-psxqw_41e45928-06b9-4854-9061-16e4053b0fb1/extract-utilities/0.log" Jan 26 18:02:36 crc kubenswrapper[4754]: I0126 18:02:36.575427 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-psxqw_41e45928-06b9-4854-9061-16e4053b0fb1/extract-content/0.log" Jan 26 18:02:36 crc kubenswrapper[4754]: I0126 18:02:36.604900 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-psxqw_41e45928-06b9-4854-9061-16e4053b0fb1/extract-content/0.log" Jan 26 18:02:36 crc kubenswrapper[4754]: I0126 18:02:36.640596 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-psxqw_41e45928-06b9-4854-9061-16e4053b0fb1/extract-utilities/0.log" Jan 26 18:02:36 crc kubenswrapper[4754]: I0126 18:02:36.687073 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-c44vn_02fabee4-adea-4bca-ba0a-e6b98f6e68dd/registry-server/0.log" Jan 26 18:02:36 crc kubenswrapper[4754]: I0126 18:02:36.796538 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-psxqw_41e45928-06b9-4854-9061-16e4053b0fb1/extract-content/0.log" Jan 26 18:02:36 crc kubenswrapper[4754]: I0126 18:02:36.840374 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-psxqw_41e45928-06b9-4854-9061-16e4053b0fb1/extract-utilities/0.log" Jan 26 18:02:36 crc kubenswrapper[4754]: I0126 18:02:36.918813 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-dn77h_2dfa533f-5161-4ef0-bc9b-397abae75b23/extract-utilities/0.log" Jan 26 18:02:37 crc kubenswrapper[4754]: I0126 18:02:37.034037 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-psxqw_41e45928-06b9-4854-9061-16e4053b0fb1/registry-server/0.log" Jan 26 18:02:37 crc kubenswrapper[4754]: I0126 18:02:37.102554 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-dn77h_2dfa533f-5161-4ef0-bc9b-397abae75b23/extract-content/0.log" Jan 26 18:02:37 crc kubenswrapper[4754]: I0126 18:02:37.182543 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-dn77h_2dfa533f-5161-4ef0-bc9b-397abae75b23/extract-utilities/0.log" Jan 26 18:02:37 crc kubenswrapper[4754]: I0126 18:02:37.188878 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-dn77h_2dfa533f-5161-4ef0-bc9b-397abae75b23/extract-content/0.log" Jan 26 18:02:37 crc kubenswrapper[4754]: I0126 18:02:37.388507 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-dn77h_2dfa533f-5161-4ef0-bc9b-397abae75b23/extract-utilities/0.log" Jan 26 18:02:37 crc kubenswrapper[4754]: I0126 18:02:37.429408 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-dn77h_2dfa533f-5161-4ef0-bc9b-397abae75b23/extract-content/0.log" Jan 26 18:02:37 crc kubenswrapper[4754]: I0126 18:02:37.961941 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-dn77h_2dfa533f-5161-4ef0-bc9b-397abae75b23/registry-server/0.log" Jan 26 18:02:49 crc kubenswrapper[4754]: I0126 18:02:49.767883 4754 scope.go:117] "RemoveContainer" containerID="b36649b0f03300e411a658fea6f425b2be30bb8d773b1d9ef5bb8f521458f9a6" Jan 26 18:02:49 crc kubenswrapper[4754]: E0126 18:02:49.768554 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x65wv_openshift-machine-config-operator(8c3718a4-f354-4284-92e0-fdfb45a692bd)\"" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" Jan 26 18:03:03 crc kubenswrapper[4754]: I0126 18:03:03.767225 4754 scope.go:117] "RemoveContainer" containerID="b36649b0f03300e411a658fea6f425b2be30bb8d773b1d9ef5bb8f521458f9a6" Jan 26 18:03:03 crc kubenswrapper[4754]: E0126 18:03:03.772357 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x65wv_openshift-machine-config-operator(8c3718a4-f354-4284-92e0-fdfb45a692bd)\"" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" Jan 26 18:03:15 crc kubenswrapper[4754]: I0126 18:03:15.767979 4754 scope.go:117] "RemoveContainer" containerID="b36649b0f03300e411a658fea6f425b2be30bb8d773b1d9ef5bb8f521458f9a6" Jan 26 18:03:15 crc kubenswrapper[4754]: E0126 18:03:15.768607 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x65wv_openshift-machine-config-operator(8c3718a4-f354-4284-92e0-fdfb45a692bd)\"" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" Jan 26 18:03:28 crc kubenswrapper[4754]: I0126 18:03:28.767579 4754 scope.go:117] "RemoveContainer" containerID="b36649b0f03300e411a658fea6f425b2be30bb8d773b1d9ef5bb8f521458f9a6" Jan 26 18:03:28 crc kubenswrapper[4754]: E0126 18:03:28.768283 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x65wv_openshift-machine-config-operator(8c3718a4-f354-4284-92e0-fdfb45a692bd)\"" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" Jan 26 18:03:38 crc kubenswrapper[4754]: I0126 18:03:38.773039 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-jqlp7"] Jan 26 18:03:38 crc kubenswrapper[4754]: E0126 18:03:38.773988 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58e32fd7-763d-484f-bc95-0c175ab37147" containerName="keystone-cron" Jan 26 18:03:38 crc kubenswrapper[4754]: I0126 18:03:38.774003 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="58e32fd7-763d-484f-bc95-0c175ab37147" containerName="keystone-cron" Jan 26 18:03:38 crc kubenswrapper[4754]: I0126 18:03:38.774221 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="58e32fd7-763d-484f-bc95-0c175ab37147" containerName="keystone-cron" Jan 26 18:03:38 crc kubenswrapper[4754]: I0126 18:03:38.775526 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jqlp7" Jan 26 18:03:38 crc kubenswrapper[4754]: I0126 18:03:38.791653 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-jqlp7"] Jan 26 18:03:38 crc kubenswrapper[4754]: I0126 18:03:38.921495 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3868c9e0-3e22-4992-b0eb-d261f67bfc58-utilities\") pod \"community-operators-jqlp7\" (UID: \"3868c9e0-3e22-4992-b0eb-d261f67bfc58\") " pod="openshift-marketplace/community-operators-jqlp7" Jan 26 18:03:38 crc kubenswrapper[4754]: I0126 18:03:38.921568 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-btv5b\" (UniqueName: \"kubernetes.io/projected/3868c9e0-3e22-4992-b0eb-d261f67bfc58-kube-api-access-btv5b\") pod \"community-operators-jqlp7\" (UID: \"3868c9e0-3e22-4992-b0eb-d261f67bfc58\") " pod="openshift-marketplace/community-operators-jqlp7" Jan 26 18:03:38 crc kubenswrapper[4754]: I0126 18:03:38.921619 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3868c9e0-3e22-4992-b0eb-d261f67bfc58-catalog-content\") pod \"community-operators-jqlp7\" (UID: \"3868c9e0-3e22-4992-b0eb-d261f67bfc58\") " pod="openshift-marketplace/community-operators-jqlp7" Jan 26 18:03:39 crc kubenswrapper[4754]: I0126 18:03:39.024028 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3868c9e0-3e22-4992-b0eb-d261f67bfc58-utilities\") pod \"community-operators-jqlp7\" (UID: \"3868c9e0-3e22-4992-b0eb-d261f67bfc58\") " pod="openshift-marketplace/community-operators-jqlp7" Jan 26 18:03:39 crc kubenswrapper[4754]: I0126 18:03:39.024117 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-btv5b\" (UniqueName: \"kubernetes.io/projected/3868c9e0-3e22-4992-b0eb-d261f67bfc58-kube-api-access-btv5b\") pod \"community-operators-jqlp7\" (UID: \"3868c9e0-3e22-4992-b0eb-d261f67bfc58\") " pod="openshift-marketplace/community-operators-jqlp7" Jan 26 18:03:39 crc kubenswrapper[4754]: I0126 18:03:39.024184 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3868c9e0-3e22-4992-b0eb-d261f67bfc58-catalog-content\") pod \"community-operators-jqlp7\" (UID: \"3868c9e0-3e22-4992-b0eb-d261f67bfc58\") " pod="openshift-marketplace/community-operators-jqlp7" Jan 26 18:03:39 crc kubenswrapper[4754]: I0126 18:03:39.024714 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3868c9e0-3e22-4992-b0eb-d261f67bfc58-utilities\") pod \"community-operators-jqlp7\" (UID: \"3868c9e0-3e22-4992-b0eb-d261f67bfc58\") " pod="openshift-marketplace/community-operators-jqlp7" Jan 26 18:03:39 crc kubenswrapper[4754]: I0126 18:03:39.024877 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3868c9e0-3e22-4992-b0eb-d261f67bfc58-catalog-content\") pod \"community-operators-jqlp7\" (UID: \"3868c9e0-3e22-4992-b0eb-d261f67bfc58\") " pod="openshift-marketplace/community-operators-jqlp7" Jan 26 18:03:39 crc kubenswrapper[4754]: I0126 18:03:39.046083 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-btv5b\" (UniqueName: \"kubernetes.io/projected/3868c9e0-3e22-4992-b0eb-d261f67bfc58-kube-api-access-btv5b\") pod \"community-operators-jqlp7\" (UID: \"3868c9e0-3e22-4992-b0eb-d261f67bfc58\") " pod="openshift-marketplace/community-operators-jqlp7" Jan 26 18:03:39 crc kubenswrapper[4754]: I0126 18:03:39.099543 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jqlp7" Jan 26 18:03:39 crc kubenswrapper[4754]: I0126 18:03:39.746658 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-jqlp7"] Jan 26 18:03:40 crc kubenswrapper[4754]: I0126 18:03:40.555258 4754 generic.go:334] "Generic (PLEG): container finished" podID="3868c9e0-3e22-4992-b0eb-d261f67bfc58" containerID="4c7d1c9acf262d9885d83adec1f3a02bf1ca12bb3188ba2d286953fd0d99f18f" exitCode=0 Jan 26 18:03:40 crc kubenswrapper[4754]: I0126 18:03:40.555558 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jqlp7" event={"ID":"3868c9e0-3e22-4992-b0eb-d261f67bfc58","Type":"ContainerDied","Data":"4c7d1c9acf262d9885d83adec1f3a02bf1ca12bb3188ba2d286953fd0d99f18f"} Jan 26 18:03:40 crc kubenswrapper[4754]: I0126 18:03:40.555585 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jqlp7" event={"ID":"3868c9e0-3e22-4992-b0eb-d261f67bfc58","Type":"ContainerStarted","Data":"5215a44c9f1ccff0ef0b45436b12e1b46784e0558f297f36fccdc7ca114e2c54"} Jan 26 18:03:41 crc kubenswrapper[4754]: I0126 18:03:41.566178 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jqlp7" event={"ID":"3868c9e0-3e22-4992-b0eb-d261f67bfc58","Type":"ContainerStarted","Data":"c2d2d541923f0e26e09f7415c3401764bab031a77c574d5de66c96cac6ef289c"} Jan 26 18:03:41 crc kubenswrapper[4754]: I0126 18:03:41.767581 4754 scope.go:117] "RemoveContainer" containerID="b36649b0f03300e411a658fea6f425b2be30bb8d773b1d9ef5bb8f521458f9a6" Jan 26 18:03:41 crc kubenswrapper[4754]: E0126 18:03:41.768019 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x65wv_openshift-machine-config-operator(8c3718a4-f354-4284-92e0-fdfb45a692bd)\"" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" Jan 26 18:03:42 crc kubenswrapper[4754]: I0126 18:03:42.577093 4754 generic.go:334] "Generic (PLEG): container finished" podID="3868c9e0-3e22-4992-b0eb-d261f67bfc58" containerID="c2d2d541923f0e26e09f7415c3401764bab031a77c574d5de66c96cac6ef289c" exitCode=0 Jan 26 18:03:42 crc kubenswrapper[4754]: I0126 18:03:42.577187 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jqlp7" event={"ID":"3868c9e0-3e22-4992-b0eb-d261f67bfc58","Type":"ContainerDied","Data":"c2d2d541923f0e26e09f7415c3401764bab031a77c574d5de66c96cac6ef289c"} Jan 26 18:03:43 crc kubenswrapper[4754]: I0126 18:03:43.597224 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jqlp7" event={"ID":"3868c9e0-3e22-4992-b0eb-d261f67bfc58","Type":"ContainerStarted","Data":"666ab54552ef4680c43e4624572539cedfb7de384cc81dfd8c81ca5eba9cb02c"} Jan 26 18:03:49 crc kubenswrapper[4754]: I0126 18:03:49.099933 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-jqlp7" Jan 26 18:03:49 crc kubenswrapper[4754]: I0126 18:03:49.100480 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-jqlp7" Jan 26 18:03:49 crc kubenswrapper[4754]: I0126 18:03:49.163723 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-jqlp7" Jan 26 18:03:49 crc kubenswrapper[4754]: I0126 18:03:49.190973 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-jqlp7" podStartSLOduration=8.734706254 podStartE2EDuration="11.190941529s" podCreationTimestamp="2026-01-26 18:03:38 +0000 UTC" firstStartedPulling="2026-01-26 18:03:40.557577798 +0000 UTC m=+4587.081758242" lastFinishedPulling="2026-01-26 18:03:43.013813083 +0000 UTC m=+4589.537993517" observedRunningTime="2026-01-26 18:03:43.620476668 +0000 UTC m=+4590.144657112" watchObservedRunningTime="2026-01-26 18:03:49.190941529 +0000 UTC m=+4595.715122003" Jan 26 18:03:49 crc kubenswrapper[4754]: I0126 18:03:49.739839 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-jqlp7" Jan 26 18:03:49 crc kubenswrapper[4754]: I0126 18:03:49.803019 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-jqlp7"] Jan 26 18:03:51 crc kubenswrapper[4754]: I0126 18:03:51.685270 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-jqlp7" podUID="3868c9e0-3e22-4992-b0eb-d261f67bfc58" containerName="registry-server" containerID="cri-o://666ab54552ef4680c43e4624572539cedfb7de384cc81dfd8c81ca5eba9cb02c" gracePeriod=2 Jan 26 18:03:52 crc kubenswrapper[4754]: I0126 18:03:52.693801 4754 generic.go:334] "Generic (PLEG): container finished" podID="3868c9e0-3e22-4992-b0eb-d261f67bfc58" containerID="666ab54552ef4680c43e4624572539cedfb7de384cc81dfd8c81ca5eba9cb02c" exitCode=0 Jan 26 18:03:52 crc kubenswrapper[4754]: I0126 18:03:52.694032 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jqlp7" event={"ID":"3868c9e0-3e22-4992-b0eb-d261f67bfc58","Type":"ContainerDied","Data":"666ab54552ef4680c43e4624572539cedfb7de384cc81dfd8c81ca5eba9cb02c"} Jan 26 18:03:52 crc kubenswrapper[4754]: I0126 18:03:52.694152 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jqlp7" event={"ID":"3868c9e0-3e22-4992-b0eb-d261f67bfc58","Type":"ContainerDied","Data":"5215a44c9f1ccff0ef0b45436b12e1b46784e0558f297f36fccdc7ca114e2c54"} Jan 26 18:03:52 crc kubenswrapper[4754]: I0126 18:03:52.694167 4754 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5215a44c9f1ccff0ef0b45436b12e1b46784e0558f297f36fccdc7ca114e2c54" Jan 26 18:03:52 crc kubenswrapper[4754]: I0126 18:03:52.757897 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jqlp7" Jan 26 18:03:52 crc kubenswrapper[4754]: I0126 18:03:52.865321 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3868c9e0-3e22-4992-b0eb-d261f67bfc58-catalog-content\") pod \"3868c9e0-3e22-4992-b0eb-d261f67bfc58\" (UID: \"3868c9e0-3e22-4992-b0eb-d261f67bfc58\") " Jan 26 18:03:52 crc kubenswrapper[4754]: I0126 18:03:52.865404 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-btv5b\" (UniqueName: \"kubernetes.io/projected/3868c9e0-3e22-4992-b0eb-d261f67bfc58-kube-api-access-btv5b\") pod \"3868c9e0-3e22-4992-b0eb-d261f67bfc58\" (UID: \"3868c9e0-3e22-4992-b0eb-d261f67bfc58\") " Jan 26 18:03:52 crc kubenswrapper[4754]: I0126 18:03:52.865485 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3868c9e0-3e22-4992-b0eb-d261f67bfc58-utilities\") pod \"3868c9e0-3e22-4992-b0eb-d261f67bfc58\" (UID: \"3868c9e0-3e22-4992-b0eb-d261f67bfc58\") " Jan 26 18:03:52 crc kubenswrapper[4754]: I0126 18:03:52.866489 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3868c9e0-3e22-4992-b0eb-d261f67bfc58-utilities" (OuterVolumeSpecName: "utilities") pod "3868c9e0-3e22-4992-b0eb-d261f67bfc58" (UID: "3868c9e0-3e22-4992-b0eb-d261f67bfc58"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 18:03:52 crc kubenswrapper[4754]: I0126 18:03:52.877949 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3868c9e0-3e22-4992-b0eb-d261f67bfc58-kube-api-access-btv5b" (OuterVolumeSpecName: "kube-api-access-btv5b") pod "3868c9e0-3e22-4992-b0eb-d261f67bfc58" (UID: "3868c9e0-3e22-4992-b0eb-d261f67bfc58"). InnerVolumeSpecName "kube-api-access-btv5b". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 18:03:52 crc kubenswrapper[4754]: I0126 18:03:52.920099 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3868c9e0-3e22-4992-b0eb-d261f67bfc58-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3868c9e0-3e22-4992-b0eb-d261f67bfc58" (UID: "3868c9e0-3e22-4992-b0eb-d261f67bfc58"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 18:03:52 crc kubenswrapper[4754]: I0126 18:03:52.967617 4754 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3868c9e0-3e22-4992-b0eb-d261f67bfc58-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 26 18:03:52 crc kubenswrapper[4754]: I0126 18:03:52.967650 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-btv5b\" (UniqueName: \"kubernetes.io/projected/3868c9e0-3e22-4992-b0eb-d261f67bfc58-kube-api-access-btv5b\") on node \"crc\" DevicePath \"\"" Jan 26 18:03:52 crc kubenswrapper[4754]: I0126 18:03:52.967684 4754 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3868c9e0-3e22-4992-b0eb-d261f67bfc58-utilities\") on node \"crc\" DevicePath \"\"" Jan 26 18:03:53 crc kubenswrapper[4754]: I0126 18:03:53.704141 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jqlp7" Jan 26 18:03:53 crc kubenswrapper[4754]: I0126 18:03:53.779983 4754 scope.go:117] "RemoveContainer" containerID="b36649b0f03300e411a658fea6f425b2be30bb8d773b1d9ef5bb8f521458f9a6" Jan 26 18:03:53 crc kubenswrapper[4754]: E0126 18:03:53.780326 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x65wv_openshift-machine-config-operator(8c3718a4-f354-4284-92e0-fdfb45a692bd)\"" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" Jan 26 18:03:53 crc kubenswrapper[4754]: I0126 18:03:53.797640 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-jqlp7"] Jan 26 18:03:53 crc kubenswrapper[4754]: I0126 18:03:53.802622 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-jqlp7"] Jan 26 18:03:55 crc kubenswrapper[4754]: I0126 18:03:55.778926 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3868c9e0-3e22-4992-b0eb-d261f67bfc58" path="/var/lib/kubelet/pods/3868c9e0-3e22-4992-b0eb-d261f67bfc58/volumes" Jan 26 18:04:07 crc kubenswrapper[4754]: I0126 18:04:07.767848 4754 scope.go:117] "RemoveContainer" containerID="b36649b0f03300e411a658fea6f425b2be30bb8d773b1d9ef5bb8f521458f9a6" Jan 26 18:04:07 crc kubenswrapper[4754]: E0126 18:04:07.769737 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x65wv_openshift-machine-config-operator(8c3718a4-f354-4284-92e0-fdfb45a692bd)\"" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" Jan 26 18:04:21 crc kubenswrapper[4754]: I0126 18:04:21.767744 4754 scope.go:117] "RemoveContainer" containerID="b36649b0f03300e411a658fea6f425b2be30bb8d773b1d9ef5bb8f521458f9a6" Jan 26 18:04:21 crc kubenswrapper[4754]: E0126 18:04:21.768893 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x65wv_openshift-machine-config-operator(8c3718a4-f354-4284-92e0-fdfb45a692bd)\"" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" Jan 26 18:04:23 crc kubenswrapper[4754]: I0126 18:04:23.017680 4754 generic.go:334] "Generic (PLEG): container finished" podID="bea1bcb2-12bb-4034-b289-9d7eda6175b2" containerID="d432469e4534c9613a668b7e303bbc4ee01cac5fb65975e5c21882834d3a8d11" exitCode=0 Jan 26 18:04:23 crc kubenswrapper[4754]: I0126 18:04:23.017774 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-zb6c2/must-gather-xhxjk" event={"ID":"bea1bcb2-12bb-4034-b289-9d7eda6175b2","Type":"ContainerDied","Data":"d432469e4534c9613a668b7e303bbc4ee01cac5fb65975e5c21882834d3a8d11"} Jan 26 18:04:23 crc kubenswrapper[4754]: I0126 18:04:23.018742 4754 scope.go:117] "RemoveContainer" containerID="d432469e4534c9613a668b7e303bbc4ee01cac5fb65975e5c21882834d3a8d11" Jan 26 18:04:23 crc kubenswrapper[4754]: I0126 18:04:23.815872 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-zb6c2_must-gather-xhxjk_bea1bcb2-12bb-4034-b289-9d7eda6175b2/gather/0.log" Jan 26 18:04:33 crc kubenswrapper[4754]: I0126 18:04:33.752022 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-zb6c2/must-gather-xhxjk"] Jan 26 18:04:33 crc kubenswrapper[4754]: I0126 18:04:33.752798 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-zb6c2/must-gather-xhxjk" podUID="bea1bcb2-12bb-4034-b289-9d7eda6175b2" containerName="copy" containerID="cri-o://479d15afd09a697338c933f25fd2f03484df7b08f42579a078a42e06f51bbcfc" gracePeriod=2 Jan 26 18:04:33 crc kubenswrapper[4754]: I0126 18:04:33.778824 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-zb6c2/must-gather-xhxjk"] Jan 26 18:04:34 crc kubenswrapper[4754]: I0126 18:04:34.139860 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-zb6c2_must-gather-xhxjk_bea1bcb2-12bb-4034-b289-9d7eda6175b2/copy/0.log" Jan 26 18:04:34 crc kubenswrapper[4754]: I0126 18:04:34.140504 4754 generic.go:334] "Generic (PLEG): container finished" podID="bea1bcb2-12bb-4034-b289-9d7eda6175b2" containerID="479d15afd09a697338c933f25fd2f03484df7b08f42579a078a42e06f51bbcfc" exitCode=143 Jan 26 18:04:34 crc kubenswrapper[4754]: I0126 18:04:34.215934 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-zb6c2_must-gather-xhxjk_bea1bcb2-12bb-4034-b289-9d7eda6175b2/copy/0.log" Jan 26 18:04:34 crc kubenswrapper[4754]: I0126 18:04:34.216295 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-zb6c2/must-gather-xhxjk" Jan 26 18:04:34 crc kubenswrapper[4754]: I0126 18:04:34.260195 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/bea1bcb2-12bb-4034-b289-9d7eda6175b2-must-gather-output\") pod \"bea1bcb2-12bb-4034-b289-9d7eda6175b2\" (UID: \"bea1bcb2-12bb-4034-b289-9d7eda6175b2\") " Jan 26 18:04:34 crc kubenswrapper[4754]: I0126 18:04:34.260313 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fdfn4\" (UniqueName: \"kubernetes.io/projected/bea1bcb2-12bb-4034-b289-9d7eda6175b2-kube-api-access-fdfn4\") pod \"bea1bcb2-12bb-4034-b289-9d7eda6175b2\" (UID: \"bea1bcb2-12bb-4034-b289-9d7eda6175b2\") " Jan 26 18:04:34 crc kubenswrapper[4754]: I0126 18:04:34.267783 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bea1bcb2-12bb-4034-b289-9d7eda6175b2-kube-api-access-fdfn4" (OuterVolumeSpecName: "kube-api-access-fdfn4") pod "bea1bcb2-12bb-4034-b289-9d7eda6175b2" (UID: "bea1bcb2-12bb-4034-b289-9d7eda6175b2"). InnerVolumeSpecName "kube-api-access-fdfn4". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 18:04:34 crc kubenswrapper[4754]: I0126 18:04:34.362038 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fdfn4\" (UniqueName: \"kubernetes.io/projected/bea1bcb2-12bb-4034-b289-9d7eda6175b2-kube-api-access-fdfn4\") on node \"crc\" DevicePath \"\"" Jan 26 18:04:34 crc kubenswrapper[4754]: I0126 18:04:34.408851 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bea1bcb2-12bb-4034-b289-9d7eda6175b2-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "bea1bcb2-12bb-4034-b289-9d7eda6175b2" (UID: "bea1bcb2-12bb-4034-b289-9d7eda6175b2"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 18:04:34 crc kubenswrapper[4754]: I0126 18:04:34.465161 4754 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/bea1bcb2-12bb-4034-b289-9d7eda6175b2-must-gather-output\") on node \"crc\" DevicePath \"\"" Jan 26 18:04:35 crc kubenswrapper[4754]: I0126 18:04:35.153342 4754 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-zb6c2_must-gather-xhxjk_bea1bcb2-12bb-4034-b289-9d7eda6175b2/copy/0.log" Jan 26 18:04:35 crc kubenswrapper[4754]: I0126 18:04:35.153721 4754 scope.go:117] "RemoveContainer" containerID="479d15afd09a697338c933f25fd2f03484df7b08f42579a078a42e06f51bbcfc" Jan 26 18:04:35 crc kubenswrapper[4754]: I0126 18:04:35.153870 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-zb6c2/must-gather-xhxjk" Jan 26 18:04:35 crc kubenswrapper[4754]: I0126 18:04:35.172885 4754 scope.go:117] "RemoveContainer" containerID="d432469e4534c9613a668b7e303bbc4ee01cac5fb65975e5c21882834d3a8d11" Jan 26 18:04:35 crc kubenswrapper[4754]: I0126 18:04:35.767411 4754 scope.go:117] "RemoveContainer" containerID="b36649b0f03300e411a658fea6f425b2be30bb8d773b1d9ef5bb8f521458f9a6" Jan 26 18:04:35 crc kubenswrapper[4754]: E0126 18:04:35.768026 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x65wv_openshift-machine-config-operator(8c3718a4-f354-4284-92e0-fdfb45a692bd)\"" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" Jan 26 18:04:35 crc kubenswrapper[4754]: I0126 18:04:35.779574 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bea1bcb2-12bb-4034-b289-9d7eda6175b2" path="/var/lib/kubelet/pods/bea1bcb2-12bb-4034-b289-9d7eda6175b2/volumes" Jan 26 18:04:42 crc kubenswrapper[4754]: I0126 18:04:42.325954 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-dpptd"] Jan 26 18:04:42 crc kubenswrapper[4754]: E0126 18:04:42.326850 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3868c9e0-3e22-4992-b0eb-d261f67bfc58" containerName="extract-utilities" Jan 26 18:04:42 crc kubenswrapper[4754]: I0126 18:04:42.326864 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="3868c9e0-3e22-4992-b0eb-d261f67bfc58" containerName="extract-utilities" Jan 26 18:04:42 crc kubenswrapper[4754]: E0126 18:04:42.326880 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3868c9e0-3e22-4992-b0eb-d261f67bfc58" containerName="extract-content" Jan 26 18:04:42 crc kubenswrapper[4754]: I0126 18:04:42.326887 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="3868c9e0-3e22-4992-b0eb-d261f67bfc58" containerName="extract-content" Jan 26 18:04:42 crc kubenswrapper[4754]: E0126 18:04:42.326911 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bea1bcb2-12bb-4034-b289-9d7eda6175b2" containerName="gather" Jan 26 18:04:42 crc kubenswrapper[4754]: I0126 18:04:42.326917 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="bea1bcb2-12bb-4034-b289-9d7eda6175b2" containerName="gather" Jan 26 18:04:42 crc kubenswrapper[4754]: E0126 18:04:42.326926 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3868c9e0-3e22-4992-b0eb-d261f67bfc58" containerName="registry-server" Jan 26 18:04:42 crc kubenswrapper[4754]: I0126 18:04:42.326932 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="3868c9e0-3e22-4992-b0eb-d261f67bfc58" containerName="registry-server" Jan 26 18:04:42 crc kubenswrapper[4754]: E0126 18:04:42.326945 4754 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bea1bcb2-12bb-4034-b289-9d7eda6175b2" containerName="copy" Jan 26 18:04:42 crc kubenswrapper[4754]: I0126 18:04:42.326951 4754 state_mem.go:107] "Deleted CPUSet assignment" podUID="bea1bcb2-12bb-4034-b289-9d7eda6175b2" containerName="copy" Jan 26 18:04:42 crc kubenswrapper[4754]: I0126 18:04:42.327110 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="bea1bcb2-12bb-4034-b289-9d7eda6175b2" containerName="gather" Jan 26 18:04:42 crc kubenswrapper[4754]: I0126 18:04:42.327122 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="bea1bcb2-12bb-4034-b289-9d7eda6175b2" containerName="copy" Jan 26 18:04:42 crc kubenswrapper[4754]: I0126 18:04:42.327139 4754 memory_manager.go:354] "RemoveStaleState removing state" podUID="3868c9e0-3e22-4992-b0eb-d261f67bfc58" containerName="registry-server" Jan 26 18:04:42 crc kubenswrapper[4754]: I0126 18:04:42.328622 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dpptd" Jan 26 18:04:42 crc kubenswrapper[4754]: I0126 18:04:42.342927 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-dpptd"] Jan 26 18:04:42 crc kubenswrapper[4754]: I0126 18:04:42.407465 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d3988422-fd7f-4971-9180-8a3a91091f3e-utilities\") pod \"certified-operators-dpptd\" (UID: \"d3988422-fd7f-4971-9180-8a3a91091f3e\") " pod="openshift-marketplace/certified-operators-dpptd" Jan 26 18:04:42 crc kubenswrapper[4754]: I0126 18:04:42.407609 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d3988422-fd7f-4971-9180-8a3a91091f3e-catalog-content\") pod \"certified-operators-dpptd\" (UID: \"d3988422-fd7f-4971-9180-8a3a91091f3e\") " pod="openshift-marketplace/certified-operators-dpptd" Jan 26 18:04:42 crc kubenswrapper[4754]: I0126 18:04:42.407631 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2t8v2\" (UniqueName: \"kubernetes.io/projected/d3988422-fd7f-4971-9180-8a3a91091f3e-kube-api-access-2t8v2\") pod \"certified-operators-dpptd\" (UID: \"d3988422-fd7f-4971-9180-8a3a91091f3e\") " pod="openshift-marketplace/certified-operators-dpptd" Jan 26 18:04:42 crc kubenswrapper[4754]: I0126 18:04:42.509497 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d3988422-fd7f-4971-9180-8a3a91091f3e-catalog-content\") pod \"certified-operators-dpptd\" (UID: \"d3988422-fd7f-4971-9180-8a3a91091f3e\") " pod="openshift-marketplace/certified-operators-dpptd" Jan 26 18:04:42 crc kubenswrapper[4754]: I0126 18:04:42.509546 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2t8v2\" (UniqueName: \"kubernetes.io/projected/d3988422-fd7f-4971-9180-8a3a91091f3e-kube-api-access-2t8v2\") pod \"certified-operators-dpptd\" (UID: \"d3988422-fd7f-4971-9180-8a3a91091f3e\") " pod="openshift-marketplace/certified-operators-dpptd" Jan 26 18:04:42 crc kubenswrapper[4754]: I0126 18:04:42.509609 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d3988422-fd7f-4971-9180-8a3a91091f3e-utilities\") pod \"certified-operators-dpptd\" (UID: \"d3988422-fd7f-4971-9180-8a3a91091f3e\") " pod="openshift-marketplace/certified-operators-dpptd" Jan 26 18:04:42 crc kubenswrapper[4754]: I0126 18:04:42.510092 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d3988422-fd7f-4971-9180-8a3a91091f3e-utilities\") pod \"certified-operators-dpptd\" (UID: \"d3988422-fd7f-4971-9180-8a3a91091f3e\") " pod="openshift-marketplace/certified-operators-dpptd" Jan 26 18:04:42 crc kubenswrapper[4754]: I0126 18:04:42.510334 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d3988422-fd7f-4971-9180-8a3a91091f3e-catalog-content\") pod \"certified-operators-dpptd\" (UID: \"d3988422-fd7f-4971-9180-8a3a91091f3e\") " pod="openshift-marketplace/certified-operators-dpptd" Jan 26 18:04:42 crc kubenswrapper[4754]: I0126 18:04:42.539965 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2t8v2\" (UniqueName: \"kubernetes.io/projected/d3988422-fd7f-4971-9180-8a3a91091f3e-kube-api-access-2t8v2\") pod \"certified-operators-dpptd\" (UID: \"d3988422-fd7f-4971-9180-8a3a91091f3e\") " pod="openshift-marketplace/certified-operators-dpptd" Jan 26 18:04:42 crc kubenswrapper[4754]: I0126 18:04:42.650652 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dpptd" Jan 26 18:04:43 crc kubenswrapper[4754]: I0126 18:04:43.201469 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-dpptd"] Jan 26 18:04:44 crc kubenswrapper[4754]: I0126 18:04:44.228477 4754 generic.go:334] "Generic (PLEG): container finished" podID="d3988422-fd7f-4971-9180-8a3a91091f3e" containerID="3fbac77ed6aa5757c4a18fe32e05570c2cd404de36dc7d095bd07f189aadce2c" exitCode=0 Jan 26 18:04:44 crc kubenswrapper[4754]: I0126 18:04:44.228594 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dpptd" event={"ID":"d3988422-fd7f-4971-9180-8a3a91091f3e","Type":"ContainerDied","Data":"3fbac77ed6aa5757c4a18fe32e05570c2cd404de36dc7d095bd07f189aadce2c"} Jan 26 18:04:44 crc kubenswrapper[4754]: I0126 18:04:44.229051 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dpptd" event={"ID":"d3988422-fd7f-4971-9180-8a3a91091f3e","Type":"ContainerStarted","Data":"a5adad0a758b8e0de02d03108e4f38147c7199ae794a1c58dce499af5d313d0f"} Jan 26 18:04:44 crc kubenswrapper[4754]: I0126 18:04:44.232106 4754 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 26 18:04:44 crc kubenswrapper[4754]: I0126 18:04:44.734076 4754 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-c6dwz"] Jan 26 18:04:44 crc kubenswrapper[4754]: I0126 18:04:44.737033 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-c6dwz" Jan 26 18:04:44 crc kubenswrapper[4754]: I0126 18:04:44.746715 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-c6dwz"] Jan 26 18:04:44 crc kubenswrapper[4754]: I0126 18:04:44.760832 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2a7b938b-2fe1-42e3-b7e2-b2f3b4636522-utilities\") pod \"redhat-operators-c6dwz\" (UID: \"2a7b938b-2fe1-42e3-b7e2-b2f3b4636522\") " pod="openshift-marketplace/redhat-operators-c6dwz" Jan 26 18:04:44 crc kubenswrapper[4754]: I0126 18:04:44.760872 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nlkl8\" (UniqueName: \"kubernetes.io/projected/2a7b938b-2fe1-42e3-b7e2-b2f3b4636522-kube-api-access-nlkl8\") pod \"redhat-operators-c6dwz\" (UID: \"2a7b938b-2fe1-42e3-b7e2-b2f3b4636522\") " pod="openshift-marketplace/redhat-operators-c6dwz" Jan 26 18:04:44 crc kubenswrapper[4754]: I0126 18:04:44.760935 4754 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2a7b938b-2fe1-42e3-b7e2-b2f3b4636522-catalog-content\") pod \"redhat-operators-c6dwz\" (UID: \"2a7b938b-2fe1-42e3-b7e2-b2f3b4636522\") " pod="openshift-marketplace/redhat-operators-c6dwz" Jan 26 18:04:44 crc kubenswrapper[4754]: I0126 18:04:44.862407 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2a7b938b-2fe1-42e3-b7e2-b2f3b4636522-catalog-content\") pod \"redhat-operators-c6dwz\" (UID: \"2a7b938b-2fe1-42e3-b7e2-b2f3b4636522\") " pod="openshift-marketplace/redhat-operators-c6dwz" Jan 26 18:04:44 crc kubenswrapper[4754]: I0126 18:04:44.862624 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2a7b938b-2fe1-42e3-b7e2-b2f3b4636522-utilities\") pod \"redhat-operators-c6dwz\" (UID: \"2a7b938b-2fe1-42e3-b7e2-b2f3b4636522\") " pod="openshift-marketplace/redhat-operators-c6dwz" Jan 26 18:04:44 crc kubenswrapper[4754]: I0126 18:04:44.862689 4754 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nlkl8\" (UniqueName: \"kubernetes.io/projected/2a7b938b-2fe1-42e3-b7e2-b2f3b4636522-kube-api-access-nlkl8\") pod \"redhat-operators-c6dwz\" (UID: \"2a7b938b-2fe1-42e3-b7e2-b2f3b4636522\") " pod="openshift-marketplace/redhat-operators-c6dwz" Jan 26 18:04:44 crc kubenswrapper[4754]: I0126 18:04:44.862869 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2a7b938b-2fe1-42e3-b7e2-b2f3b4636522-catalog-content\") pod \"redhat-operators-c6dwz\" (UID: \"2a7b938b-2fe1-42e3-b7e2-b2f3b4636522\") " pod="openshift-marketplace/redhat-operators-c6dwz" Jan 26 18:04:44 crc kubenswrapper[4754]: I0126 18:04:44.863209 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2a7b938b-2fe1-42e3-b7e2-b2f3b4636522-utilities\") pod \"redhat-operators-c6dwz\" (UID: \"2a7b938b-2fe1-42e3-b7e2-b2f3b4636522\") " pod="openshift-marketplace/redhat-operators-c6dwz" Jan 26 18:04:44 crc kubenswrapper[4754]: I0126 18:04:44.884962 4754 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nlkl8\" (UniqueName: \"kubernetes.io/projected/2a7b938b-2fe1-42e3-b7e2-b2f3b4636522-kube-api-access-nlkl8\") pod \"redhat-operators-c6dwz\" (UID: \"2a7b938b-2fe1-42e3-b7e2-b2f3b4636522\") " pod="openshift-marketplace/redhat-operators-c6dwz" Jan 26 18:04:45 crc kubenswrapper[4754]: I0126 18:04:45.073518 4754 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-c6dwz" Jan 26 18:04:45 crc kubenswrapper[4754]: I0126 18:04:45.909464 4754 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-c6dwz"] Jan 26 18:04:45 crc kubenswrapper[4754]: W0126 18:04:45.917117 4754 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2a7b938b_2fe1_42e3_b7e2_b2f3b4636522.slice/crio-55698ee2bf4eff2ce43fbb08665333db64e239b914cb9819e57d5397e3094da0 WatchSource:0}: Error finding container 55698ee2bf4eff2ce43fbb08665333db64e239b914cb9819e57d5397e3094da0: Status 404 returned error can't find the container with id 55698ee2bf4eff2ce43fbb08665333db64e239b914cb9819e57d5397e3094da0 Jan 26 18:04:46 crc kubenswrapper[4754]: I0126 18:04:46.250857 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-c6dwz" event={"ID":"2a7b938b-2fe1-42e3-b7e2-b2f3b4636522","Type":"ContainerStarted","Data":"55698ee2bf4eff2ce43fbb08665333db64e239b914cb9819e57d5397e3094da0"} Jan 26 18:04:46 crc kubenswrapper[4754]: I0126 18:04:46.253827 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dpptd" event={"ID":"d3988422-fd7f-4971-9180-8a3a91091f3e","Type":"ContainerStarted","Data":"7be5ea1ce32213f26fec6265bce354bdb00a7345228261972d567430e2ce0d51"} Jan 26 18:04:47 crc kubenswrapper[4754]: I0126 18:04:47.265634 4754 generic.go:334] "Generic (PLEG): container finished" podID="d3988422-fd7f-4971-9180-8a3a91091f3e" containerID="7be5ea1ce32213f26fec6265bce354bdb00a7345228261972d567430e2ce0d51" exitCode=0 Jan 26 18:04:47 crc kubenswrapper[4754]: I0126 18:04:47.265752 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dpptd" event={"ID":"d3988422-fd7f-4971-9180-8a3a91091f3e","Type":"ContainerDied","Data":"7be5ea1ce32213f26fec6265bce354bdb00a7345228261972d567430e2ce0d51"} Jan 26 18:04:47 crc kubenswrapper[4754]: I0126 18:04:47.267362 4754 generic.go:334] "Generic (PLEG): container finished" podID="2a7b938b-2fe1-42e3-b7e2-b2f3b4636522" containerID="fb293a9771bbfce23a2bba9103437149ff103e79eb4acc8e2648e8c22877f4de" exitCode=0 Jan 26 18:04:47 crc kubenswrapper[4754]: I0126 18:04:47.267400 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-c6dwz" event={"ID":"2a7b938b-2fe1-42e3-b7e2-b2f3b4636522","Type":"ContainerDied","Data":"fb293a9771bbfce23a2bba9103437149ff103e79eb4acc8e2648e8c22877f4de"} Jan 26 18:04:48 crc kubenswrapper[4754]: I0126 18:04:48.277299 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dpptd" event={"ID":"d3988422-fd7f-4971-9180-8a3a91091f3e","Type":"ContainerStarted","Data":"86abc93e34f85a49613bee0e62f4a0088543786ca9e467e74af3556e92d6b48c"} Jan 26 18:04:48 crc kubenswrapper[4754]: I0126 18:04:48.303103 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-dpptd" podStartSLOduration=2.842723811 podStartE2EDuration="6.303081975s" podCreationTimestamp="2026-01-26 18:04:42 +0000 UTC" firstStartedPulling="2026-01-26 18:04:44.231526683 +0000 UTC m=+4650.755707147" lastFinishedPulling="2026-01-26 18:04:47.691884877 +0000 UTC m=+4654.216065311" observedRunningTime="2026-01-26 18:04:48.299282732 +0000 UTC m=+4654.823463186" watchObservedRunningTime="2026-01-26 18:04:48.303081975 +0000 UTC m=+4654.827262419" Jan 26 18:04:49 crc kubenswrapper[4754]: I0126 18:04:49.292678 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-c6dwz" event={"ID":"2a7b938b-2fe1-42e3-b7e2-b2f3b4636522","Type":"ContainerStarted","Data":"3000d1f35a8b1cc0ff9ca5fedb1413d7d1d42fcb9024239aebda862b5a3fe581"} Jan 26 18:04:49 crc kubenswrapper[4754]: I0126 18:04:49.767246 4754 scope.go:117] "RemoveContainer" containerID="b36649b0f03300e411a658fea6f425b2be30bb8d773b1d9ef5bb8f521458f9a6" Jan 26 18:04:49 crc kubenswrapper[4754]: E0126 18:04:49.767758 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x65wv_openshift-machine-config-operator(8c3718a4-f354-4284-92e0-fdfb45a692bd)\"" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" Jan 26 18:04:50 crc kubenswrapper[4754]: I0126 18:04:50.304990 4754 generic.go:334] "Generic (PLEG): container finished" podID="2a7b938b-2fe1-42e3-b7e2-b2f3b4636522" containerID="3000d1f35a8b1cc0ff9ca5fedb1413d7d1d42fcb9024239aebda862b5a3fe581" exitCode=0 Jan 26 18:04:50 crc kubenswrapper[4754]: I0126 18:04:50.305294 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-c6dwz" event={"ID":"2a7b938b-2fe1-42e3-b7e2-b2f3b4636522","Type":"ContainerDied","Data":"3000d1f35a8b1cc0ff9ca5fedb1413d7d1d42fcb9024239aebda862b5a3fe581"} Jan 26 18:04:52 crc kubenswrapper[4754]: I0126 18:04:52.324336 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-c6dwz" event={"ID":"2a7b938b-2fe1-42e3-b7e2-b2f3b4636522","Type":"ContainerStarted","Data":"90c00fcb5c463329d4c6e3603bb97ea043755c82a808ed691d3ad46212d044c7"} Jan 26 18:04:52 crc kubenswrapper[4754]: I0126 18:04:52.350423 4754 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-c6dwz" podStartSLOduration=4.6320788440000005 podStartE2EDuration="8.350401514s" podCreationTimestamp="2026-01-26 18:04:44 +0000 UTC" firstStartedPulling="2026-01-26 18:04:47.269554471 +0000 UTC m=+4653.793734925" lastFinishedPulling="2026-01-26 18:04:50.987877161 +0000 UTC m=+4657.512057595" observedRunningTime="2026-01-26 18:04:52.34185866 +0000 UTC m=+4658.866039124" watchObservedRunningTime="2026-01-26 18:04:52.350401514 +0000 UTC m=+4658.874581968" Jan 26 18:04:52 crc kubenswrapper[4754]: I0126 18:04:52.653052 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-dpptd" Jan 26 18:04:52 crc kubenswrapper[4754]: I0126 18:04:52.653495 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-dpptd" Jan 26 18:04:52 crc kubenswrapper[4754]: I0126 18:04:52.695611 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-dpptd" Jan 26 18:04:53 crc kubenswrapper[4754]: I0126 18:04:53.404967 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-dpptd" Jan 26 18:04:54 crc kubenswrapper[4754]: I0126 18:04:54.914505 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-dpptd"] Jan 26 18:04:55 crc kubenswrapper[4754]: I0126 18:04:55.073950 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-c6dwz" Jan 26 18:04:55 crc kubenswrapper[4754]: I0126 18:04:55.074021 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-c6dwz" Jan 26 18:04:55 crc kubenswrapper[4754]: I0126 18:04:55.350415 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-dpptd" podUID="d3988422-fd7f-4971-9180-8a3a91091f3e" containerName="registry-server" containerID="cri-o://86abc93e34f85a49613bee0e62f4a0088543786ca9e467e74af3556e92d6b48c" gracePeriod=2 Jan 26 18:04:56 crc kubenswrapper[4754]: I0126 18:04:56.130118 4754 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-c6dwz" podUID="2a7b938b-2fe1-42e3-b7e2-b2f3b4636522" containerName="registry-server" probeResult="failure" output=< Jan 26 18:04:56 crc kubenswrapper[4754]: timeout: failed to connect service ":50051" within 1s Jan 26 18:04:56 crc kubenswrapper[4754]: > Jan 26 18:04:59 crc kubenswrapper[4754]: I0126 18:04:59.393804 4754 generic.go:334] "Generic (PLEG): container finished" podID="d3988422-fd7f-4971-9180-8a3a91091f3e" containerID="86abc93e34f85a49613bee0e62f4a0088543786ca9e467e74af3556e92d6b48c" exitCode=0 Jan 26 18:04:59 crc kubenswrapper[4754]: I0126 18:04:59.393895 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dpptd" event={"ID":"d3988422-fd7f-4971-9180-8a3a91091f3e","Type":"ContainerDied","Data":"86abc93e34f85a49613bee0e62f4a0088543786ca9e467e74af3556e92d6b48c"} Jan 26 18:05:00 crc kubenswrapper[4754]: I0126 18:05:00.032882 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dpptd" Jan 26 18:05:00 crc kubenswrapper[4754]: I0126 18:05:00.177123 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2t8v2\" (UniqueName: \"kubernetes.io/projected/d3988422-fd7f-4971-9180-8a3a91091f3e-kube-api-access-2t8v2\") pod \"d3988422-fd7f-4971-9180-8a3a91091f3e\" (UID: \"d3988422-fd7f-4971-9180-8a3a91091f3e\") " Jan 26 18:05:00 crc kubenswrapper[4754]: I0126 18:05:00.177297 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d3988422-fd7f-4971-9180-8a3a91091f3e-catalog-content\") pod \"d3988422-fd7f-4971-9180-8a3a91091f3e\" (UID: \"d3988422-fd7f-4971-9180-8a3a91091f3e\") " Jan 26 18:05:00 crc kubenswrapper[4754]: I0126 18:05:00.177328 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d3988422-fd7f-4971-9180-8a3a91091f3e-utilities\") pod \"d3988422-fd7f-4971-9180-8a3a91091f3e\" (UID: \"d3988422-fd7f-4971-9180-8a3a91091f3e\") " Jan 26 18:05:00 crc kubenswrapper[4754]: I0126 18:05:00.178099 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d3988422-fd7f-4971-9180-8a3a91091f3e-utilities" (OuterVolumeSpecName: "utilities") pod "d3988422-fd7f-4971-9180-8a3a91091f3e" (UID: "d3988422-fd7f-4971-9180-8a3a91091f3e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 18:05:00 crc kubenswrapper[4754]: I0126 18:05:00.185024 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d3988422-fd7f-4971-9180-8a3a91091f3e-kube-api-access-2t8v2" (OuterVolumeSpecName: "kube-api-access-2t8v2") pod "d3988422-fd7f-4971-9180-8a3a91091f3e" (UID: "d3988422-fd7f-4971-9180-8a3a91091f3e"). InnerVolumeSpecName "kube-api-access-2t8v2". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 18:05:00 crc kubenswrapper[4754]: I0126 18:05:00.229157 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d3988422-fd7f-4971-9180-8a3a91091f3e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d3988422-fd7f-4971-9180-8a3a91091f3e" (UID: "d3988422-fd7f-4971-9180-8a3a91091f3e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 18:05:00 crc kubenswrapper[4754]: I0126 18:05:00.279511 4754 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d3988422-fd7f-4971-9180-8a3a91091f3e-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 26 18:05:00 crc kubenswrapper[4754]: I0126 18:05:00.279543 4754 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d3988422-fd7f-4971-9180-8a3a91091f3e-utilities\") on node \"crc\" DevicePath \"\"" Jan 26 18:05:00 crc kubenswrapper[4754]: I0126 18:05:00.279554 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2t8v2\" (UniqueName: \"kubernetes.io/projected/d3988422-fd7f-4971-9180-8a3a91091f3e-kube-api-access-2t8v2\") on node \"crc\" DevicePath \"\"" Jan 26 18:05:00 crc kubenswrapper[4754]: I0126 18:05:00.406943 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dpptd" event={"ID":"d3988422-fd7f-4971-9180-8a3a91091f3e","Type":"ContainerDied","Data":"a5adad0a758b8e0de02d03108e4f38147c7199ae794a1c58dce499af5d313d0f"} Jan 26 18:05:00 crc kubenswrapper[4754]: I0126 18:05:00.407004 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dpptd" Jan 26 18:05:00 crc kubenswrapper[4754]: I0126 18:05:00.407024 4754 scope.go:117] "RemoveContainer" containerID="86abc93e34f85a49613bee0e62f4a0088543786ca9e467e74af3556e92d6b48c" Jan 26 18:05:00 crc kubenswrapper[4754]: I0126 18:05:00.452059 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-dpptd"] Jan 26 18:05:00 crc kubenswrapper[4754]: I0126 18:05:00.454812 4754 scope.go:117] "RemoveContainer" containerID="7be5ea1ce32213f26fec6265bce354bdb00a7345228261972d567430e2ce0d51" Jan 26 18:05:00 crc kubenswrapper[4754]: I0126 18:05:00.460014 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-dpptd"] Jan 26 18:05:00 crc kubenswrapper[4754]: I0126 18:05:00.489581 4754 scope.go:117] "RemoveContainer" containerID="3fbac77ed6aa5757c4a18fe32e05570c2cd404de36dc7d095bd07f189aadce2c" Jan 26 18:05:01 crc kubenswrapper[4754]: I0126 18:05:01.784765 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d3988422-fd7f-4971-9180-8a3a91091f3e" path="/var/lib/kubelet/pods/d3988422-fd7f-4971-9180-8a3a91091f3e/volumes" Jan 26 18:05:04 crc kubenswrapper[4754]: I0126 18:05:04.766983 4754 scope.go:117] "RemoveContainer" containerID="b36649b0f03300e411a658fea6f425b2be30bb8d773b1d9ef5bb8f521458f9a6" Jan 26 18:05:04 crc kubenswrapper[4754]: E0126 18:05:04.767624 4754 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x65wv_openshift-machine-config-operator(8c3718a4-f354-4284-92e0-fdfb45a692bd)\"" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" Jan 26 18:05:05 crc kubenswrapper[4754]: I0126 18:05:05.121921 4754 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-c6dwz" Jan 26 18:05:05 crc kubenswrapper[4754]: I0126 18:05:05.171130 4754 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-c6dwz" Jan 26 18:05:05 crc kubenswrapper[4754]: I0126 18:05:05.371192 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-c6dwz"] Jan 26 18:05:06 crc kubenswrapper[4754]: I0126 18:05:06.463035 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-c6dwz" podUID="2a7b938b-2fe1-42e3-b7e2-b2f3b4636522" containerName="registry-server" containerID="cri-o://90c00fcb5c463329d4c6e3603bb97ea043755c82a808ed691d3ad46212d044c7" gracePeriod=2 Jan 26 18:05:06 crc kubenswrapper[4754]: I0126 18:05:06.967799 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-c6dwz" Jan 26 18:05:07 crc kubenswrapper[4754]: I0126 18:05:07.123583 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2a7b938b-2fe1-42e3-b7e2-b2f3b4636522-catalog-content\") pod \"2a7b938b-2fe1-42e3-b7e2-b2f3b4636522\" (UID: \"2a7b938b-2fe1-42e3-b7e2-b2f3b4636522\") " Jan 26 18:05:07 crc kubenswrapper[4754]: I0126 18:05:07.123788 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2a7b938b-2fe1-42e3-b7e2-b2f3b4636522-utilities\") pod \"2a7b938b-2fe1-42e3-b7e2-b2f3b4636522\" (UID: \"2a7b938b-2fe1-42e3-b7e2-b2f3b4636522\") " Jan 26 18:05:07 crc kubenswrapper[4754]: I0126 18:05:07.123915 4754 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nlkl8\" (UniqueName: \"kubernetes.io/projected/2a7b938b-2fe1-42e3-b7e2-b2f3b4636522-kube-api-access-nlkl8\") pod \"2a7b938b-2fe1-42e3-b7e2-b2f3b4636522\" (UID: \"2a7b938b-2fe1-42e3-b7e2-b2f3b4636522\") " Jan 26 18:05:07 crc kubenswrapper[4754]: I0126 18:05:07.124510 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2a7b938b-2fe1-42e3-b7e2-b2f3b4636522-utilities" (OuterVolumeSpecName: "utilities") pod "2a7b938b-2fe1-42e3-b7e2-b2f3b4636522" (UID: "2a7b938b-2fe1-42e3-b7e2-b2f3b4636522"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 18:05:07 crc kubenswrapper[4754]: I0126 18:05:07.136573 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2a7b938b-2fe1-42e3-b7e2-b2f3b4636522-kube-api-access-nlkl8" (OuterVolumeSpecName: "kube-api-access-nlkl8") pod "2a7b938b-2fe1-42e3-b7e2-b2f3b4636522" (UID: "2a7b938b-2fe1-42e3-b7e2-b2f3b4636522"). InnerVolumeSpecName "kube-api-access-nlkl8". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 18:05:07 crc kubenswrapper[4754]: I0126 18:05:07.227084 4754 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2a7b938b-2fe1-42e3-b7e2-b2f3b4636522-utilities\") on node \"crc\" DevicePath \"\"" Jan 26 18:05:07 crc kubenswrapper[4754]: I0126 18:05:07.227121 4754 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nlkl8\" (UniqueName: \"kubernetes.io/projected/2a7b938b-2fe1-42e3-b7e2-b2f3b4636522-kube-api-access-nlkl8\") on node \"crc\" DevicePath \"\"" Jan 26 18:05:07 crc kubenswrapper[4754]: I0126 18:05:07.265956 4754 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2a7b938b-2fe1-42e3-b7e2-b2f3b4636522-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2a7b938b-2fe1-42e3-b7e2-b2f3b4636522" (UID: "2a7b938b-2fe1-42e3-b7e2-b2f3b4636522"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 18:05:07 crc kubenswrapper[4754]: I0126 18:05:07.328804 4754 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2a7b938b-2fe1-42e3-b7e2-b2f3b4636522-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 26 18:05:07 crc kubenswrapper[4754]: I0126 18:05:07.475614 4754 generic.go:334] "Generic (PLEG): container finished" podID="2a7b938b-2fe1-42e3-b7e2-b2f3b4636522" containerID="90c00fcb5c463329d4c6e3603bb97ea043755c82a808ed691d3ad46212d044c7" exitCode=0 Jan 26 18:05:07 crc kubenswrapper[4754]: I0126 18:05:07.475656 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-c6dwz" event={"ID":"2a7b938b-2fe1-42e3-b7e2-b2f3b4636522","Type":"ContainerDied","Data":"90c00fcb5c463329d4c6e3603bb97ea043755c82a808ed691d3ad46212d044c7"} Jan 26 18:05:07 crc kubenswrapper[4754]: I0126 18:05:07.475697 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-c6dwz" event={"ID":"2a7b938b-2fe1-42e3-b7e2-b2f3b4636522","Type":"ContainerDied","Data":"55698ee2bf4eff2ce43fbb08665333db64e239b914cb9819e57d5397e3094da0"} Jan 26 18:05:07 crc kubenswrapper[4754]: I0126 18:05:07.475713 4754 scope.go:117] "RemoveContainer" containerID="90c00fcb5c463329d4c6e3603bb97ea043755c82a808ed691d3ad46212d044c7" Jan 26 18:05:07 crc kubenswrapper[4754]: I0126 18:05:07.475834 4754 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-c6dwz" Jan 26 18:05:07 crc kubenswrapper[4754]: I0126 18:05:07.500176 4754 scope.go:117] "RemoveContainer" containerID="3000d1f35a8b1cc0ff9ca5fedb1413d7d1d42fcb9024239aebda862b5a3fe581" Jan 26 18:05:07 crc kubenswrapper[4754]: I0126 18:05:07.533559 4754 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-c6dwz"] Jan 26 18:05:07 crc kubenswrapper[4754]: I0126 18:05:07.537202 4754 scope.go:117] "RemoveContainer" containerID="fb293a9771bbfce23a2bba9103437149ff103e79eb4acc8e2648e8c22877f4de" Jan 26 18:05:07 crc kubenswrapper[4754]: I0126 18:05:07.549823 4754 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-c6dwz"] Jan 26 18:05:07 crc kubenswrapper[4754]: I0126 18:05:07.574260 4754 scope.go:117] "RemoveContainer" containerID="90c00fcb5c463329d4c6e3603bb97ea043755c82a808ed691d3ad46212d044c7" Jan 26 18:05:07 crc kubenswrapper[4754]: E0126 18:05:07.574973 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"90c00fcb5c463329d4c6e3603bb97ea043755c82a808ed691d3ad46212d044c7\": container with ID starting with 90c00fcb5c463329d4c6e3603bb97ea043755c82a808ed691d3ad46212d044c7 not found: ID does not exist" containerID="90c00fcb5c463329d4c6e3603bb97ea043755c82a808ed691d3ad46212d044c7" Jan 26 18:05:07 crc kubenswrapper[4754]: I0126 18:05:07.575011 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"90c00fcb5c463329d4c6e3603bb97ea043755c82a808ed691d3ad46212d044c7"} err="failed to get container status \"90c00fcb5c463329d4c6e3603bb97ea043755c82a808ed691d3ad46212d044c7\": rpc error: code = NotFound desc = could not find container \"90c00fcb5c463329d4c6e3603bb97ea043755c82a808ed691d3ad46212d044c7\": container with ID starting with 90c00fcb5c463329d4c6e3603bb97ea043755c82a808ed691d3ad46212d044c7 not found: ID does not exist" Jan 26 18:05:07 crc kubenswrapper[4754]: I0126 18:05:07.575033 4754 scope.go:117] "RemoveContainer" containerID="3000d1f35a8b1cc0ff9ca5fedb1413d7d1d42fcb9024239aebda862b5a3fe581" Jan 26 18:05:07 crc kubenswrapper[4754]: E0126 18:05:07.575485 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3000d1f35a8b1cc0ff9ca5fedb1413d7d1d42fcb9024239aebda862b5a3fe581\": container with ID starting with 3000d1f35a8b1cc0ff9ca5fedb1413d7d1d42fcb9024239aebda862b5a3fe581 not found: ID does not exist" containerID="3000d1f35a8b1cc0ff9ca5fedb1413d7d1d42fcb9024239aebda862b5a3fe581" Jan 26 18:05:07 crc kubenswrapper[4754]: I0126 18:05:07.575506 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3000d1f35a8b1cc0ff9ca5fedb1413d7d1d42fcb9024239aebda862b5a3fe581"} err="failed to get container status \"3000d1f35a8b1cc0ff9ca5fedb1413d7d1d42fcb9024239aebda862b5a3fe581\": rpc error: code = NotFound desc = could not find container \"3000d1f35a8b1cc0ff9ca5fedb1413d7d1d42fcb9024239aebda862b5a3fe581\": container with ID starting with 3000d1f35a8b1cc0ff9ca5fedb1413d7d1d42fcb9024239aebda862b5a3fe581 not found: ID does not exist" Jan 26 18:05:07 crc kubenswrapper[4754]: I0126 18:05:07.575518 4754 scope.go:117] "RemoveContainer" containerID="fb293a9771bbfce23a2bba9103437149ff103e79eb4acc8e2648e8c22877f4de" Jan 26 18:05:07 crc kubenswrapper[4754]: E0126 18:05:07.575902 4754 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fb293a9771bbfce23a2bba9103437149ff103e79eb4acc8e2648e8c22877f4de\": container with ID starting with fb293a9771bbfce23a2bba9103437149ff103e79eb4acc8e2648e8c22877f4de not found: ID does not exist" containerID="fb293a9771bbfce23a2bba9103437149ff103e79eb4acc8e2648e8c22877f4de" Jan 26 18:05:07 crc kubenswrapper[4754]: I0126 18:05:07.575924 4754 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fb293a9771bbfce23a2bba9103437149ff103e79eb4acc8e2648e8c22877f4de"} err="failed to get container status \"fb293a9771bbfce23a2bba9103437149ff103e79eb4acc8e2648e8c22877f4de\": rpc error: code = NotFound desc = could not find container \"fb293a9771bbfce23a2bba9103437149ff103e79eb4acc8e2648e8c22877f4de\": container with ID starting with fb293a9771bbfce23a2bba9103437149ff103e79eb4acc8e2648e8c22877f4de not found: ID does not exist" Jan 26 18:05:07 crc kubenswrapper[4754]: I0126 18:05:07.779736 4754 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2a7b938b-2fe1-42e3-b7e2-b2f3b4636522" path="/var/lib/kubelet/pods/2a7b938b-2fe1-42e3-b7e2-b2f3b4636522/volumes" Jan 26 18:05:15 crc kubenswrapper[4754]: I0126 18:05:15.768298 4754 scope.go:117] "RemoveContainer" containerID="b36649b0f03300e411a658fea6f425b2be30bb8d773b1d9ef5bb8f521458f9a6" Jan 26 18:05:16 crc kubenswrapper[4754]: I0126 18:05:16.594495 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" event={"ID":"8c3718a4-f354-4284-92e0-fdfb45a692bd","Type":"ContainerStarted","Data":"f6f8d201926ac0f12ac99d54195c10c45660be1b2d4574efa22c782c862e87c5"} Jan 26 18:05:28 crc kubenswrapper[4754]: I0126 18:05:28.667560 4754 scope.go:117] "RemoveContainer" containerID="c7923bfbe51ba7551e08e70ea0f69993fab9e8005f2bbb3eb514c63a240c7f43" Jan 26 18:07:37 crc kubenswrapper[4754]: I0126 18:07:37.129398 4754 patch_prober.go:28] interesting pod/machine-config-daemon-x65wv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 26 18:07:37 crc kubenswrapper[4754]: I0126 18:07:37.131744 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 26 18:08:07 crc kubenswrapper[4754]: I0126 18:08:07.129159 4754 patch_prober.go:28] interesting pod/machine-config-daemon-x65wv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 26 18:08:07 crc kubenswrapper[4754]: I0126 18:08:07.129624 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 26 18:08:37 crc kubenswrapper[4754]: I0126 18:08:37.130021 4754 patch_prober.go:28] interesting pod/machine-config-daemon-x65wv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 26 18:08:37 crc kubenswrapper[4754]: I0126 18:08:37.131052 4754 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 26 18:08:37 crc kubenswrapper[4754]: I0126 18:08:37.131113 4754 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" Jan 26 18:08:37 crc kubenswrapper[4754]: I0126 18:08:37.132218 4754 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f6f8d201926ac0f12ac99d54195c10c45660be1b2d4574efa22c782c862e87c5"} pod="openshift-machine-config-operator/machine-config-daemon-x65wv" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 26 18:08:37 crc kubenswrapper[4754]: I0126 18:08:37.132275 4754 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" podUID="8c3718a4-f354-4284-92e0-fdfb45a692bd" containerName="machine-config-daemon" containerID="cri-o://f6f8d201926ac0f12ac99d54195c10c45660be1b2d4574efa22c782c862e87c5" gracePeriod=600 Jan 26 18:08:37 crc kubenswrapper[4754]: I0126 18:08:37.510319 4754 generic.go:334] "Generic (PLEG): container finished" podID="8c3718a4-f354-4284-92e0-fdfb45a692bd" containerID="f6f8d201926ac0f12ac99d54195c10c45660be1b2d4574efa22c782c862e87c5" exitCode=0 Jan 26 18:08:37 crc kubenswrapper[4754]: I0126 18:08:37.510631 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" event={"ID":"8c3718a4-f354-4284-92e0-fdfb45a692bd","Type":"ContainerDied","Data":"f6f8d201926ac0f12ac99d54195c10c45660be1b2d4574efa22c782c862e87c5"} Jan 26 18:08:37 crc kubenswrapper[4754]: I0126 18:08:37.510753 4754 scope.go:117] "RemoveContainer" containerID="b36649b0f03300e411a658fea6f425b2be30bb8d773b1d9ef5bb8f521458f9a6" Jan 26 18:08:38 crc kubenswrapper[4754]: I0126 18:08:38.520792 4754 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-x65wv" event={"ID":"8c3718a4-f354-4284-92e0-fdfb45a692bd","Type":"ContainerStarted","Data":"82604ae7e1fb25d8c986aabd751a96ab80274255e9d1a2ecaaa3bc8ae04fb593"} var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515135726715024461 0ustar coreroot  Om77'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015135726716017377 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015135714716016517 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015135714716015467 5ustar corecore